var/home/core/zuul-output/0000755000175000017500000000000015115322113014517 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015115333146015473 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000004673145315115333140017706 0ustar rootrootDec 07 16:02:16 crc systemd[1]: Starting Kubernetes Kubelet... Dec 07 16:02:16 crc restorecon[4689]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:16 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 07 16:02:17 crc restorecon[4689]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 07 16:02:17 crc restorecon[4689]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Dec 07 16:02:17 crc kubenswrapper[4716]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 07 16:02:17 crc kubenswrapper[4716]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Dec 07 16:02:17 crc kubenswrapper[4716]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 07 16:02:17 crc kubenswrapper[4716]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 07 16:02:17 crc kubenswrapper[4716]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 07 16:02:17 crc kubenswrapper[4716]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.421299 4716 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.424659 4716 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.424682 4716 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.424687 4716 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.424693 4716 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.424699 4716 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.424705 4716 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.424710 4716 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.424714 4716 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.424718 4716 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.424723 4716 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.424728 4716 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.424732 4716 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.424738 4716 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.424743 4716 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.424748 4716 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.424752 4716 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.424756 4716 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.424760 4716 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.424775 4716 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.424780 4716 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.424784 4716 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.424790 4716 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.424796 4716 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.424801 4716 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.424806 4716 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.424811 4716 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.424816 4716 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.424820 4716 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.424825 4716 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.424830 4716 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.424834 4716 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.424839 4716 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.424844 4716 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.424850 4716 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.424855 4716 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.424861 4716 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.424869 4716 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.424876 4716 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.424881 4716 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.424887 4716 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.424892 4716 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.424898 4716 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.424902 4716 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.424907 4716 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.424912 4716 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.424917 4716 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.424921 4716 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.424926 4716 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.424932 4716 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.424938 4716 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.424943 4716 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.424947 4716 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.424952 4716 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.424957 4716 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.424963 4716 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.424969 4716 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.424973 4716 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.424978 4716 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.424982 4716 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.424986 4716 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.424990 4716 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.424995 4716 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.425000 4716 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.425004 4716 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.425008 4716 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.425013 4716 feature_gate.go:330] unrecognized feature gate: Example Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.425017 4716 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.425024 4716 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.425028 4716 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.425032 4716 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.425037 4716 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.425881 4716 flags.go:64] FLAG: --address="0.0.0.0" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.425896 4716 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.425906 4716 flags.go:64] FLAG: --anonymous-auth="true" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.425912 4716 flags.go:64] FLAG: --application-metrics-count-limit="100" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.425918 4716 flags.go:64] FLAG: --authentication-token-webhook="false" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.425923 4716 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.425930 4716 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.425937 4716 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.425941 4716 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.425946 4716 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.425951 4716 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.425957 4716 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.425962 4716 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.425967 4716 flags.go:64] FLAG: --cgroup-root="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.425972 4716 flags.go:64] FLAG: --cgroups-per-qos="true" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.425977 4716 flags.go:64] FLAG: --client-ca-file="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.425983 4716 flags.go:64] FLAG: --cloud-config="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.425989 4716 flags.go:64] FLAG: --cloud-provider="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.425993 4716 flags.go:64] FLAG: --cluster-dns="[]" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426002 4716 flags.go:64] FLAG: --cluster-domain="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426007 4716 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426013 4716 flags.go:64] FLAG: --config-dir="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426017 4716 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426022 4716 flags.go:64] FLAG: --container-log-max-files="5" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426029 4716 flags.go:64] FLAG: --container-log-max-size="10Mi" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426034 4716 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426039 4716 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426044 4716 flags.go:64] FLAG: --containerd-namespace="k8s.io" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426049 4716 flags.go:64] FLAG: --contention-profiling="false" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426054 4716 flags.go:64] FLAG: --cpu-cfs-quota="true" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426058 4716 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426064 4716 flags.go:64] FLAG: --cpu-manager-policy="none" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426069 4716 flags.go:64] FLAG: --cpu-manager-policy-options="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426091 4716 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426096 4716 flags.go:64] FLAG: --enable-controller-attach-detach="true" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426101 4716 flags.go:64] FLAG: --enable-debugging-handlers="true" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426106 4716 flags.go:64] FLAG: --enable-load-reader="false" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426111 4716 flags.go:64] FLAG: --enable-server="true" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426116 4716 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426122 4716 flags.go:64] FLAG: --event-burst="100" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426128 4716 flags.go:64] FLAG: --event-qps="50" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426133 4716 flags.go:64] FLAG: --event-storage-age-limit="default=0" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426138 4716 flags.go:64] FLAG: --event-storage-event-limit="default=0" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426143 4716 flags.go:64] FLAG: --eviction-hard="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426150 4716 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426155 4716 flags.go:64] FLAG: --eviction-minimum-reclaim="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426160 4716 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426165 4716 flags.go:64] FLAG: --eviction-soft="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426173 4716 flags.go:64] FLAG: --eviction-soft-grace-period="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426178 4716 flags.go:64] FLAG: --exit-on-lock-contention="false" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426183 4716 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426188 4716 flags.go:64] FLAG: --experimental-mounter-path="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426194 4716 flags.go:64] FLAG: --fail-cgroupv1="false" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426199 4716 flags.go:64] FLAG: --fail-swap-on="true" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426204 4716 flags.go:64] FLAG: --feature-gates="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426211 4716 flags.go:64] FLAG: --file-check-frequency="20s" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426217 4716 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426222 4716 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426227 4716 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426233 4716 flags.go:64] FLAG: --healthz-port="10248" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426238 4716 flags.go:64] FLAG: --help="false" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426243 4716 flags.go:64] FLAG: --hostname-override="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426248 4716 flags.go:64] FLAG: --housekeeping-interval="10s" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426253 4716 flags.go:64] FLAG: --http-check-frequency="20s" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426258 4716 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426264 4716 flags.go:64] FLAG: --image-credential-provider-config="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426270 4716 flags.go:64] FLAG: --image-gc-high-threshold="85" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426274 4716 flags.go:64] FLAG: --image-gc-low-threshold="80" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426281 4716 flags.go:64] FLAG: --image-service-endpoint="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426286 4716 flags.go:64] FLAG: --kernel-memcg-notification="false" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426291 4716 flags.go:64] FLAG: --kube-api-burst="100" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426296 4716 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426301 4716 flags.go:64] FLAG: --kube-api-qps="50" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426306 4716 flags.go:64] FLAG: --kube-reserved="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426311 4716 flags.go:64] FLAG: --kube-reserved-cgroup="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426316 4716 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426321 4716 flags.go:64] FLAG: --kubelet-cgroups="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426325 4716 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426330 4716 flags.go:64] FLAG: --lock-file="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426335 4716 flags.go:64] FLAG: --log-cadvisor-usage="false" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426341 4716 flags.go:64] FLAG: --log-flush-frequency="5s" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426346 4716 flags.go:64] FLAG: --log-json-info-buffer-size="0" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426354 4716 flags.go:64] FLAG: --log-json-split-stream="false" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426359 4716 flags.go:64] FLAG: --log-text-info-buffer-size="0" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426364 4716 flags.go:64] FLAG: --log-text-split-stream="false" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426371 4716 flags.go:64] FLAG: --logging-format="text" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426376 4716 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426381 4716 flags.go:64] FLAG: --make-iptables-util-chains="true" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426386 4716 flags.go:64] FLAG: --manifest-url="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426391 4716 flags.go:64] FLAG: --manifest-url-header="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426397 4716 flags.go:64] FLAG: --max-housekeeping-interval="15s" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426402 4716 flags.go:64] FLAG: --max-open-files="1000000" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426408 4716 flags.go:64] FLAG: --max-pods="110" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426413 4716 flags.go:64] FLAG: --maximum-dead-containers="-1" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426418 4716 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426423 4716 flags.go:64] FLAG: --memory-manager-policy="None" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426428 4716 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426432 4716 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426437 4716 flags.go:64] FLAG: --node-ip="192.168.126.11" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426442 4716 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426456 4716 flags.go:64] FLAG: --node-status-max-images="50" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426461 4716 flags.go:64] FLAG: --node-status-update-frequency="10s" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426466 4716 flags.go:64] FLAG: --oom-score-adj="-999" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426472 4716 flags.go:64] FLAG: --pod-cidr="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426479 4716 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426488 4716 flags.go:64] FLAG: --pod-manifest-path="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426494 4716 flags.go:64] FLAG: --pod-max-pids="-1" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426500 4716 flags.go:64] FLAG: --pods-per-core="0" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426506 4716 flags.go:64] FLAG: --port="10250" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426511 4716 flags.go:64] FLAG: --protect-kernel-defaults="false" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426516 4716 flags.go:64] FLAG: --provider-id="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426521 4716 flags.go:64] FLAG: --qos-reserved="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426528 4716 flags.go:64] FLAG: --read-only-port="10255" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426534 4716 flags.go:64] FLAG: --register-node="true" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426539 4716 flags.go:64] FLAG: --register-schedulable="true" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426544 4716 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426553 4716 flags.go:64] FLAG: --registry-burst="10" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426558 4716 flags.go:64] FLAG: --registry-qps="5" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426562 4716 flags.go:64] FLAG: --reserved-cpus="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426568 4716 flags.go:64] FLAG: --reserved-memory="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426792 4716 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426804 4716 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426809 4716 flags.go:64] FLAG: --rotate-certificates="false" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426815 4716 flags.go:64] FLAG: --rotate-server-certificates="false" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426820 4716 flags.go:64] FLAG: --runonce="false" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426825 4716 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426831 4716 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426838 4716 flags.go:64] FLAG: --seccomp-default="false" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426843 4716 flags.go:64] FLAG: --serialize-image-pulls="true" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426848 4716 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426854 4716 flags.go:64] FLAG: --storage-driver-db="cadvisor" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426860 4716 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426865 4716 flags.go:64] FLAG: --storage-driver-password="root" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426870 4716 flags.go:64] FLAG: --storage-driver-secure="false" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426876 4716 flags.go:64] FLAG: --storage-driver-table="stats" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426880 4716 flags.go:64] FLAG: --storage-driver-user="root" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426885 4716 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426891 4716 flags.go:64] FLAG: --sync-frequency="1m0s" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426896 4716 flags.go:64] FLAG: --system-cgroups="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426900 4716 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426910 4716 flags.go:64] FLAG: --system-reserved-cgroup="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426915 4716 flags.go:64] FLAG: --tls-cert-file="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426921 4716 flags.go:64] FLAG: --tls-cipher-suites="[]" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426928 4716 flags.go:64] FLAG: --tls-min-version="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426934 4716 flags.go:64] FLAG: --tls-private-key-file="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426939 4716 flags.go:64] FLAG: --topology-manager-policy="none" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426968 4716 flags.go:64] FLAG: --topology-manager-policy-options="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426973 4716 flags.go:64] FLAG: --topology-manager-scope="container" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426979 4716 flags.go:64] FLAG: --v="2" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426987 4716 flags.go:64] FLAG: --version="false" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.426993 4716 flags.go:64] FLAG: --vmodule="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.427002 4716 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.427007 4716 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.427178 4716 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.427187 4716 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.427192 4716 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.427197 4716 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.427202 4716 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.427207 4716 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.427214 4716 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.427219 4716 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.427224 4716 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.427230 4716 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.427236 4716 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.427241 4716 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.427246 4716 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.427252 4716 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.427258 4716 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.427263 4716 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.427269 4716 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.427275 4716 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.427280 4716 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.427285 4716 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.427289 4716 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.427295 4716 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.427299 4716 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.427305 4716 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.427311 4716 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.427315 4716 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.427320 4716 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.427324 4716 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.427329 4716 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.427333 4716 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.427337 4716 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.427342 4716 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.427346 4716 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.427350 4716 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.427354 4716 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.427358 4716 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.427362 4716 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.427367 4716 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.427371 4716 feature_gate.go:330] unrecognized feature gate: Example Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.427375 4716 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.427379 4716 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.427384 4716 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.427388 4716 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.427392 4716 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.427396 4716 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.427401 4716 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.427405 4716 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.427410 4716 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.427414 4716 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.427418 4716 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.427422 4716 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.427427 4716 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.427432 4716 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.427438 4716 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.427442 4716 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.427447 4716 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.427452 4716 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.427456 4716 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.427460 4716 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.427465 4716 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.427469 4716 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.427473 4716 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.427478 4716 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.427482 4716 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.427487 4716 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.427491 4716 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.427495 4716 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.427500 4716 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.427504 4716 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.427508 4716 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.427512 4716 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.427529 4716 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.437296 4716 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.437358 4716 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.437488 4716 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.437514 4716 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.437523 4716 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.437534 4716 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.437544 4716 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.437552 4716 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.437561 4716 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.437569 4716 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.437577 4716 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.437584 4716 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.437592 4716 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.437600 4716 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.437611 4716 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.437626 4716 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.437635 4716 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.437644 4716 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.437654 4716 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.437663 4716 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.437672 4716 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.437681 4716 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.437690 4716 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.437699 4716 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.437707 4716 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.437715 4716 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.437723 4716 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.437731 4716 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.437739 4716 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.437747 4716 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.437758 4716 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.437768 4716 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.437777 4716 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.437787 4716 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.437796 4716 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.437805 4716 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.437816 4716 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.437825 4716 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.437833 4716 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.437844 4716 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.437853 4716 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.437862 4716 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.437871 4716 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.437879 4716 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.437887 4716 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.437896 4716 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.437903 4716 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.437911 4716 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.437919 4716 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.437927 4716 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.437935 4716 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.437943 4716 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.437951 4716 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.437959 4716 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.437966 4716 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.437975 4716 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.437983 4716 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.437990 4716 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.437997 4716 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438008 4716 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438016 4716 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438024 4716 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438032 4716 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438040 4716 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438049 4716 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438058 4716 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438066 4716 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438073 4716 feature_gate.go:330] unrecognized feature gate: Example Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438107 4716 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438114 4716 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438122 4716 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438130 4716 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438139 4716 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.438154 4716 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438399 4716 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438414 4716 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438425 4716 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438434 4716 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438444 4716 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438453 4716 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438462 4716 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438470 4716 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438478 4716 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438490 4716 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438498 4716 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438506 4716 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438557 4716 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438568 4716 feature_gate.go:330] unrecognized feature gate: Example Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438577 4716 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438586 4716 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438597 4716 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438606 4716 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438616 4716 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438625 4716 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438634 4716 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438642 4716 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438650 4716 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438658 4716 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438667 4716 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438676 4716 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438685 4716 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438693 4716 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438701 4716 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438709 4716 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438717 4716 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438724 4716 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438732 4716 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438740 4716 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438749 4716 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438756 4716 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438764 4716 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438772 4716 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438779 4716 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438789 4716 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438797 4716 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438806 4716 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438814 4716 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438822 4716 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438830 4716 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438838 4716 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438848 4716 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438858 4716 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438866 4716 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438874 4716 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438882 4716 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438890 4716 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438898 4716 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438906 4716 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438914 4716 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438921 4716 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438929 4716 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438936 4716 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438961 4716 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438969 4716 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438977 4716 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438984 4716 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.438992 4716 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.439000 4716 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.439008 4716 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.439019 4716 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.439028 4716 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.439036 4716 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.439044 4716 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.439053 4716 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.439062 4716 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.439074 4716 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.439685 4716 server.go:940] "Client rotation is on, will bootstrap in background" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.444469 4716 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.444632 4716 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.445343 4716 server.go:997] "Starting client certificate rotation" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.445393 4716 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.445731 4716 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-11-16 14:55:56.254558983 +0000 UTC Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.445854 4716 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.461716 4716 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 07 16:02:17 crc kubenswrapper[4716]: E1207 16:02:17.463591 4716 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.18:6443: connect: connection refused" logger="UnhandledError" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.464502 4716 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.478410 4716 log.go:25] "Validated CRI v1 runtime API" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.500571 4716 log.go:25] "Validated CRI v1 image API" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.503354 4716 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.507294 4716 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-12-07-15-58-11-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.507360 4716 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.543310 4716 manager.go:217] Machine: {Timestamp:2025-12-07 16:02:17.539997756 +0000 UTC m=+0.230282748 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654124544 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:a57d09b0-0224-46eb-a3ac-a2b60dad2250 BootID:a8d1b7d4-7ec4-40c7-86b0-291db1f38894 Filesystems:[{Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108169 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827060224 Type:vfs Inodes:4108169 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:26:97:b3 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:26:97:b3 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:c5:0a:61 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:f7:fa:a1 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:c7:45:d0 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:8c:87:b6 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:52:f9:45:ce:8b:23 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:0a:ae:bc:1e:78:8e Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654124544 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.543759 4716 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.544215 4716 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.545002 4716 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.545426 4716 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.545509 4716 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.545890 4716 topology_manager.go:138] "Creating topology manager with none policy" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.545910 4716 container_manager_linux.go:303] "Creating device plugin manager" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.546250 4716 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.546333 4716 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.546781 4716 state_mem.go:36] "Initialized new in-memory state store" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.547476 4716 server.go:1245] "Using root directory" path="/var/lib/kubelet" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.548401 4716 kubelet.go:418] "Attempting to sync node with API server" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.548434 4716 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.548474 4716 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.548496 4716 kubelet.go:324] "Adding apiserver pod source" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.548583 4716 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.551134 4716 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.18:6443: connect: connection refused Dec 07 16:02:17 crc kubenswrapper[4716]: E1207 16:02:17.551327 4716 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.18:6443: connect: connection refused" logger="UnhandledError" Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.551317 4716 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.18:6443: connect: connection refused Dec 07 16:02:17 crc kubenswrapper[4716]: E1207 16:02:17.551440 4716 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.18:6443: connect: connection refused" logger="UnhandledError" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.552331 4716 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.552757 4716 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.553533 4716 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.554332 4716 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.554376 4716 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.554390 4716 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.554405 4716 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.554426 4716 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.554440 4716 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.554453 4716 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.554477 4716 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.554491 4716 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.554505 4716 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.554544 4716 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.554558 4716 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.555255 4716 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.556001 4716 server.go:1280] "Started kubelet" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.556163 4716 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.18:6443: connect: connection refused Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.556660 4716 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.556654 4716 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.557922 4716 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 07 16:02:17 crc kubenswrapper[4716]: E1207 16:02:17.558380 4716 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.18:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187efa52cfc65866 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-07 16:02:17.555957862 +0000 UTC m=+0.246242804,LastTimestamp:2025-12-07 16:02:17.555957862 +0000 UTC m=+0.246242804,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 07 16:02:17 crc systemd[1]: Started Kubernetes Kubelet. Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.559938 4716 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.559969 4716 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.560387 4716 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-17 05:10:47.107288959 +0000 UTC Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.560465 4716 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 973h8m29.546831079s for next certificate rotation Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.560472 4716 server.go:460] "Adding debug handlers to kubelet server" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.560559 4716 volume_manager.go:287] "The desired_state_of_world populator starts" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.560584 4716 volume_manager.go:289] "Starting Kubelet Volume Manager" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.561911 4716 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Dec 07 16:02:17 crc kubenswrapper[4716]: E1207 16:02:17.562327 4716 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 07 16:02:17 crc kubenswrapper[4716]: E1207 16:02:17.563322 4716 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.18:6443: connect: connection refused" interval="200ms" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.563768 4716 factory.go:55] Registering systemd factory Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.563824 4716 factory.go:221] Registration of the systemd container factory successfully Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.565377 4716 factory.go:153] Registering CRI-O factory Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.565488 4716 factory.go:221] Registration of the crio container factory successfully Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.565696 4716 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.565761 4716 factory.go:103] Registering Raw factory Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.565790 4716 manager.go:1196] Started watching for new ooms in manager Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.565857 4716 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.18:6443: connect: connection refused Dec 07 16:02:17 crc kubenswrapper[4716]: E1207 16:02:17.565996 4716 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.18:6443: connect: connection refused" logger="UnhandledError" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.567274 4716 manager.go:319] Starting recovery of all containers Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.577230 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.577281 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.577295 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.577305 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.577315 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.577323 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.577332 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.577344 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.577354 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.577365 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.577374 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.577384 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.577396 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.577407 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.577417 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.577427 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.577437 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.577447 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.577462 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.577472 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.577484 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.577496 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.577507 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.577517 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.577527 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.577536 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.577549 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.577560 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.577570 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.577593 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.577606 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.577618 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.577628 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.577663 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.577674 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.577686 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.577696 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.577708 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.577719 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.577729 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.577771 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.577782 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.577793 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.577822 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.577832 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.577844 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.577856 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.577866 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.577877 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.577888 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.577898 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.577909 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.577924 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.577935 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.577946 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.577957 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.577969 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.577979 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.577991 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.578000 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.578011 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.578021 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.578031 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.578042 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.578052 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.578063 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.578153 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.578166 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.578176 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.578186 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.578197 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.578207 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.578220 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.578230 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.578240 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.578251 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.578263 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.578273 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.578284 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.578294 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.578305 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.578314 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.578324 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.578334 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.578344 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.578354 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.578363 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.578373 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.578383 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.578393 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.578403 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.578414 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.578424 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.580235 4716 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.580310 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.580342 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.580364 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.580388 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.580411 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.580434 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.580455 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.580497 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.580518 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.580540 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.580560 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.580594 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.580619 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.580645 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.580670 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.580693 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.580716 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.580740 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.580765 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.580789 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.580813 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.580833 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.580855 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.580876 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.580897 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.580918 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.580943 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.580968 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.580990 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.581011 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.581032 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.581060 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.581136 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.581159 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.581182 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.581212 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.581234 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.581255 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.581277 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.581299 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.581321 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.581342 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.581364 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.581387 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.581410 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.581434 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.581457 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.581478 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.581502 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.581521 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.581545 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.581567 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.581590 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.581611 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.581633 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.581653 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.581676 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.581696 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.581717 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.581737 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.581757 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.581781 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.581802 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.581822 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.581847 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.581870 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.581889 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.581913 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.581934 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.581958 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.581981 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.582004 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.582025 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.582047 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.582070 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.582115 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.582142 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.582166 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.582188 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.582208 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.582233 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.582255 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.582277 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.582298 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.582319 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.582340 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.582360 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.582381 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.582403 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.582425 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.582445 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.582466 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.582488 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.582510 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.582532 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.582552 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.582576 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.582596 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.582618 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.582642 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.582662 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.582685 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.582705 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.582729 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.582751 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.582771 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.582794 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.582814 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.582835 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.582856 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.582879 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.582898 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.582954 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.582976 4716 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.582995 4716 reconstruct.go:97] "Volume reconstruction finished" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.583010 4716 reconciler.go:26] "Reconciler: start to sync state" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.592807 4716 manager.go:324] Recovery completed Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.602531 4716 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.604030 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.604063 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.604087 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.604638 4716 cpu_manager.go:225] "Starting CPU manager" policy="none" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.604652 4716 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.604669 4716 state_mem.go:36] "Initialized new in-memory state store" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.651841 4716 policy_none.go:49] "None policy: Start" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.652713 4716 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.656148 4716 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.656192 4716 state_mem.go:35] "Initializing new in-memory state store" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.656356 4716 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.656414 4716 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.656438 4716 kubelet.go:2335] "Starting kubelet main sync loop" Dec 07 16:02:17 crc kubenswrapper[4716]: E1207 16:02:17.656491 4716 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 07 16:02:17 crc kubenswrapper[4716]: W1207 16:02:17.657800 4716 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.18:6443: connect: connection refused Dec 07 16:02:17 crc kubenswrapper[4716]: E1207 16:02:17.657887 4716 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.18:6443: connect: connection refused" logger="UnhandledError" Dec 07 16:02:17 crc kubenswrapper[4716]: E1207 16:02:17.663202 4716 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.723850 4716 manager.go:334] "Starting Device Plugin manager" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.723921 4716 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.723938 4716 server.go:79] "Starting device plugin registration server" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.724493 4716 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.724519 4716 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.724772 4716 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.724862 4716 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.724879 4716 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 07 16:02:17 crc kubenswrapper[4716]: E1207 16:02:17.734434 4716 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.757323 4716 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc"] Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.757523 4716 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.758536 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.758574 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.758587 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.758715 4716 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.759240 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.759312 4716 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.759346 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.759370 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.759380 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.759464 4716 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.759580 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.759612 4716 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.760451 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.760472 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.760486 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.760494 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.760508 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.760568 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.760581 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.760590 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.760684 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.760767 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.760785 4716 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.760796 4716 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.761542 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.761562 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.761571 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.761549 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.761638 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.761649 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.761656 4716 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.761802 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.761840 4716 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.762348 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.762395 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.762407 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.762501 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.762532 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.762545 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.762604 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.762639 4716 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.763396 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.763419 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.763428 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:17 crc kubenswrapper[4716]: E1207 16:02:17.763836 4716 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.18:6443: connect: connection refused" interval="400ms" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.785155 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.785218 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.785327 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.785426 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.785471 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.785501 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.785552 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.785605 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.785639 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.785671 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.785691 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.785713 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.785738 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.785759 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.785789 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.826583 4716 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.827617 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.827668 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.827678 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.827711 4716 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 07 16:02:17 crc kubenswrapper[4716]: E1207 16:02:17.828331 4716 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.18:6443: connect: connection refused" node="crc" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.887169 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.887224 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.887248 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.887265 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.887282 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.887299 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.887315 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.887332 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.887347 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.887346 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.887370 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.887387 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.887372 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.887406 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.887411 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.887411 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.887449 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.887478 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.887433 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.887439 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.887346 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.887460 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.887463 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.887422 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.887556 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.887571 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.887394 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.887468 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.887632 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.887712 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.893180 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 16:02:17 crc kubenswrapper[4716]: I1207 16:02:17.895265 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 07 16:02:18 crc kubenswrapper[4716]: I1207 16:02:18.029147 4716 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 16:02:18 crc kubenswrapper[4716]: I1207 16:02:18.030999 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:18 crc kubenswrapper[4716]: I1207 16:02:18.031057 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:18 crc kubenswrapper[4716]: I1207 16:02:18.031067 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:18 crc kubenswrapper[4716]: I1207 16:02:18.031124 4716 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 07 16:02:18 crc kubenswrapper[4716]: E1207 16:02:18.031751 4716 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.18:6443: connect: connection refused" node="crc" Dec 07 16:02:18 crc kubenswrapper[4716]: I1207 16:02:18.115262 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 07 16:02:18 crc kubenswrapper[4716]: W1207 16:02:18.129630 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-95f4b6effe1913dda2cbd46680583b974f7c732b53990f0259a0dc7e44f73ac5 WatchSource:0}: Error finding container 95f4b6effe1913dda2cbd46680583b974f7c732b53990f0259a0dc7e44f73ac5: Status 404 returned error can't find the container with id 95f4b6effe1913dda2cbd46680583b974f7c732b53990f0259a0dc7e44f73ac5 Dec 07 16:02:18 crc kubenswrapper[4716]: I1207 16:02:18.145014 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 07 16:02:18 crc kubenswrapper[4716]: I1207 16:02:18.156100 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 07 16:02:18 crc kubenswrapper[4716]: W1207 16:02:18.157053 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-a745ba7defcb95b5611c76ccf86a480c8a7d36b8c4579bed6d8aa682a841cf0f WatchSource:0}: Error finding container a745ba7defcb95b5611c76ccf86a480c8a7d36b8c4579bed6d8aa682a841cf0f: Status 404 returned error can't find the container with id a745ba7defcb95b5611c76ccf86a480c8a7d36b8c4579bed6d8aa682a841cf0f Dec 07 16:02:18 crc kubenswrapper[4716]: E1207 16:02:18.165253 4716 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.18:6443: connect: connection refused" interval="800ms" Dec 07 16:02:18 crc kubenswrapper[4716]: W1207 16:02:18.174816 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-7a46a5e2abdc85e83aca0bcfdb8b082f20151502d63a5b36461d6ec1ad04f155 WatchSource:0}: Error finding container 7a46a5e2abdc85e83aca0bcfdb8b082f20151502d63a5b36461d6ec1ad04f155: Status 404 returned error can't find the container with id 7a46a5e2abdc85e83aca0bcfdb8b082f20151502d63a5b36461d6ec1ad04f155 Dec 07 16:02:18 crc kubenswrapper[4716]: I1207 16:02:18.432641 4716 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 16:02:18 crc kubenswrapper[4716]: I1207 16:02:18.435057 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:18 crc kubenswrapper[4716]: I1207 16:02:18.435130 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:18 crc kubenswrapper[4716]: I1207 16:02:18.435143 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:18 crc kubenswrapper[4716]: I1207 16:02:18.435190 4716 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 07 16:02:18 crc kubenswrapper[4716]: E1207 16:02:18.435848 4716 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.18:6443: connect: connection refused" node="crc" Dec 07 16:02:18 crc kubenswrapper[4716]: I1207 16:02:18.557459 4716 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.18:6443: connect: connection refused Dec 07 16:02:18 crc kubenswrapper[4716]: I1207 16:02:18.663753 4716 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="ad8df9b5bd5f5a8dd44ce2899cfa29c19435149b170d339d339183cd3df09e99" exitCode=0 Dec 07 16:02:18 crc kubenswrapper[4716]: I1207 16:02:18.663826 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"ad8df9b5bd5f5a8dd44ce2899cfa29c19435149b170d339d339183cd3df09e99"} Dec 07 16:02:18 crc kubenswrapper[4716]: I1207 16:02:18.663917 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"95f4b6effe1913dda2cbd46680583b974f7c732b53990f0259a0dc7e44f73ac5"} Dec 07 16:02:18 crc kubenswrapper[4716]: I1207 16:02:18.664005 4716 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 16:02:18 crc kubenswrapper[4716]: I1207 16:02:18.664801 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:18 crc kubenswrapper[4716]: I1207 16:02:18.664859 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:18 crc kubenswrapper[4716]: I1207 16:02:18.664879 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:18 crc kubenswrapper[4716]: I1207 16:02:18.665704 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"d14d018fa044e9e80fe81e7e38df935992d14c773ed2b2193fd5cb873e5d0947"} Dec 07 16:02:18 crc kubenswrapper[4716]: I1207 16:02:18.665772 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"6a9e9cf048220ca14c996847e9ced9f2367b9331dc6c2836dd023658855cd79d"} Dec 07 16:02:18 crc kubenswrapper[4716]: I1207 16:02:18.665789 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"f84f2728d5774884910b8709b3b30e72667cc4d4a62eaceb8a96cde1e857aa9a"} Dec 07 16:02:18 crc kubenswrapper[4716]: I1207 16:02:18.667656 4716 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a" exitCode=0 Dec 07 16:02:18 crc kubenswrapper[4716]: I1207 16:02:18.667683 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a"} Dec 07 16:02:18 crc kubenswrapper[4716]: I1207 16:02:18.667725 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"996837b00ff2fc88a6762ad4b7b8e00cf6d26ea0988857b096ff082ed96e2ffa"} Dec 07 16:02:18 crc kubenswrapper[4716]: I1207 16:02:18.667846 4716 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 16:02:18 crc kubenswrapper[4716]: I1207 16:02:18.672703 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:18 crc kubenswrapper[4716]: I1207 16:02:18.672741 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:18 crc kubenswrapper[4716]: I1207 16:02:18.672753 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:18 crc kubenswrapper[4716]: I1207 16:02:18.673568 4716 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="09b38cd4b389ea477c802a842ee4f2297a64d64684624117163e819539c257c6" exitCode=0 Dec 07 16:02:18 crc kubenswrapper[4716]: I1207 16:02:18.673659 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"09b38cd4b389ea477c802a842ee4f2297a64d64684624117163e819539c257c6"} Dec 07 16:02:18 crc kubenswrapper[4716]: I1207 16:02:18.673722 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"7a46a5e2abdc85e83aca0bcfdb8b082f20151502d63a5b36461d6ec1ad04f155"} Dec 07 16:02:18 crc kubenswrapper[4716]: I1207 16:02:18.674134 4716 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 16:02:18 crc kubenswrapper[4716]: I1207 16:02:18.675354 4716 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="e3d77f99006bada2a3a676398c5ddb366c0c727f3837f10674a9d582dd83e2cd" exitCode=0 Dec 07 16:02:18 crc kubenswrapper[4716]: I1207 16:02:18.675394 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"e3d77f99006bada2a3a676398c5ddb366c0c727f3837f10674a9d582dd83e2cd"} Dec 07 16:02:18 crc kubenswrapper[4716]: I1207 16:02:18.675549 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"a745ba7defcb95b5611c76ccf86a480c8a7d36b8c4579bed6d8aa682a841cf0f"} Dec 07 16:02:18 crc kubenswrapper[4716]: I1207 16:02:18.675592 4716 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 16:02:18 crc kubenswrapper[4716]: I1207 16:02:18.675637 4716 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 16:02:18 crc kubenswrapper[4716]: I1207 16:02:18.676777 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:18 crc kubenswrapper[4716]: I1207 16:02:18.676940 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:18 crc kubenswrapper[4716]: I1207 16:02:18.676966 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:18 crc kubenswrapper[4716]: I1207 16:02:18.681862 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:18 crc kubenswrapper[4716]: I1207 16:02:18.681889 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:18 crc kubenswrapper[4716]: I1207 16:02:18.681923 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:18 crc kubenswrapper[4716]: I1207 16:02:18.682029 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:18 crc kubenswrapper[4716]: I1207 16:02:18.683070 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:18 crc kubenswrapper[4716]: I1207 16:02:18.683121 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:18 crc kubenswrapper[4716]: W1207 16:02:18.863298 4716 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.18:6443: connect: connection refused Dec 07 16:02:18 crc kubenswrapper[4716]: E1207 16:02:18.863362 4716 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.18:6443: connect: connection refused" logger="UnhandledError" Dec 07 16:02:18 crc kubenswrapper[4716]: E1207 16:02:18.966798 4716 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.18:6443: connect: connection refused" interval="1.6s" Dec 07 16:02:19 crc kubenswrapper[4716]: W1207 16:02:19.038351 4716 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.18:6443: connect: connection refused Dec 07 16:02:19 crc kubenswrapper[4716]: E1207 16:02:19.038419 4716 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.18:6443: connect: connection refused" logger="UnhandledError" Dec 07 16:02:19 crc kubenswrapper[4716]: W1207 16:02:19.080973 4716 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.18:6443: connect: connection refused Dec 07 16:02:19 crc kubenswrapper[4716]: E1207 16:02:19.081043 4716 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.18:6443: connect: connection refused" logger="UnhandledError" Dec 07 16:02:19 crc kubenswrapper[4716]: W1207 16:02:19.135588 4716 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.18:6443: connect: connection refused Dec 07 16:02:19 crc kubenswrapper[4716]: E1207 16:02:19.135729 4716 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.18:6443: connect: connection refused" logger="UnhandledError" Dec 07 16:02:19 crc kubenswrapper[4716]: I1207 16:02:19.236428 4716 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 16:02:19 crc kubenswrapper[4716]: I1207 16:02:19.237575 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:19 crc kubenswrapper[4716]: I1207 16:02:19.237615 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:19 crc kubenswrapper[4716]: I1207 16:02:19.237625 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:19 crc kubenswrapper[4716]: I1207 16:02:19.237651 4716 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 07 16:02:19 crc kubenswrapper[4716]: I1207 16:02:19.571678 4716 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 07 16:02:19 crc kubenswrapper[4716]: I1207 16:02:19.678887 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"a1ab9067e4d9988fd94f93d7798954e4c8888875f47244d797cb6b0557ce5582"} Dec 07 16:02:19 crc kubenswrapper[4716]: I1207 16:02:19.679016 4716 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 16:02:19 crc kubenswrapper[4716]: I1207 16:02:19.679769 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:19 crc kubenswrapper[4716]: I1207 16:02:19.679798 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:19 crc kubenswrapper[4716]: I1207 16:02:19.679807 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:19 crc kubenswrapper[4716]: I1207 16:02:19.681060 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"5d8c9a434c6cfb9e4fee6a8955f1379291f153c007876d1ffb498ce561c73ba8"} Dec 07 16:02:19 crc kubenswrapper[4716]: I1207 16:02:19.681122 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"c1358745250b35426602fc7417e4f85dd411cdd0fb3e8f703c432d72128ca7c5"} Dec 07 16:02:19 crc kubenswrapper[4716]: I1207 16:02:19.681136 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"1ff9c456318bbdb54ec3f1ac3e2c2366968f18ad084158b7100d142acb668948"} Dec 07 16:02:19 crc kubenswrapper[4716]: I1207 16:02:19.681189 4716 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 16:02:19 crc kubenswrapper[4716]: I1207 16:02:19.681700 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:19 crc kubenswrapper[4716]: I1207 16:02:19.681728 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:19 crc kubenswrapper[4716]: I1207 16:02:19.681739 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:19 crc kubenswrapper[4716]: I1207 16:02:19.683885 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"9ac6ce7701f1c6aa1ea8fcf7a3035e3b0eaa28ae3cb6afc442372bd9117bdde4"} Dec 07 16:02:19 crc kubenswrapper[4716]: I1207 16:02:19.683915 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"d6d3c911874095b3c7cb062d03070e5090a7bb2456cb20cb4ce078f502515cda"} Dec 07 16:02:19 crc kubenswrapper[4716]: I1207 16:02:19.683939 4716 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 16:02:19 crc kubenswrapper[4716]: I1207 16:02:19.684710 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:19 crc kubenswrapper[4716]: I1207 16:02:19.684736 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:19 crc kubenswrapper[4716]: I1207 16:02:19.684747 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:19 crc kubenswrapper[4716]: I1207 16:02:19.686274 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"0d31de0326a05216fa10442bdb47151c6e073c97e9a16c34e793028e5e253756"} Dec 07 16:02:19 crc kubenswrapper[4716]: I1207 16:02:19.686309 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"27bd4a54f920b2725987c1c7fd2d9fdf16cd08738c246c333a4e9164a3a582a0"} Dec 07 16:02:19 crc kubenswrapper[4716]: I1207 16:02:19.686321 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"6276df652efe6ca935e745f24ef20f87a16465475139230c63b9999234a4d12a"} Dec 07 16:02:19 crc kubenswrapper[4716]: I1207 16:02:19.686329 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"e9c3b1dd202586ee755ca89fbe12b30bbb12f95a62514f6e4d93cbbdbe55181e"} Dec 07 16:02:19 crc kubenswrapper[4716]: I1207 16:02:19.686337 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"73f85248b4d68a9beecdacb3a01f433e97a0ffea2247ce92a41ef96855509377"} Dec 07 16:02:19 crc kubenswrapper[4716]: I1207 16:02:19.686401 4716 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 16:02:19 crc kubenswrapper[4716]: I1207 16:02:19.686912 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:19 crc kubenswrapper[4716]: I1207 16:02:19.686931 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:19 crc kubenswrapper[4716]: I1207 16:02:19.686941 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:19 crc kubenswrapper[4716]: I1207 16:02:19.688504 4716 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="c00f7a8e024e0e0dfd7b73e16d1d5437602e79ea95b73c21f24b209edef60e98" exitCode=0 Dec 07 16:02:19 crc kubenswrapper[4716]: I1207 16:02:19.688540 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"c00f7a8e024e0e0dfd7b73e16d1d5437602e79ea95b73c21f24b209edef60e98"} Dec 07 16:02:19 crc kubenswrapper[4716]: I1207 16:02:19.688649 4716 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 16:02:19 crc kubenswrapper[4716]: I1207 16:02:19.689327 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:19 crc kubenswrapper[4716]: I1207 16:02:19.689365 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:19 crc kubenswrapper[4716]: I1207 16:02:19.689376 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:20 crc kubenswrapper[4716]: I1207 16:02:20.647681 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 07 16:02:20 crc kubenswrapper[4716]: I1207 16:02:20.695115 4716 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="c3f6392d426fa2cd788ec3a5c3bff4ac39483a54dfb2624a1c4370522413a350" exitCode=0 Dec 07 16:02:20 crc kubenswrapper[4716]: I1207 16:02:20.695120 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"c3f6392d426fa2cd788ec3a5c3bff4ac39483a54dfb2624a1c4370522413a350"} Dec 07 16:02:20 crc kubenswrapper[4716]: I1207 16:02:20.695302 4716 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 07 16:02:20 crc kubenswrapper[4716]: I1207 16:02:20.695370 4716 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 16:02:20 crc kubenswrapper[4716]: I1207 16:02:20.695370 4716 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 16:02:20 crc kubenswrapper[4716]: I1207 16:02:20.695379 4716 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 16:02:20 crc kubenswrapper[4716]: I1207 16:02:20.696991 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:20 crc kubenswrapper[4716]: I1207 16:02:20.697051 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:20 crc kubenswrapper[4716]: I1207 16:02:20.697053 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:20 crc kubenswrapper[4716]: I1207 16:02:20.697069 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:20 crc kubenswrapper[4716]: I1207 16:02:20.697110 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:20 crc kubenswrapper[4716]: I1207 16:02:20.697140 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:20 crc kubenswrapper[4716]: I1207 16:02:20.697156 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:20 crc kubenswrapper[4716]: I1207 16:02:20.697167 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:20 crc kubenswrapper[4716]: I1207 16:02:20.697174 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:21 crc kubenswrapper[4716]: I1207 16:02:21.704996 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"54fd3c2941782921f99a6cdc3de4e892234548bde527fec15a8ac92e513566e4"} Dec 07 16:02:21 crc kubenswrapper[4716]: I1207 16:02:21.705063 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"404426a171b8676eddd4a00996c76afab7e269643fa9608f1e1b6bdba385071d"} Dec 07 16:02:21 crc kubenswrapper[4716]: I1207 16:02:21.705095 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"5a0a0eca3c2087a457c0e448e649345ec4595a9c8a019aed2cddc5f30c77871e"} Dec 07 16:02:21 crc kubenswrapper[4716]: I1207 16:02:21.705109 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"84a36b76f20618c7885d611ecad9db29bac5101c27572be31d3acbb90c581291"} Dec 07 16:02:21 crc kubenswrapper[4716]: I1207 16:02:21.705121 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"9e23abfdca314ed43d594333c9f9e401cc8ef7dad39b8ebac84a836b06ee5c41"} Dec 07 16:02:21 crc kubenswrapper[4716]: I1207 16:02:21.705506 4716 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 16:02:21 crc kubenswrapper[4716]: I1207 16:02:21.706506 4716 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 16:02:21 crc kubenswrapper[4716]: I1207 16:02:21.707788 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:21 crc kubenswrapper[4716]: I1207 16:02:21.707821 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:21 crc kubenswrapper[4716]: I1207 16:02:21.707834 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:21 crc kubenswrapper[4716]: I1207 16:02:21.708732 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:21 crc kubenswrapper[4716]: I1207 16:02:21.708824 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:21 crc kubenswrapper[4716]: I1207 16:02:21.708843 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:21 crc kubenswrapper[4716]: I1207 16:02:21.771494 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 07 16:02:22 crc kubenswrapper[4716]: I1207 16:02:22.662539 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 16:02:22 crc kubenswrapper[4716]: I1207 16:02:22.662718 4716 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 16:02:22 crc kubenswrapper[4716]: I1207 16:02:22.663130 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 16:02:22 crc kubenswrapper[4716]: I1207 16:02:22.663989 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:22 crc kubenswrapper[4716]: I1207 16:02:22.664033 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:22 crc kubenswrapper[4716]: I1207 16:02:22.664050 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:22 crc kubenswrapper[4716]: I1207 16:02:22.707642 4716 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 16:02:22 crc kubenswrapper[4716]: I1207 16:02:22.707687 4716 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 16:02:22 crc kubenswrapper[4716]: I1207 16:02:22.708506 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:22 crc kubenswrapper[4716]: I1207 16:02:22.708541 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:22 crc kubenswrapper[4716]: I1207 16:02:22.708554 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:22 crc kubenswrapper[4716]: I1207 16:02:22.709479 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:22 crc kubenswrapper[4716]: I1207 16:02:22.709515 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:22 crc kubenswrapper[4716]: I1207 16:02:22.709532 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:23 crc kubenswrapper[4716]: I1207 16:02:23.270470 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Dec 07 16:02:23 crc kubenswrapper[4716]: I1207 16:02:23.270686 4716 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 16:02:23 crc kubenswrapper[4716]: I1207 16:02:23.271784 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:23 crc kubenswrapper[4716]: I1207 16:02:23.271820 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:23 crc kubenswrapper[4716]: I1207 16:02:23.271832 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:23 crc kubenswrapper[4716]: I1207 16:02:23.484583 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 07 16:02:23 crc kubenswrapper[4716]: I1207 16:02:23.489565 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 07 16:02:23 crc kubenswrapper[4716]: I1207 16:02:23.648422 4716 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 07 16:02:23 crc kubenswrapper[4716]: I1207 16:02:23.648573 4716 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 07 16:02:23 crc kubenswrapper[4716]: I1207 16:02:23.709782 4716 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 16:02:23 crc kubenswrapper[4716]: I1207 16:02:23.710214 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 07 16:02:23 crc kubenswrapper[4716]: I1207 16:02:23.710943 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:23 crc kubenswrapper[4716]: I1207 16:02:23.710993 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:23 crc kubenswrapper[4716]: I1207 16:02:23.711010 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:24 crc kubenswrapper[4716]: I1207 16:02:24.541801 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 07 16:02:24 crc kubenswrapper[4716]: I1207 16:02:24.542031 4716 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 16:02:24 crc kubenswrapper[4716]: I1207 16:02:24.543252 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:24 crc kubenswrapper[4716]: I1207 16:02:24.543289 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:24 crc kubenswrapper[4716]: I1207 16:02:24.543299 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:24 crc kubenswrapper[4716]: I1207 16:02:24.629502 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 16:02:24 crc kubenswrapper[4716]: I1207 16:02:24.629697 4716 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 16:02:24 crc kubenswrapper[4716]: I1207 16:02:24.630706 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:24 crc kubenswrapper[4716]: I1207 16:02:24.630746 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:24 crc kubenswrapper[4716]: I1207 16:02:24.630758 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:24 crc kubenswrapper[4716]: I1207 16:02:24.712686 4716 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 16:02:24 crc kubenswrapper[4716]: I1207 16:02:24.713763 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:24 crc kubenswrapper[4716]: I1207 16:02:24.713820 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:24 crc kubenswrapper[4716]: I1207 16:02:24.713837 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:27 crc kubenswrapper[4716]: E1207 16:02:27.734682 4716 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 07 16:02:29 crc kubenswrapper[4716]: E1207 16:02:29.239013 4716 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": net/http: TLS handshake timeout" node="crc" Dec 07 16:02:29 crc kubenswrapper[4716]: I1207 16:02:29.335528 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 07 16:02:29 crc kubenswrapper[4716]: I1207 16:02:29.335653 4716 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 16:02:29 crc kubenswrapper[4716]: I1207 16:02:29.336699 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:29 crc kubenswrapper[4716]: I1207 16:02:29.336770 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:29 crc kubenswrapper[4716]: I1207 16:02:29.336789 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:29 crc kubenswrapper[4716]: I1207 16:02:29.557817 4716 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Dec 07 16:02:29 crc kubenswrapper[4716]: E1207 16:02:29.573492 4716 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 07 16:02:29 crc kubenswrapper[4716]: I1207 16:02:29.626186 4716 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 07 16:02:29 crc kubenswrapper[4716]: I1207 16:02:29.626293 4716 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 07 16:02:29 crc kubenswrapper[4716]: I1207 16:02:29.636426 4716 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 07 16:02:29 crc kubenswrapper[4716]: I1207 16:02:29.636539 4716 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 07 16:02:30 crc kubenswrapper[4716]: I1207 16:02:30.580250 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Dec 07 16:02:30 crc kubenswrapper[4716]: I1207 16:02:30.580471 4716 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 16:02:30 crc kubenswrapper[4716]: I1207 16:02:30.581550 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:30 crc kubenswrapper[4716]: I1207 16:02:30.581589 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:30 crc kubenswrapper[4716]: I1207 16:02:30.581598 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:30 crc kubenswrapper[4716]: I1207 16:02:30.839351 4716 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 16:02:30 crc kubenswrapper[4716]: I1207 16:02:30.840345 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:30 crc kubenswrapper[4716]: I1207 16:02:30.840378 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:30 crc kubenswrapper[4716]: I1207 16:02:30.840387 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:30 crc kubenswrapper[4716]: I1207 16:02:30.840407 4716 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 07 16:02:32 crc kubenswrapper[4716]: I1207 16:02:32.669669 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 16:02:32 crc kubenswrapper[4716]: I1207 16:02:32.669868 4716 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 16:02:32 crc kubenswrapper[4716]: I1207 16:02:32.670596 4716 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 07 16:02:32 crc kubenswrapper[4716]: I1207 16:02:32.670682 4716 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 07 16:02:32 crc kubenswrapper[4716]: I1207 16:02:32.671709 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:32 crc kubenswrapper[4716]: I1207 16:02:32.671769 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:32 crc kubenswrapper[4716]: I1207 16:02:32.671782 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:32 crc kubenswrapper[4716]: I1207 16:02:32.677298 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 16:02:32 crc kubenswrapper[4716]: I1207 16:02:32.732776 4716 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 07 16:02:32 crc kubenswrapper[4716]: I1207 16:02:32.733427 4716 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 07 16:02:32 crc kubenswrapper[4716]: I1207 16:02:32.733492 4716 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 07 16:02:32 crc kubenswrapper[4716]: I1207 16:02:32.734243 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:32 crc kubenswrapper[4716]: I1207 16:02:32.734292 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:32 crc kubenswrapper[4716]: I1207 16:02:32.734308 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:33 crc kubenswrapper[4716]: I1207 16:02:33.622169 4716 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 07 16:02:33 crc kubenswrapper[4716]: I1207 16:02:33.644559 4716 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Dec 07 16:02:33 crc kubenswrapper[4716]: I1207 16:02:33.648150 4716 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 07 16:02:33 crc kubenswrapper[4716]: I1207 16:02:33.648199 4716 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 07 16:02:34 crc kubenswrapper[4716]: E1207 16:02:34.627190 4716 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="3.2s" Dec 07 16:02:34 crc kubenswrapper[4716]: I1207 16:02:34.628820 4716 trace.go:236] Trace[16039352]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (07-Dec-2025 16:02:21.294) (total time: 13334ms): Dec 07 16:02:34 crc kubenswrapper[4716]: Trace[16039352]: ---"Objects listed" error: 13334ms (16:02:34.628) Dec 07 16:02:34 crc kubenswrapper[4716]: Trace[16039352]: [13.334296698s] [13.334296698s] END Dec 07 16:02:34 crc kubenswrapper[4716]: I1207 16:02:34.628840 4716 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 07 16:02:34 crc kubenswrapper[4716]: I1207 16:02:34.629030 4716 trace.go:236] Trace[1967501804]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (07-Dec-2025 16:02:21.775) (total time: 12853ms): Dec 07 16:02:34 crc kubenswrapper[4716]: Trace[1967501804]: ---"Objects listed" error: 12853ms (16:02:34.628) Dec 07 16:02:34 crc kubenswrapper[4716]: Trace[1967501804]: [12.853386919s] [12.853386919s] END Dec 07 16:02:34 crc kubenswrapper[4716]: I1207 16:02:34.629066 4716 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 07 16:02:34 crc kubenswrapper[4716]: I1207 16:02:34.630092 4716 trace.go:236] Trace[271933009]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (07-Dec-2025 16:02:22.111) (total time: 12518ms): Dec 07 16:02:34 crc kubenswrapper[4716]: Trace[271933009]: ---"Objects listed" error: 12518ms (16:02:34.629) Dec 07 16:02:34 crc kubenswrapper[4716]: Trace[271933009]: [12.518945085s] [12.518945085s] END Dec 07 16:02:34 crc kubenswrapper[4716]: I1207 16:02:34.630121 4716 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 07 16:02:34 crc kubenswrapper[4716]: I1207 16:02:34.630269 4716 trace.go:236] Trace[1660534691]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (07-Dec-2025 16:02:22.135) (total time: 12494ms): Dec 07 16:02:34 crc kubenswrapper[4716]: Trace[1660534691]: ---"Objects listed" error: 12494ms (16:02:34.630) Dec 07 16:02:34 crc kubenswrapper[4716]: Trace[1660534691]: [12.494263116s] [12.494263116s] END Dec 07 16:02:34 crc kubenswrapper[4716]: I1207 16:02:34.630290 4716 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 07 16:02:34 crc kubenswrapper[4716]: I1207 16:02:34.630947 4716 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.558215 4716 apiserver.go:52] "Watching apiserver" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.561031 4716 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.561439 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c"] Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.561954 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.562287 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.562422 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.562634 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.563103 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 07 16:02:35 crc kubenswrapper[4716]: E1207 16:02:35.563403 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 16:02:35 crc kubenswrapper[4716]: E1207 16:02:35.564849 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.565993 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:02:35 crc kubenswrapper[4716]: E1207 16:02:35.566049 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.567330 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.567497 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.567500 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.567646 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.571518 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.571698 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.571833 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.572112 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.572245 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.618385 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.632066 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.637184 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.637216 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.637236 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.637255 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.637275 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.637294 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.637310 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.637333 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.637352 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.637368 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.637386 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.637403 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.638041 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.638150 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.638316 4716 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.638420 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.642361 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.650309 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 07 16:02:35 crc kubenswrapper[4716]: E1207 16:02:35.651294 4716 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 07 16:02:35 crc kubenswrapper[4716]: E1207 16:02:35.651338 4716 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 07 16:02:35 crc kubenswrapper[4716]: E1207 16:02:35.651375 4716 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.651432 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 07 16:02:35 crc kubenswrapper[4716]: E1207 16:02:35.651499 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-07 16:02:36.151472586 +0000 UTC m=+18.841757528 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.656552 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 07 16:02:35 crc kubenswrapper[4716]: E1207 16:02:35.658276 4716 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 07 16:02:35 crc kubenswrapper[4716]: E1207 16:02:35.658303 4716 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 07 16:02:35 crc kubenswrapper[4716]: E1207 16:02:35.658329 4716 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 07 16:02:35 crc kubenswrapper[4716]: E1207 16:02:35.658389 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-07 16:02:36.158371366 +0000 UTC m=+18.848656288 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.658780 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.663151 4716 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.664439 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.664888 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.678290 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.687667 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.697524 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.707542 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.738063 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.738123 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.738149 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.738171 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.738193 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.738216 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.738240 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.738262 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.738284 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.738305 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.738325 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.738345 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.738366 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.738386 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.738408 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.738434 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.738457 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.738477 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.738499 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.738518 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.738539 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.738559 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.738578 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.738598 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.738619 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.738639 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.738697 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.738720 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.738742 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.738764 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.738784 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.738825 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.738868 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.738891 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.738911 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.738950 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.738971 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.738992 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.739014 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.739037 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.739057 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.739098 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.739120 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.739143 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.739283 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.739337 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.739387 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.739409 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.739433 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.739454 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.739475 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.739496 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.739517 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.739539 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.739561 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.739582 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.739604 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.739613 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.739626 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.739645 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.739704 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.739747 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.739779 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.739817 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.739865 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.739895 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.739915 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.739954 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.739989 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.740022 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.740032 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.740066 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.740137 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.740171 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.740177 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.740203 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.740237 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.740274 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.740308 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.740322 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.740343 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.740377 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.740418 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.740451 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.740484 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.740515 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.740548 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.740580 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.740615 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.740604 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.740654 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.740690 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.740726 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.740769 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.740817 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.740879 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.740933 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.740993 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.741055 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.741154 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.741204 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.741239 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.741274 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.741307 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.741350 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.741383 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.741423 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.741459 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.741490 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.741556 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.741594 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.741626 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.741659 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.741755 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.741796 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.741828 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.741865 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.741896 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.741930 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.741965 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.742475 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.742537 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.742586 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.742623 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.742674 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.742713 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.742748 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.742780 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.742817 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.742850 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.742903 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.742953 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.742995 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.743028 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.743060 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.743124 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.743166 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.743210 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.743299 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.743413 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.743567 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.743621 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.743721 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.744370 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.744446 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.744489 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.744526 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.744562 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.744599 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.744648 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.744716 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.744751 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.744785 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.744818 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.744865 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.744898 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.744934 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.744968 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.745383 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.745455 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.745495 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.745531 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.745568 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.745602 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.745646 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.745681 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.745724 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.745783 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.745835 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.745869 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.745912 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.745955 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.746010 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.746057 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.740773 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.740836 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.740838 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.746107 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.740893 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.740958 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.741045 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.741149 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.741745 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.741901 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.741942 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.742133 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.742196 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.742256 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.742310 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.742455 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.742506 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.742555 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.742628 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.742660 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.743280 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.743471 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.743567 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.743606 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.743675 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.743837 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.743917 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.743973 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.743956 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.744334 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.744519 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.744596 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.744659 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.744666 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.744689 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.744709 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.744765 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.744118 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.745477 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.745499 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.745637 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.746496 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.746127 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.746585 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.746671 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.746693 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.746144 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.746356 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.745708 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.746810 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.747447 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.747451 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.747446 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.747554 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.748755 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.759284 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.746129 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.763435 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.763908 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.764671 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.765494 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.765529 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.771971 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.761222 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.759644 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: E1207 16:02:35.759700 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 16:02:36.259671505 +0000 UTC m=+18.949956417 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.759902 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.774632 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.760776 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.761325 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.761366 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.774783 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.774856 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.775156 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.775262 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.775306 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.775339 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.775371 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.775396 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.775429 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.761749 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.761967 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.762598 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.762590 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.762763 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.763172 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.764189 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.764329 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.764469 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.764012 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.764564 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.764620 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.774723 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.775039 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.775120 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.775182 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.775189 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.775711 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.778073 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.778342 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.779732 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.780844 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.786505 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.775064 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.786554 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.786838 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.787621 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.787780 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.788402 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.788471 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.788505 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.788528 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.788521 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.788593 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.788848 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.788860 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.788893 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.788894 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.789122 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.789280 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.789303 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.789530 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.789846 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.789982 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.790071 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.789649 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.790620 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.790700 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.790888 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.790927 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.791146 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.791181 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.791383 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.791530 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.791536 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.791317 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.791830 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.791842 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.791744 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.792220 4716 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="0d31de0326a05216fa10442bdb47151c6e073c97e9a16c34e793028e5e253756" exitCode=255 Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.792209 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"0d31de0326a05216fa10442bdb47151c6e073c97e9a16c34e793028e5e253756"} Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.792033 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.788925 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.792371 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.792418 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.792446 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.792473 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.792493 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.792665 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.792757 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.792828 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.793166 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.793171 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.793912 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.794147 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.794567 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.794811 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.794936 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.794967 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.796178 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.794321 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.797509 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.798405 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.798443 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.798738 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.799037 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.799438 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.803468 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.803563 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.804028 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.806407 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.806442 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.807059 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.807129 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.807151 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.807195 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.807233 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.807558 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:02:35 crc kubenswrapper[4716]: E1207 16:02:35.807682 4716 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.807988 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: E1207 16:02:35.808124 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-07 16:02:36.30772741 +0000 UTC m=+18.998012332 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.808343 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.808645 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.808707 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.808748 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.808768 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.808775 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: E1207 16:02:35.808829 4716 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 07 16:02:35 crc kubenswrapper[4716]: E1207 16:02:35.808889 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-07 16:02:36.308871553 +0000 UTC m=+18.999156465 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.808936 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.808953 4716 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.808963 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.808973 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.808983 4716 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.808992 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809001 4716 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809015 4716 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809026 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809040 4716 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809052 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809064 4716 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809104 4716 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809123 4716 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809135 4716 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809146 4716 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809162 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809174 4716 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809185 4716 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809195 4716 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809206 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809217 4716 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809227 4716 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809236 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809245 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809256 4716 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809277 4716 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809289 4716 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809301 4716 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809304 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809310 4716 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809340 4716 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809353 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809362 4716 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809372 4716 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809382 4716 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809396 4716 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809406 4716 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809420 4716 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809428 4716 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809437 4716 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809446 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809455 4716 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809463 4716 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809472 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809481 4716 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809491 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809500 4716 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809510 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809518 4716 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809528 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809537 4716 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809545 4716 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809554 4716 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809563 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809572 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809582 4716 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809591 4716 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809599 4716 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809601 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809609 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809629 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809639 4716 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809648 4716 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809656 4716 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809670 4716 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809680 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809689 4716 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809698 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809706 4716 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809714 4716 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809722 4716 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809730 4716 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809740 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809749 4716 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809757 4716 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809766 4716 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809776 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809786 4716 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809796 4716 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809806 4716 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809818 4716 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809827 4716 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809835 4716 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809844 4716 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809854 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809862 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809871 4716 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809880 4716 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809888 4716 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809896 4716 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809904 4716 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809913 4716 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809922 4716 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809930 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809938 4716 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809946 4716 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809954 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809964 4716 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809973 4716 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809982 4716 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.809991 4716 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.810000 4716 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.810056 4716 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.810066 4716 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.810111 4716 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.810121 4716 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.810130 4716 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.810141 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.810150 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.810158 4716 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.810185 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.810193 4716 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.810204 4716 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.810212 4716 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.810254 4716 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.810266 4716 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.810281 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.810291 4716 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.810301 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.810309 4716 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.810318 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.810327 4716 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.810335 4716 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.810345 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.810353 4716 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.810362 4716 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.810372 4716 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.810380 4716 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.810388 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.810397 4716 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.810406 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.810415 4716 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.810424 4716 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.810432 4716 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.810440 4716 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.810449 4716 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.810458 4716 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.810467 4716 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.810476 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.810485 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.810494 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.810503 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.810512 4716 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.810521 4716 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.810532 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.810541 4716 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.810550 4716 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.810559 4716 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.810569 4716 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.810577 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.810586 4716 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.810594 4716 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.810602 4716 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.810611 4716 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.810619 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.810628 4716 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.810636 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.810820 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.813019 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.813214 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.813544 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.813772 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.814070 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.814242 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.814439 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.814617 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.814784 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.815057 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.818185 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.818459 4716 scope.go:117] "RemoveContainer" containerID="0d31de0326a05216fa10442bdb47151c6e073c97e9a16c34e793028e5e253756" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.818563 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.818963 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.819043 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.819712 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.819884 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.820033 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.820211 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.821000 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.821371 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.822158 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.822538 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.824436 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.827970 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.827899 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.828163 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.829653 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.829743 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.829971 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.830305 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.832484 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.832541 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.833064 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.833286 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.834172 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.838983 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.841161 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.841751 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.848487 4716 kubelet_node_status.go:115] "Node was previously registered" node="crc" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.848590 4716 kubelet_node_status.go:79] "Successfully registered node" node="crc" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.850049 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.850115 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.850129 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.850150 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.850163 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:35Z","lastTransitionTime":"2025-12-07T16:02:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.850432 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.863571 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.867558 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.868326 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 07 16:02:35 crc kubenswrapper[4716]: E1207 16:02:35.868642 4716 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a8d1b7d4-7ec4-40c7-86b0-291db1f38894\\\",\\\"systemUUID\\\":\\\"a57d09b0-0224-46eb-a3ac-a2b60dad2250\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.876349 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.876416 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.876437 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.876461 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.876482 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:35Z","lastTransitionTime":"2025-12-07T16:02:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.880942 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.885025 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.888218 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 07 16:02:35 crc kubenswrapper[4716]: E1207 16:02:35.891500 4716 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a8d1b7d4-7ec4-40c7-86b0-291db1f38894\\\",\\\"systemUUID\\\":\\\"a57d09b0-0224-46eb-a3ac-a2b60dad2250\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.896364 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.896402 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.896413 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.896430 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.896444 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:35Z","lastTransitionTime":"2025-12-07T16:02:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.896518 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 07 16:02:35 crc kubenswrapper[4716]: W1207 16:02:35.900751 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-1c3dda54efcf87ff8456dca3a2557162aa52d7617f74a99ee8e954c79f2d9534 WatchSource:0}: Error finding container 1c3dda54efcf87ff8456dca3a2557162aa52d7617f74a99ee8e954c79f2d9534: Status 404 returned error can't find the container with id 1c3dda54efcf87ff8456dca3a2557162aa52d7617f74a99ee8e954c79f2d9534 Dec 07 16:02:35 crc kubenswrapper[4716]: W1207 16:02:35.901204 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-2512ec8467369836d607215c28a5a691d7b6cc7dff0073bf9d6c8f7b6dd0839a WatchSource:0}: Error finding container 2512ec8467369836d607215c28a5a691d7b6cc7dff0073bf9d6c8f7b6dd0839a: Status 404 returned error can't find the container with id 2512ec8467369836d607215c28a5a691d7b6cc7dff0073bf9d6c8f7b6dd0839a Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.905695 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 07 16:02:35 crc kubenswrapper[4716]: E1207 16:02:35.905823 4716 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a8d1b7d4-7ec4-40c7-86b0-291db1f38894\\\",\\\"systemUUID\\\":\\\"a57d09b0-0224-46eb-a3ac-a2b60dad2250\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.910256 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.910284 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.910314 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.910327 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.910336 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:35Z","lastTransitionTime":"2025-12-07T16:02:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.911044 4716 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.911069 4716 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.911101 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.911115 4716 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.911126 4716 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.911138 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.911149 4716 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.911160 4716 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.911171 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.911182 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.911218 4716 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.911230 4716 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.911242 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.911254 4716 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.911266 4716 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.911279 4716 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.911290 4716 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.911301 4716 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.911312 4716 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.911323 4716 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.911335 4716 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.911346 4716 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.911357 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.911368 4716 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.911378 4716 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.911390 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.911400 4716 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.911413 4716 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.911425 4716 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.911436 4716 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.911447 4716 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.911458 4716 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.911469 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.911482 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.911496 4716 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.911508 4716 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.911520 4716 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.911568 4716 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Dec 07 16:02:35 crc kubenswrapper[4716]: E1207 16:02:35.919049 4716 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a8d1b7d4-7ec4-40c7-86b0-291db1f38894\\\",\\\"systemUUID\\\":\\\"a57d09b0-0224-46eb-a3ac-a2b60dad2250\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.923110 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.923144 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.923156 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.923169 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.923180 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:35Z","lastTransitionTime":"2025-12-07T16:02:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:35 crc kubenswrapper[4716]: E1207 16:02:35.936121 4716 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a8d1b7d4-7ec4-40c7-86b0-291db1f38894\\\",\\\"systemUUID\\\":\\\"a57d09b0-0224-46eb-a3ac-a2b60dad2250\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 07 16:02:35 crc kubenswrapper[4716]: E1207 16:02:35.938971 4716 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.941906 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.941931 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.941939 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.941952 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:35 crc kubenswrapper[4716]: I1207 16:02:35.941961 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:35Z","lastTransitionTime":"2025-12-07T16:02:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.044971 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.045001 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.045012 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.045028 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.045040 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:36Z","lastTransitionTime":"2025-12-07T16:02:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.146788 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.146826 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.146836 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.146850 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.146859 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:36Z","lastTransitionTime":"2025-12-07T16:02:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.214376 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.214425 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 16:02:36 crc kubenswrapper[4716]: E1207 16:02:36.214570 4716 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 07 16:02:36 crc kubenswrapper[4716]: E1207 16:02:36.214580 4716 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 07 16:02:36 crc kubenswrapper[4716]: E1207 16:02:36.214618 4716 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 07 16:02:36 crc kubenswrapper[4716]: E1207 16:02:36.214630 4716 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 07 16:02:36 crc kubenswrapper[4716]: E1207 16:02:36.214678 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-07 16:02:37.214663498 +0000 UTC m=+19.904948410 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 07 16:02:36 crc kubenswrapper[4716]: E1207 16:02:36.214587 4716 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 07 16:02:36 crc kubenswrapper[4716]: E1207 16:02:36.214719 4716 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 07 16:02:36 crc kubenswrapper[4716]: E1207 16:02:36.214777 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-07 16:02:37.21476026 +0000 UTC m=+19.905045272 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.249107 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.249148 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.249158 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.249173 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.249183 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:36Z","lastTransitionTime":"2025-12-07T16:02:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.314700 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.314786 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.314828 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:02:36 crc kubenswrapper[4716]: E1207 16:02:36.314938 4716 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 07 16:02:36 crc kubenswrapper[4716]: E1207 16:02:36.314990 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-07 16:02:37.314974479 +0000 UTC m=+20.005259401 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.315406 4716 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 16:02:36 crc kubenswrapper[4716]: E1207 16:02:36.315502 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 16:02:37.315485843 +0000 UTC m=+20.005770765 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:02:36 crc kubenswrapper[4716]: E1207 16:02:36.315558 4716 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 07 16:02:36 crc kubenswrapper[4716]: E1207 16:02:36.315595 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-07 16:02:37.315583846 +0000 UTC m=+20.005868768 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.351347 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.351381 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.351393 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.351409 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.351420 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:36Z","lastTransitionTime":"2025-12-07T16:02:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.453291 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.453315 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.453322 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.453334 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.453343 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:36Z","lastTransitionTime":"2025-12-07T16:02:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.555913 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.555955 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.555965 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.555979 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.555991 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:36Z","lastTransitionTime":"2025-12-07T16:02:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.656981 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 16:02:36 crc kubenswrapper[4716]: E1207 16:02:36.657708 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.661048 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.661102 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.661112 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.661126 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.661136 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:36Z","lastTransitionTime":"2025-12-07T16:02:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.763097 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.763132 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.763141 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.763158 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.763168 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:36Z","lastTransitionTime":"2025-12-07T16:02:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.795399 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"f95ef61779507288a03bda8eddf745e105f705998a117a30789675bfa452d88f"} Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.795444 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"a0b955b23163e7c8c458a3bbe2f681633e6a663ce76748febcc0b3a3c37f3ca9"} Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.795454 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"3037c5b910734282b47e37fe8d9f2766540551b063119af781d0384b4a64b377"} Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.796630 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"2512ec8467369836d607215c28a5a691d7b6cc7dff0073bf9d6c8f7b6dd0839a"} Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.797845 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"ef3335ae3875a75ab64685fc7e3590177280d651e8f8292c6f5ea6f472e55789"} Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.797893 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"1c3dda54efcf87ff8456dca3a2557162aa52d7617f74a99ee8e954c79f2d9534"} Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.799880 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.801019 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"dd318647816e4a27849666b0ec4fc2ca4905295661c557dc381e1144f22cf284"} Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.801753 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.817035 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:36Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.830898 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:36Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.843814 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f95ef61779507288a03bda8eddf745e105f705998a117a30789675bfa452d88f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b955b23163e7c8c458a3bbe2f681633e6a663ce76748febcc0b3a3c37f3ca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:36Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.855979 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:36Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.865034 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.865067 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.865093 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.865110 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.865123 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:36Z","lastTransitionTime":"2025-12-07T16:02:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.868755 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdee9957-48b9-458f-ba71-e0cd39f80596\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73f85248b4d68a9beecdacb3a01f433e97a0ffea2247ce92a41ef96855509377\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6276df652efe6ca935e745f24ef20f87a16465475139230c63b9999234a4d12a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3b1dd202586ee755ca89fbe12b30bbb12f95a62514f6e4d93cbbdbe55181e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d31de0326a05216fa10442bdb47151c6e073c97e9a16c34e793028e5e253756\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d31de0326a05216fa10442bdb47151c6e073c97e9a16c34e793028e5e253756\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"le observer\\\\nW1207 16:02:34.645464 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1207 16:02:34.645587 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 16:02:34.646501 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2082283479/tls.crt::/tmp/serving-cert-2082283479/tls.key\\\\\\\"\\\\nI1207 16:02:34.972041 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 16:02:34.973702 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 16:02:34.973717 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 16:02:34.973738 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 16:02:34.973743 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 16:02:34.977185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1207 16:02:34.977204 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1207 16:02:34.977209 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977217 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 16:02:34.977225 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 16:02:34.977228 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 16:02:34.977232 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1207 16:02:34.979033 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27bd4a54f920b2725987c1c7fd2d9fdf16cd08738c246c333a4e9164a3a582a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:36Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.881978 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:36Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.901058 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:36Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.914864 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f95ef61779507288a03bda8eddf745e105f705998a117a30789675bfa452d88f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b955b23163e7c8c458a3bbe2f681633e6a663ce76748febcc0b3a3c37f3ca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:36Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.929642 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef3335ae3875a75ab64685fc7e3590177280d651e8f8292c6f5ea6f472e55789\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:36Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.942414 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:36Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.953317 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:36Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.965859 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:36Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.967059 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.967109 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.967122 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.967139 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.967151 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:36Z","lastTransitionTime":"2025-12-07T16:02:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:36 crc kubenswrapper[4716]: I1207 16:02:36.985776 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdee9957-48b9-458f-ba71-e0cd39f80596\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73f85248b4d68a9beecdacb3a01f433e97a0ffea2247ce92a41ef96855509377\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6276df652efe6ca935e745f24ef20f87a16465475139230c63b9999234a4d12a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3b1dd202586ee755ca89fbe12b30bbb12f95a62514f6e4d93cbbdbe55181e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd318647816e4a27849666b0ec4fc2ca4905295661c557dc381e1144f22cf284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d31de0326a05216fa10442bdb47151c6e073c97e9a16c34e793028e5e253756\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"le observer\\\\nW1207 16:02:34.645464 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1207 16:02:34.645587 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 16:02:34.646501 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2082283479/tls.crt::/tmp/serving-cert-2082283479/tls.key\\\\\\\"\\\\nI1207 16:02:34.972041 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 16:02:34.973702 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 16:02:34.973717 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 16:02:34.973738 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 16:02:34.973743 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 16:02:34.977185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1207 16:02:34.977204 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1207 16:02:34.977209 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977217 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 16:02:34.977225 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 16:02:34.977228 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 16:02:34.977232 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1207 16:02:34.979033 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27bd4a54f920b2725987c1c7fd2d9fdf16cd08738c246c333a4e9164a3a582a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:36Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.003059 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:37Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.069581 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.069607 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.069616 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.069629 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.069638 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:37Z","lastTransitionTime":"2025-12-07T16:02:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.171634 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.171667 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.171679 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.171695 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.171707 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:37Z","lastTransitionTime":"2025-12-07T16:02:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.221756 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.221808 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 16:02:37 crc kubenswrapper[4716]: E1207 16:02:37.221904 4716 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 07 16:02:37 crc kubenswrapper[4716]: E1207 16:02:37.221919 4716 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 07 16:02:37 crc kubenswrapper[4716]: E1207 16:02:37.221929 4716 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 07 16:02:37 crc kubenswrapper[4716]: E1207 16:02:37.221962 4716 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 07 16:02:37 crc kubenswrapper[4716]: E1207 16:02:37.221994 4716 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 07 16:02:37 crc kubenswrapper[4716]: E1207 16:02:37.222009 4716 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 07 16:02:37 crc kubenswrapper[4716]: E1207 16:02:37.221978 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-07 16:02:39.221965319 +0000 UTC m=+21.912250231 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 07 16:02:37 crc kubenswrapper[4716]: E1207 16:02:37.222112 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-07 16:02:39.222067911 +0000 UTC m=+21.912352823 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.273951 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.273991 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.274000 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.274014 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.274025 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:37Z","lastTransitionTime":"2025-12-07T16:02:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.323102 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.323166 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.323207 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:02:37 crc kubenswrapper[4716]: E1207 16:02:37.323281 4716 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 07 16:02:37 crc kubenswrapper[4716]: E1207 16:02:37.323337 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-07 16:02:39.32332263 +0000 UTC m=+22.013607552 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 07 16:02:37 crc kubenswrapper[4716]: E1207 16:02:37.323380 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 16:02:39.32334503 +0000 UTC m=+22.013629972 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:02:37 crc kubenswrapper[4716]: E1207 16:02:37.323382 4716 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 07 16:02:37 crc kubenswrapper[4716]: E1207 16:02:37.323483 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-07 16:02:39.323469394 +0000 UTC m=+22.013754346 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.375932 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.376000 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.376025 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.376054 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.376109 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:37Z","lastTransitionTime":"2025-12-07T16:02:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.477893 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.477972 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.477999 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.478028 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.478049 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:37Z","lastTransitionTime":"2025-12-07T16:02:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.580782 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.580824 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.580833 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.580848 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.580857 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:37Z","lastTransitionTime":"2025-12-07T16:02:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.657504 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:02:37 crc kubenswrapper[4716]: E1207 16:02:37.657680 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.657708 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 16:02:37 crc kubenswrapper[4716]: E1207 16:02:37.657936 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.663570 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.664360 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.665406 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.666291 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.667055 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.669190 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.670469 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.671745 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.673409 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.674952 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.676218 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.680475 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.681608 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.683386 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.683445 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.683462 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.683487 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.683505 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:37Z","lastTransitionTime":"2025-12-07T16:02:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.683829 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.684791 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdee9957-48b9-458f-ba71-e0cd39f80596\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73f85248b4d68a9beecdacb3a01f433e97a0ffea2247ce92a41ef96855509377\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6276df652efe6ca935e745f24ef20f87a16465475139230c63b9999234a4d12a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3b1dd202586ee755ca89fbe12b30bbb12f95a62514f6e4d93cbbdbe55181e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd318647816e4a27849666b0ec4fc2ca4905295661c557dc381e1144f22cf284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d31de0326a05216fa10442bdb47151c6e073c97e9a16c34e793028e5e253756\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"le observer\\\\nW1207 16:02:34.645464 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1207 16:02:34.645587 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 16:02:34.646501 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2082283479/tls.crt::/tmp/serving-cert-2082283479/tls.key\\\\\\\"\\\\nI1207 16:02:34.972041 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 16:02:34.973702 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 16:02:34.973717 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 16:02:34.973738 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 16:02:34.973743 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 16:02:34.977185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1207 16:02:34.977204 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1207 16:02:34.977209 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977217 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 16:02:34.977225 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 16:02:34.977228 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 16:02:34.977232 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1207 16:02:34.979033 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27bd4a54f920b2725987c1c7fd2d9fdf16cd08738c246c333a4e9164a3a582a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:37Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.685002 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.686482 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.688898 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.690113 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.691690 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.693483 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.694129 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.694932 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.695594 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.696511 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.697116 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.697943 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.698760 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.699482 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.700382 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.701199 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.701850 4716 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.701981 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.704399 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.704640 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:37Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.704951 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.705397 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.706693 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.707940 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.709165 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.710587 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.712029 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.713011 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.714308 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.715616 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.716847 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.719321 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.720845 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.720915 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:37Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.722639 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.723490 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.723957 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.724404 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.724868 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.725361 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.725903 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.726370 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.744421 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:37Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.763322 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef3335ae3875a75ab64685fc7e3590177280d651e8f8292c6f5ea6f472e55789\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:37Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.784647 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:37Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.787888 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.787924 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.787933 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.787947 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.787957 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:37Z","lastTransitionTime":"2025-12-07T16:02:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.803316 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f95ef61779507288a03bda8eddf745e105f705998a117a30789675bfa452d88f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b955b23163e7c8c458a3bbe2f681633e6a663ce76748febcc0b3a3c37f3ca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:37Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.889662 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.889687 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.889695 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.889708 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.889717 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:37Z","lastTransitionTime":"2025-12-07T16:02:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.991971 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.992041 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.992059 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.992108 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:37 crc kubenswrapper[4716]: I1207 16:02:37.992126 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:37Z","lastTransitionTime":"2025-12-07T16:02:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:38 crc kubenswrapper[4716]: I1207 16:02:38.094309 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:38 crc kubenswrapper[4716]: I1207 16:02:38.094347 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:38 crc kubenswrapper[4716]: I1207 16:02:38.094358 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:38 crc kubenswrapper[4716]: I1207 16:02:38.094375 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:38 crc kubenswrapper[4716]: I1207 16:02:38.094387 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:38Z","lastTransitionTime":"2025-12-07T16:02:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:38 crc kubenswrapper[4716]: I1207 16:02:38.196322 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:38 crc kubenswrapper[4716]: I1207 16:02:38.196396 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:38 crc kubenswrapper[4716]: I1207 16:02:38.196407 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:38 crc kubenswrapper[4716]: I1207 16:02:38.196424 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:38 crc kubenswrapper[4716]: I1207 16:02:38.196435 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:38Z","lastTransitionTime":"2025-12-07T16:02:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:38 crc kubenswrapper[4716]: I1207 16:02:38.299224 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:38 crc kubenswrapper[4716]: I1207 16:02:38.299250 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:38 crc kubenswrapper[4716]: I1207 16:02:38.299259 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:38 crc kubenswrapper[4716]: I1207 16:02:38.299270 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:38 crc kubenswrapper[4716]: I1207 16:02:38.299279 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:38Z","lastTransitionTime":"2025-12-07T16:02:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:38 crc kubenswrapper[4716]: I1207 16:02:38.401764 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:38 crc kubenswrapper[4716]: I1207 16:02:38.401793 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:38 crc kubenswrapper[4716]: I1207 16:02:38.401802 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:38 crc kubenswrapper[4716]: I1207 16:02:38.401814 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:38 crc kubenswrapper[4716]: I1207 16:02:38.401824 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:38Z","lastTransitionTime":"2025-12-07T16:02:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:38 crc kubenswrapper[4716]: I1207 16:02:38.503536 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:38 crc kubenswrapper[4716]: I1207 16:02:38.503573 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:38 crc kubenswrapper[4716]: I1207 16:02:38.503586 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:38 crc kubenswrapper[4716]: I1207 16:02:38.503602 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:38 crc kubenswrapper[4716]: I1207 16:02:38.503613 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:38Z","lastTransitionTime":"2025-12-07T16:02:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:38 crc kubenswrapper[4716]: I1207 16:02:38.607006 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:38 crc kubenswrapper[4716]: I1207 16:02:38.607065 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:38 crc kubenswrapper[4716]: I1207 16:02:38.607108 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:38 crc kubenswrapper[4716]: I1207 16:02:38.607131 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:38 crc kubenswrapper[4716]: I1207 16:02:38.607149 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:38Z","lastTransitionTime":"2025-12-07T16:02:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:38 crc kubenswrapper[4716]: I1207 16:02:38.657615 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 16:02:38 crc kubenswrapper[4716]: E1207 16:02:38.657748 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 16:02:38 crc kubenswrapper[4716]: I1207 16:02:38.709955 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:38 crc kubenswrapper[4716]: I1207 16:02:38.710006 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:38 crc kubenswrapper[4716]: I1207 16:02:38.710023 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:38 crc kubenswrapper[4716]: I1207 16:02:38.710042 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:38 crc kubenswrapper[4716]: I1207 16:02:38.710058 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:38Z","lastTransitionTime":"2025-12-07T16:02:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:38 crc kubenswrapper[4716]: I1207 16:02:38.806536 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"e4742ea84b3a2c9461d9b900beee7354a2ded24c39ad23e007cbf270ae12377c"} Dec 07 16:02:38 crc kubenswrapper[4716]: I1207 16:02:38.812470 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:38 crc kubenswrapper[4716]: I1207 16:02:38.812509 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:38 crc kubenswrapper[4716]: I1207 16:02:38.812519 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:38 crc kubenswrapper[4716]: I1207 16:02:38.812533 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:38 crc kubenswrapper[4716]: I1207 16:02:38.812543 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:38Z","lastTransitionTime":"2025-12-07T16:02:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:38 crc kubenswrapper[4716]: I1207 16:02:38.819659 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef3335ae3875a75ab64685fc7e3590177280d651e8f8292c6f5ea6f472e55789\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:38Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:38 crc kubenswrapper[4716]: I1207 16:02:38.850769 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:38Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:38 crc kubenswrapper[4716]: I1207 16:02:38.871906 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f95ef61779507288a03bda8eddf745e105f705998a117a30789675bfa452d88f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b955b23163e7c8c458a3bbe2f681633e6a663ce76748febcc0b3a3c37f3ca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:38Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:38 crc kubenswrapper[4716]: I1207 16:02:38.886205 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4742ea84b3a2c9461d9b900beee7354a2ded24c39ad23e007cbf270ae12377c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:38Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:38 crc kubenswrapper[4716]: I1207 16:02:38.901999 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:38Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:38 crc kubenswrapper[4716]: I1207 16:02:38.914965 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:38 crc kubenswrapper[4716]: I1207 16:02:38.915004 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:38 crc kubenswrapper[4716]: I1207 16:02:38.915014 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:38 crc kubenswrapper[4716]: I1207 16:02:38.915027 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:38 crc kubenswrapper[4716]: I1207 16:02:38.915037 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:38Z","lastTransitionTime":"2025-12-07T16:02:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:38 crc kubenswrapper[4716]: I1207 16:02:38.915111 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:38Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:38 crc kubenswrapper[4716]: I1207 16:02:38.933305 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdee9957-48b9-458f-ba71-e0cd39f80596\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73f85248b4d68a9beecdacb3a01f433e97a0ffea2247ce92a41ef96855509377\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6276df652efe6ca935e745f24ef20f87a16465475139230c63b9999234a4d12a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3b1dd202586ee755ca89fbe12b30bbb12f95a62514f6e4d93cbbdbe55181e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd318647816e4a27849666b0ec4fc2ca4905295661c557dc381e1144f22cf284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d31de0326a05216fa10442bdb47151c6e073c97e9a16c34e793028e5e253756\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"le observer\\\\nW1207 16:02:34.645464 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1207 16:02:34.645587 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 16:02:34.646501 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2082283479/tls.crt::/tmp/serving-cert-2082283479/tls.key\\\\\\\"\\\\nI1207 16:02:34.972041 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 16:02:34.973702 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 16:02:34.973717 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 16:02:34.973738 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 16:02:34.973743 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 16:02:34.977185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1207 16:02:34.977204 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1207 16:02:34.977209 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977217 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 16:02:34.977225 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 16:02:34.977228 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 16:02:34.977232 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1207 16:02:34.979033 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27bd4a54f920b2725987c1c7fd2d9fdf16cd08738c246c333a4e9164a3a582a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:38Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.016797 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.016854 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.016871 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.016894 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.016909 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:39Z","lastTransitionTime":"2025-12-07T16:02:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.119162 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.119195 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.119203 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.119218 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.119228 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:39Z","lastTransitionTime":"2025-12-07T16:02:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.220884 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.220913 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.220923 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.220939 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.220948 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:39Z","lastTransitionTime":"2025-12-07T16:02:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.238839 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.238903 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 16:02:39 crc kubenswrapper[4716]: E1207 16:02:39.239062 4716 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 07 16:02:39 crc kubenswrapper[4716]: E1207 16:02:39.239103 4716 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 07 16:02:39 crc kubenswrapper[4716]: E1207 16:02:39.239118 4716 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 07 16:02:39 crc kubenswrapper[4716]: E1207 16:02:39.239170 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-07 16:02:43.239154903 +0000 UTC m=+25.929439815 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 07 16:02:39 crc kubenswrapper[4716]: E1207 16:02:39.239235 4716 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 07 16:02:39 crc kubenswrapper[4716]: E1207 16:02:39.239248 4716 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 07 16:02:39 crc kubenswrapper[4716]: E1207 16:02:39.239258 4716 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 07 16:02:39 crc kubenswrapper[4716]: E1207 16:02:39.239282 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-07 16:02:43.239274187 +0000 UTC m=+25.929559099 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.276887 4716 csr.go:261] certificate signing request csr-dq5lb is approved, waiting to be issued Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.311109 4716 csr.go:257] certificate signing request csr-dq5lb is issued Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.322911 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.322961 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.322972 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.322995 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.323008 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:39Z","lastTransitionTime":"2025-12-07T16:02:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.339403 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 16:02:39 crc kubenswrapper[4716]: E1207 16:02:39.339530 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 16:02:43.339507146 +0000 UTC m=+26.029792058 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.339642 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.339670 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:02:39 crc kubenswrapper[4716]: E1207 16:02:39.339769 4716 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 07 16:02:39 crc kubenswrapper[4716]: E1207 16:02:39.339822 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-07 16:02:43.339813404 +0000 UTC m=+26.030098316 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 07 16:02:39 crc kubenswrapper[4716]: E1207 16:02:39.339930 4716 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 07 16:02:39 crc kubenswrapper[4716]: E1207 16:02:39.340022 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-07 16:02:43.34000033 +0000 UTC m=+26.030285242 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.425465 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.425509 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.425525 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.425543 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.425556 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:39Z","lastTransitionTime":"2025-12-07T16:02:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.429671 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-2jjld"] Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.429967 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-2jjld" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.431962 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.434947 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.435138 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.445071 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:39Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.454703 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:39Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.465034 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2jjld" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dee6f5c-861c-4819-9309-7791574c7b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmdns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2jjld\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:39Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.478908 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdee9957-48b9-458f-ba71-e0cd39f80596\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73f85248b4d68a9beecdacb3a01f433e97a0ffea2247ce92a41ef96855509377\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6276df652efe6ca935e745f24ef20f87a16465475139230c63b9999234a4d12a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3b1dd202586ee755ca89fbe12b30bbb12f95a62514f6e4d93cbbdbe55181e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd318647816e4a27849666b0ec4fc2ca4905295661c557dc381e1144f22cf284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d31de0326a05216fa10442bdb47151c6e073c97e9a16c34e793028e5e253756\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"le observer\\\\nW1207 16:02:34.645464 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1207 16:02:34.645587 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 16:02:34.646501 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2082283479/tls.crt::/tmp/serving-cert-2082283479/tls.key\\\\\\\"\\\\nI1207 16:02:34.972041 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 16:02:34.973702 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 16:02:34.973717 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 16:02:34.973738 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 16:02:34.973743 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 16:02:34.977185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1207 16:02:34.977204 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1207 16:02:34.977209 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977217 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 16:02:34.977225 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 16:02:34.977228 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 16:02:34.977232 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1207 16:02:34.979033 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27bd4a54f920b2725987c1c7fd2d9fdf16cd08738c246c333a4e9164a3a582a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:39Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.495542 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4742ea84b3a2c9461d9b900beee7354a2ded24c39ad23e007cbf270ae12377c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:39Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.511010 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f95ef61779507288a03bda8eddf745e105f705998a117a30789675bfa452d88f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b955b23163e7c8c458a3bbe2f681633e6a663ce76748febcc0b3a3c37f3ca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:39Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.523110 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef3335ae3875a75ab64685fc7e3590177280d651e8f8292c6f5ea6f472e55789\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:39Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.527596 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.527645 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.527655 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.527671 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.527682 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:39Z","lastTransitionTime":"2025-12-07T16:02:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.542164 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rmdns\" (UniqueName: \"kubernetes.io/projected/3dee6f5c-861c-4819-9309-7791574c7b80-kube-api-access-rmdns\") pod \"node-resolver-2jjld\" (UID: \"3dee6f5c-861c-4819-9309-7791574c7b80\") " pod="openshift-dns/node-resolver-2jjld" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.542243 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/3dee6f5c-861c-4819-9309-7791574c7b80-hosts-file\") pod \"node-resolver-2jjld\" (UID: \"3dee6f5c-861c-4819-9309-7791574c7b80\") " pod="openshift-dns/node-resolver-2jjld" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.597231 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:39Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.629583 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.629621 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.629632 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.629653 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.629665 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:39Z","lastTransitionTime":"2025-12-07T16:02:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.642781 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/3dee6f5c-861c-4819-9309-7791574c7b80-hosts-file\") pod \"node-resolver-2jjld\" (UID: \"3dee6f5c-861c-4819-9309-7791574c7b80\") " pod="openshift-dns/node-resolver-2jjld" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.642851 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rmdns\" (UniqueName: \"kubernetes.io/projected/3dee6f5c-861c-4819-9309-7791574c7b80-kube-api-access-rmdns\") pod \"node-resolver-2jjld\" (UID: \"3dee6f5c-861c-4819-9309-7791574c7b80\") " pod="openshift-dns/node-resolver-2jjld" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.642911 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/3dee6f5c-861c-4819-9309-7791574c7b80-hosts-file\") pod \"node-resolver-2jjld\" (UID: \"3dee6f5c-861c-4819-9309-7791574c7b80\") " pod="openshift-dns/node-resolver-2jjld" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.656892 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:02:39 crc kubenswrapper[4716]: E1207 16:02:39.657013 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.657145 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 16:02:39 crc kubenswrapper[4716]: E1207 16:02:39.657400 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.659608 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rmdns\" (UniqueName: \"kubernetes.io/projected/3dee6f5c-861c-4819-9309-7791574c7b80-kube-api-access-rmdns\") pod \"node-resolver-2jjld\" (UID: \"3dee6f5c-861c-4819-9309-7791574c7b80\") " pod="openshift-dns/node-resolver-2jjld" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.731902 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.731941 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.731953 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.731970 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.731982 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:39Z","lastTransitionTime":"2025-12-07T16:02:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.740171 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-2jjld" Dec 07 16:02:39 crc kubenswrapper[4716]: W1207 16:02:39.753435 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dee6f5c_861c_4819_9309_7791574c7b80.slice/crio-1cd445c5b5ac1bf19c6ba36cc2419e780e7e0686c8239d0912f3e1347cdcc9e3 WatchSource:0}: Error finding container 1cd445c5b5ac1bf19c6ba36cc2419e780e7e0686c8239d0912f3e1347cdcc9e3: Status 404 returned error can't find the container with id 1cd445c5b5ac1bf19c6ba36cc2419e780e7e0686c8239d0912f3e1347cdcc9e3 Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.810701 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-2jjld" event={"ID":"3dee6f5c-861c-4819-9309-7791574c7b80","Type":"ContainerStarted","Data":"1cd445c5b5ac1bf19c6ba36cc2419e780e7e0686c8239d0912f3e1347cdcc9e3"} Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.819182 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-zcxxp"] Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.819538 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.820870 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.822536 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.822679 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.822862 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.823453 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.836448 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:39Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.837346 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.837398 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.837412 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.837431 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.837446 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:39Z","lastTransitionTime":"2025-12-07T16:02:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.856579 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:39Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.868198 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2jjld" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dee6f5c-861c-4819-9309-7791574c7b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmdns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2jjld\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:39Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.879121 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c15b59eb-565d-4556-a4ce-75afdf159dc8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jks8z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jks8z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zcxxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:39Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.891723 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdee9957-48b9-458f-ba71-e0cd39f80596\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73f85248b4d68a9beecdacb3a01f433e97a0ffea2247ce92a41ef96855509377\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6276df652efe6ca935e745f24ef20f87a16465475139230c63b9999234a4d12a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3b1dd202586ee755ca89fbe12b30bbb12f95a62514f6e4d93cbbdbe55181e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd318647816e4a27849666b0ec4fc2ca4905295661c557dc381e1144f22cf284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d31de0326a05216fa10442bdb47151c6e073c97e9a16c34e793028e5e253756\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"le observer\\\\nW1207 16:02:34.645464 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1207 16:02:34.645587 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 16:02:34.646501 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2082283479/tls.crt::/tmp/serving-cert-2082283479/tls.key\\\\\\\"\\\\nI1207 16:02:34.972041 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 16:02:34.973702 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 16:02:34.973717 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 16:02:34.973738 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 16:02:34.973743 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 16:02:34.977185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1207 16:02:34.977204 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1207 16:02:34.977209 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977217 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 16:02:34.977225 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 16:02:34.977228 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 16:02:34.977232 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1207 16:02:34.979033 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27bd4a54f920b2725987c1c7fd2d9fdf16cd08738c246c333a4e9164a3a582a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:39Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.902250 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4742ea84b3a2c9461d9b900beee7354a2ded24c39ad23e007cbf270ae12377c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:39Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.914563 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f95ef61779507288a03bda8eddf745e105f705998a117a30789675bfa452d88f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b955b23163e7c8c458a3bbe2f681633e6a663ce76748febcc0b3a3c37f3ca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:39Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.925675 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef3335ae3875a75ab64685fc7e3590177280d651e8f8292c6f5ea6f472e55789\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:39Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.940192 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.940234 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.940244 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.940257 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.940266 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:39Z","lastTransitionTime":"2025-12-07T16:02:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.945980 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:39Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.946040 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/c15b59eb-565d-4556-a4ce-75afdf159dc8-proxy-tls\") pod \"machine-config-daemon-zcxxp\" (UID: \"c15b59eb-565d-4556-a4ce-75afdf159dc8\") " pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.946115 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jks8z\" (UniqueName: \"kubernetes.io/projected/c15b59eb-565d-4556-a4ce-75afdf159dc8-kube-api-access-jks8z\") pod \"machine-config-daemon-zcxxp\" (UID: \"c15b59eb-565d-4556-a4ce-75afdf159dc8\") " pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.946197 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/c15b59eb-565d-4556-a4ce-75afdf159dc8-rootfs\") pod \"machine-config-daemon-zcxxp\" (UID: \"c15b59eb-565d-4556-a4ce-75afdf159dc8\") " pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" Dec 07 16:02:39 crc kubenswrapper[4716]: I1207 16:02:39.946228 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c15b59eb-565d-4556-a4ce-75afdf159dc8-mcd-auth-proxy-config\") pod \"machine-config-daemon-zcxxp\" (UID: \"c15b59eb-565d-4556-a4ce-75afdf159dc8\") " pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.042485 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.042536 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.042548 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.042564 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.042578 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:40Z","lastTransitionTime":"2025-12-07T16:02:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.046882 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/c15b59eb-565d-4556-a4ce-75afdf159dc8-rootfs\") pod \"machine-config-daemon-zcxxp\" (UID: \"c15b59eb-565d-4556-a4ce-75afdf159dc8\") " pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.046922 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c15b59eb-565d-4556-a4ce-75afdf159dc8-mcd-auth-proxy-config\") pod \"machine-config-daemon-zcxxp\" (UID: \"c15b59eb-565d-4556-a4ce-75afdf159dc8\") " pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.046976 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/c15b59eb-565d-4556-a4ce-75afdf159dc8-proxy-tls\") pod \"machine-config-daemon-zcxxp\" (UID: \"c15b59eb-565d-4556-a4ce-75afdf159dc8\") " pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.047004 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jks8z\" (UniqueName: \"kubernetes.io/projected/c15b59eb-565d-4556-a4ce-75afdf159dc8-kube-api-access-jks8z\") pod \"machine-config-daemon-zcxxp\" (UID: \"c15b59eb-565d-4556-a4ce-75afdf159dc8\") " pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.047029 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/c15b59eb-565d-4556-a4ce-75afdf159dc8-rootfs\") pod \"machine-config-daemon-zcxxp\" (UID: \"c15b59eb-565d-4556-a4ce-75afdf159dc8\") " pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.047742 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c15b59eb-565d-4556-a4ce-75afdf159dc8-mcd-auth-proxy-config\") pod \"machine-config-daemon-zcxxp\" (UID: \"c15b59eb-565d-4556-a4ce-75afdf159dc8\") " pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.052305 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/c15b59eb-565d-4556-a4ce-75afdf159dc8-proxy-tls\") pod \"machine-config-daemon-zcxxp\" (UID: \"c15b59eb-565d-4556-a4ce-75afdf159dc8\") " pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.070413 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jks8z\" (UniqueName: \"kubernetes.io/projected/c15b59eb-565d-4556-a4ce-75afdf159dc8-kube-api-access-jks8z\") pod \"machine-config-daemon-zcxxp\" (UID: \"c15b59eb-565d-4556-a4ce-75afdf159dc8\") " pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.130422 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" Dec 07 16:02:40 crc kubenswrapper[4716]: W1207 16:02:40.141048 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc15b59eb_565d_4556_a4ce_75afdf159dc8.slice/crio-390c0fff32d3270bc2d944de6b3abd6b6150de4c91e750d472c5ded6f05ae3ea WatchSource:0}: Error finding container 390c0fff32d3270bc2d944de6b3abd6b6150de4c91e750d472c5ded6f05ae3ea: Status 404 returned error can't find the container with id 390c0fff32d3270bc2d944de6b3abd6b6150de4c91e750d472c5ded6f05ae3ea Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.144506 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.144543 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.144553 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.144568 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.144578 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:40Z","lastTransitionTime":"2025-12-07T16:02:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.219012 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-8ps9j"] Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.219539 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-k4wr2"] Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.219695 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-8ps9j" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.219766 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-k4wr2" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.225742 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.225821 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.226276 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.226460 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.226504 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.226656 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.226711 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.247570 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.247614 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.247624 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.247641 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.247652 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:40Z","lastTransitionTime":"2025-12-07T16:02:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.261327 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef3335ae3875a75ab64685fc7e3590177280d651e8f8292c6f5ea6f472e55789\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:40Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.276142 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:40Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.298947 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f95ef61779507288a03bda8eddf745e105f705998a117a30789675bfa452d88f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b955b23163e7c8c458a3bbe2f681633e6a663ce76748febcc0b3a3c37f3ca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:40Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.312950 4716 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-12-07 15:57:39 +0000 UTC, rotation deadline is 2026-08-26 13:04:46.460685271 +0000 UTC Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.313030 4716 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 6285h2m6.147656989s for next certificate rotation Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.316395 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8ps9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98015cea-d753-4e7a-b2b5-5ae7733a81e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8ps9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:40Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.329516 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4742ea84b3a2c9461d9b900beee7354a2ded24c39ad23e007cbf270ae12377c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:40Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.347839 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:40Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.349828 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.349953 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.350060 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.350070 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0baf8575-26cf-4a72-8a6d-ae94c078eaad-host-run-netns\") pod \"multus-k4wr2\" (UID: \"0baf8575-26cf-4a72-8a6d-ae94c078eaad\") " pod="openshift-multus/multus-k4wr2" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.350195 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/0baf8575-26cf-4a72-8a6d-ae94c078eaad-multus-daemon-config\") pod \"multus-k4wr2\" (UID: \"0baf8575-26cf-4a72-8a6d-ae94c078eaad\") " pod="openshift-multus/multus-k4wr2" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.350214 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/98015cea-d753-4e7a-b2b5-5ae7733a81e3-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-8ps9j\" (UID: \"98015cea-d753-4e7a-b2b5-5ae7733a81e3\") " pod="openshift-multus/multus-additional-cni-plugins-8ps9j" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.350231 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/0baf8575-26cf-4a72-8a6d-ae94c078eaad-host-var-lib-cni-multus\") pod \"multus-k4wr2\" (UID: \"0baf8575-26cf-4a72-8a6d-ae94c078eaad\") " pod="openshift-multus/multus-k4wr2" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.350249 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/98015cea-d753-4e7a-b2b5-5ae7733a81e3-system-cni-dir\") pod \"multus-additional-cni-plugins-8ps9j\" (UID: \"98015cea-d753-4e7a-b2b5-5ae7733a81e3\") " pod="openshift-multus/multus-additional-cni-plugins-8ps9j" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.350265 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/98015cea-d753-4e7a-b2b5-5ae7733a81e3-tuning-conf-dir\") pod \"multus-additional-cni-plugins-8ps9j\" (UID: \"98015cea-d753-4e7a-b2b5-5ae7733a81e3\") " pod="openshift-multus/multus-additional-cni-plugins-8ps9j" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.350279 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/0baf8575-26cf-4a72-8a6d-ae94c078eaad-cni-binary-copy\") pod \"multus-k4wr2\" (UID: \"0baf8575-26cf-4a72-8a6d-ae94c078eaad\") " pod="openshift-multus/multus-k4wr2" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.350294 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/0baf8575-26cf-4a72-8a6d-ae94c078eaad-host-run-multus-certs\") pod \"multus-k4wr2\" (UID: \"0baf8575-26cf-4a72-8a6d-ae94c078eaad\") " pod="openshift-multus/multus-k4wr2" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.350159 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.350339 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:40Z","lastTransitionTime":"2025-12-07T16:02:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.350363 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/0baf8575-26cf-4a72-8a6d-ae94c078eaad-host-var-lib-kubelet\") pod \"multus-k4wr2\" (UID: \"0baf8575-26cf-4a72-8a6d-ae94c078eaad\") " pod="openshift-multus/multus-k4wr2" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.350465 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/0baf8575-26cf-4a72-8a6d-ae94c078eaad-multus-socket-dir-parent\") pod \"multus-k4wr2\" (UID: \"0baf8575-26cf-4a72-8a6d-ae94c078eaad\") " pod="openshift-multus/multus-k4wr2" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.350488 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/0baf8575-26cf-4a72-8a6d-ae94c078eaad-os-release\") pod \"multus-k4wr2\" (UID: \"0baf8575-26cf-4a72-8a6d-ae94c078eaad\") " pod="openshift-multus/multus-k4wr2" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.350505 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/98015cea-d753-4e7a-b2b5-5ae7733a81e3-cnibin\") pod \"multus-additional-cni-plugins-8ps9j\" (UID: \"98015cea-d753-4e7a-b2b5-5ae7733a81e3\") " pod="openshift-multus/multus-additional-cni-plugins-8ps9j" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.350530 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-llw85\" (UniqueName: \"kubernetes.io/projected/98015cea-d753-4e7a-b2b5-5ae7733a81e3-kube-api-access-llw85\") pod \"multus-additional-cni-plugins-8ps9j\" (UID: \"98015cea-d753-4e7a-b2b5-5ae7733a81e3\") " pod="openshift-multus/multus-additional-cni-plugins-8ps9j" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.350549 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/0baf8575-26cf-4a72-8a6d-ae94c078eaad-hostroot\") pod \"multus-k4wr2\" (UID: \"0baf8575-26cf-4a72-8a6d-ae94c078eaad\") " pod="openshift-multus/multus-k4wr2" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.350566 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/98015cea-d753-4e7a-b2b5-5ae7733a81e3-cni-binary-copy\") pod \"multus-additional-cni-plugins-8ps9j\" (UID: \"98015cea-d753-4e7a-b2b5-5ae7733a81e3\") " pod="openshift-multus/multus-additional-cni-plugins-8ps9j" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.350584 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0baf8575-26cf-4a72-8a6d-ae94c078eaad-host-var-lib-cni-bin\") pod \"multus-k4wr2\" (UID: \"0baf8575-26cf-4a72-8a6d-ae94c078eaad\") " pod="openshift-multus/multus-k4wr2" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.350635 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/0baf8575-26cf-4a72-8a6d-ae94c078eaad-system-cni-dir\") pod \"multus-k4wr2\" (UID: \"0baf8575-26cf-4a72-8a6d-ae94c078eaad\") " pod="openshift-multus/multus-k4wr2" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.350655 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/0baf8575-26cf-4a72-8a6d-ae94c078eaad-multus-cni-dir\") pod \"multus-k4wr2\" (UID: \"0baf8575-26cf-4a72-8a6d-ae94c078eaad\") " pod="openshift-multus/multus-k4wr2" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.350690 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/0baf8575-26cf-4a72-8a6d-ae94c078eaad-host-run-k8s-cni-cncf-io\") pod \"multus-k4wr2\" (UID: \"0baf8575-26cf-4a72-8a6d-ae94c078eaad\") " pod="openshift-multus/multus-k4wr2" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.350707 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/0baf8575-26cf-4a72-8a6d-ae94c078eaad-multus-conf-dir\") pod \"multus-k4wr2\" (UID: \"0baf8575-26cf-4a72-8a6d-ae94c078eaad\") " pod="openshift-multus/multus-k4wr2" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.350725 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/98015cea-d753-4e7a-b2b5-5ae7733a81e3-os-release\") pod \"multus-additional-cni-plugins-8ps9j\" (UID: \"98015cea-d753-4e7a-b2b5-5ae7733a81e3\") " pod="openshift-multus/multus-additional-cni-plugins-8ps9j" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.350742 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0baf8575-26cf-4a72-8a6d-ae94c078eaad-etc-kubernetes\") pod \"multus-k4wr2\" (UID: \"0baf8575-26cf-4a72-8a6d-ae94c078eaad\") " pod="openshift-multus/multus-k4wr2" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.350779 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p2trd\" (UniqueName: \"kubernetes.io/projected/0baf8575-26cf-4a72-8a6d-ae94c078eaad-kube-api-access-p2trd\") pod \"multus-k4wr2\" (UID: \"0baf8575-26cf-4a72-8a6d-ae94c078eaad\") " pod="openshift-multus/multus-k4wr2" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.350797 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/0baf8575-26cf-4a72-8a6d-ae94c078eaad-cnibin\") pod \"multus-k4wr2\" (UID: \"0baf8575-26cf-4a72-8a6d-ae94c078eaad\") " pod="openshift-multus/multus-k4wr2" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.361665 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c15b59eb-565d-4556-a4ce-75afdf159dc8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jks8z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jks8z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zcxxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:40Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.379917 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdee9957-48b9-458f-ba71-e0cd39f80596\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73f85248b4d68a9beecdacb3a01f433e97a0ffea2247ce92a41ef96855509377\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6276df652efe6ca935e745f24ef20f87a16465475139230c63b9999234a4d12a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3b1dd202586ee755ca89fbe12b30bbb12f95a62514f6e4d93cbbdbe55181e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd318647816e4a27849666b0ec4fc2ca4905295661c557dc381e1144f22cf284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d31de0326a05216fa10442bdb47151c6e073c97e9a16c34e793028e5e253756\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"le observer\\\\nW1207 16:02:34.645464 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1207 16:02:34.645587 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 16:02:34.646501 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2082283479/tls.crt::/tmp/serving-cert-2082283479/tls.key\\\\\\\"\\\\nI1207 16:02:34.972041 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 16:02:34.973702 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 16:02:34.973717 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 16:02:34.973738 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 16:02:34.973743 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 16:02:34.977185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1207 16:02:34.977204 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1207 16:02:34.977209 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977217 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 16:02:34.977225 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 16:02:34.977228 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 16:02:34.977232 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1207 16:02:34.979033 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27bd4a54f920b2725987c1c7fd2d9fdf16cd08738c246c333a4e9164a3a582a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:40Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.393965 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:40Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.407538 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2jjld" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dee6f5c-861c-4819-9309-7791574c7b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmdns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2jjld\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:40Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.419318 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4wr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0baf8575-26cf-4a72-8a6d-ae94c078eaad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p2trd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4wr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:40Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.430035 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:40Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.442717 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f95ef61779507288a03bda8eddf745e105f705998a117a30789675bfa452d88f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b955b23163e7c8c458a3bbe2f681633e6a663ce76748febcc0b3a3c37f3ca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:40Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.451696 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/0baf8575-26cf-4a72-8a6d-ae94c078eaad-cnibin\") pod \"multus-k4wr2\" (UID: \"0baf8575-26cf-4a72-8a6d-ae94c078eaad\") " pod="openshift-multus/multus-k4wr2" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.451747 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/98015cea-d753-4e7a-b2b5-5ae7733a81e3-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-8ps9j\" (UID: \"98015cea-d753-4e7a-b2b5-5ae7733a81e3\") " pod="openshift-multus/multus-additional-cni-plugins-8ps9j" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.451769 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0baf8575-26cf-4a72-8a6d-ae94c078eaad-host-run-netns\") pod \"multus-k4wr2\" (UID: \"0baf8575-26cf-4a72-8a6d-ae94c078eaad\") " pod="openshift-multus/multus-k4wr2" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.451794 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/0baf8575-26cf-4a72-8a6d-ae94c078eaad-multus-daemon-config\") pod \"multus-k4wr2\" (UID: \"0baf8575-26cf-4a72-8a6d-ae94c078eaad\") " pod="openshift-multus/multus-k4wr2" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.451800 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/0baf8575-26cf-4a72-8a6d-ae94c078eaad-cnibin\") pod \"multus-k4wr2\" (UID: \"0baf8575-26cf-4a72-8a6d-ae94c078eaad\") " pod="openshift-multus/multus-k4wr2" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.451819 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/0baf8575-26cf-4a72-8a6d-ae94c078eaad-host-var-lib-cni-multus\") pod \"multus-k4wr2\" (UID: \"0baf8575-26cf-4a72-8a6d-ae94c078eaad\") " pod="openshift-multus/multus-k4wr2" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.451844 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/98015cea-d753-4e7a-b2b5-5ae7733a81e3-system-cni-dir\") pod \"multus-additional-cni-plugins-8ps9j\" (UID: \"98015cea-d753-4e7a-b2b5-5ae7733a81e3\") " pod="openshift-multus/multus-additional-cni-plugins-8ps9j" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.451859 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/0baf8575-26cf-4a72-8a6d-ae94c078eaad-host-run-multus-certs\") pod \"multus-k4wr2\" (UID: \"0baf8575-26cf-4a72-8a6d-ae94c078eaad\") " pod="openshift-multus/multus-k4wr2" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.451875 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/98015cea-d753-4e7a-b2b5-5ae7733a81e3-tuning-conf-dir\") pod \"multus-additional-cni-plugins-8ps9j\" (UID: \"98015cea-d753-4e7a-b2b5-5ae7733a81e3\") " pod="openshift-multus/multus-additional-cni-plugins-8ps9j" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.451892 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/0baf8575-26cf-4a72-8a6d-ae94c078eaad-cni-binary-copy\") pod \"multus-k4wr2\" (UID: \"0baf8575-26cf-4a72-8a6d-ae94c078eaad\") " pod="openshift-multus/multus-k4wr2" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.451894 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/0baf8575-26cf-4a72-8a6d-ae94c078eaad-host-var-lib-cni-multus\") pod \"multus-k4wr2\" (UID: \"0baf8575-26cf-4a72-8a6d-ae94c078eaad\") " pod="openshift-multus/multus-k4wr2" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.452352 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/98015cea-d753-4e7a-b2b5-5ae7733a81e3-tuning-conf-dir\") pod \"multus-additional-cni-plugins-8ps9j\" (UID: \"98015cea-d753-4e7a-b2b5-5ae7733a81e3\") " pod="openshift-multus/multus-additional-cni-plugins-8ps9j" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.452707 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/98015cea-d753-4e7a-b2b5-5ae7733a81e3-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-8ps9j\" (UID: \"98015cea-d753-4e7a-b2b5-5ae7733a81e3\") " pod="openshift-multus/multus-additional-cni-plugins-8ps9j" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.451894 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0baf8575-26cf-4a72-8a6d-ae94c078eaad-host-run-netns\") pod \"multus-k4wr2\" (UID: \"0baf8575-26cf-4a72-8a6d-ae94c078eaad\") " pod="openshift-multus/multus-k4wr2" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.451935 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/0baf8575-26cf-4a72-8a6d-ae94c078eaad-host-run-multus-certs\") pod \"multus-k4wr2\" (UID: \"0baf8575-26cf-4a72-8a6d-ae94c078eaad\") " pod="openshift-multus/multus-k4wr2" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.451983 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/98015cea-d753-4e7a-b2b5-5ae7733a81e3-system-cni-dir\") pod \"multus-additional-cni-plugins-8ps9j\" (UID: \"98015cea-d753-4e7a-b2b5-5ae7733a81e3\") " pod="openshift-multus/multus-additional-cni-plugins-8ps9j" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.452720 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/0baf8575-26cf-4a72-8a6d-ae94c078eaad-cni-binary-copy\") pod \"multus-k4wr2\" (UID: \"0baf8575-26cf-4a72-8a6d-ae94c078eaad\") " pod="openshift-multus/multus-k4wr2" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.451965 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/0baf8575-26cf-4a72-8a6d-ae94c078eaad-multus-socket-dir-parent\") pod \"multus-k4wr2\" (UID: \"0baf8575-26cf-4a72-8a6d-ae94c078eaad\") " pod="openshift-multus/multus-k4wr2" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.451906 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/0baf8575-26cf-4a72-8a6d-ae94c078eaad-multus-socket-dir-parent\") pod \"multus-k4wr2\" (UID: \"0baf8575-26cf-4a72-8a6d-ae94c078eaad\") " pod="openshift-multus/multus-k4wr2" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.452877 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/0baf8575-26cf-4a72-8a6d-ae94c078eaad-host-var-lib-kubelet\") pod \"multus-k4wr2\" (UID: \"0baf8575-26cf-4a72-8a6d-ae94c078eaad\") " pod="openshift-multus/multus-k4wr2" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.452943 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/0baf8575-26cf-4a72-8a6d-ae94c078eaad-os-release\") pod \"multus-k4wr2\" (UID: \"0baf8575-26cf-4a72-8a6d-ae94c078eaad\") " pod="openshift-multus/multus-k4wr2" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.452968 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/98015cea-d753-4e7a-b2b5-5ae7733a81e3-cnibin\") pod \"multus-additional-cni-plugins-8ps9j\" (UID: \"98015cea-d753-4e7a-b2b5-5ae7733a81e3\") " pod="openshift-multus/multus-additional-cni-plugins-8ps9j" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.453004 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/98015cea-d753-4e7a-b2b5-5ae7733a81e3-cni-binary-copy\") pod \"multus-additional-cni-plugins-8ps9j\" (UID: \"98015cea-d753-4e7a-b2b5-5ae7733a81e3\") " pod="openshift-multus/multus-additional-cni-plugins-8ps9j" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.453033 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/98015cea-d753-4e7a-b2b5-5ae7733a81e3-cnibin\") pod \"multus-additional-cni-plugins-8ps9j\" (UID: \"98015cea-d753-4e7a-b2b5-5ae7733a81e3\") " pod="openshift-multus/multus-additional-cni-plugins-8ps9j" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.453050 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/0baf8575-26cf-4a72-8a6d-ae94c078eaad-os-release\") pod \"multus-k4wr2\" (UID: \"0baf8575-26cf-4a72-8a6d-ae94c078eaad\") " pod="openshift-multus/multus-k4wr2" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.453030 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-llw85\" (UniqueName: \"kubernetes.io/projected/98015cea-d753-4e7a-b2b5-5ae7733a81e3-kube-api-access-llw85\") pod \"multus-additional-cni-plugins-8ps9j\" (UID: \"98015cea-d753-4e7a-b2b5-5ae7733a81e3\") " pod="openshift-multus/multus-additional-cni-plugins-8ps9j" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.453122 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/0baf8575-26cf-4a72-8a6d-ae94c078eaad-hostroot\") pod \"multus-k4wr2\" (UID: \"0baf8575-26cf-4a72-8a6d-ae94c078eaad\") " pod="openshift-multus/multus-k4wr2" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.453150 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0baf8575-26cf-4a72-8a6d-ae94c078eaad-host-var-lib-cni-bin\") pod \"multus-k4wr2\" (UID: \"0baf8575-26cf-4a72-8a6d-ae94c078eaad\") " pod="openshift-multus/multus-k4wr2" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.453176 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/0baf8575-26cf-4a72-8a6d-ae94c078eaad-system-cni-dir\") pod \"multus-k4wr2\" (UID: \"0baf8575-26cf-4a72-8a6d-ae94c078eaad\") " pod="openshift-multus/multus-k4wr2" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.453199 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/0baf8575-26cf-4a72-8a6d-ae94c078eaad-multus-conf-dir\") pod \"multus-k4wr2\" (UID: \"0baf8575-26cf-4a72-8a6d-ae94c078eaad\") " pod="openshift-multus/multus-k4wr2" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.453220 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/0baf8575-26cf-4a72-8a6d-ae94c078eaad-multus-cni-dir\") pod \"multus-k4wr2\" (UID: \"0baf8575-26cf-4a72-8a6d-ae94c078eaad\") " pod="openshift-multus/multus-k4wr2" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.453238 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/0baf8575-26cf-4a72-8a6d-ae94c078eaad-host-run-k8s-cni-cncf-io\") pod \"multus-k4wr2\" (UID: \"0baf8575-26cf-4a72-8a6d-ae94c078eaad\") " pod="openshift-multus/multus-k4wr2" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.453260 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/98015cea-d753-4e7a-b2b5-5ae7733a81e3-os-release\") pod \"multus-additional-cni-plugins-8ps9j\" (UID: \"98015cea-d753-4e7a-b2b5-5ae7733a81e3\") " pod="openshift-multus/multus-additional-cni-plugins-8ps9j" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.453282 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0baf8575-26cf-4a72-8a6d-ae94c078eaad-etc-kubernetes\") pod \"multus-k4wr2\" (UID: \"0baf8575-26cf-4a72-8a6d-ae94c078eaad\") " pod="openshift-multus/multus-k4wr2" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.453305 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p2trd\" (UniqueName: \"kubernetes.io/projected/0baf8575-26cf-4a72-8a6d-ae94c078eaad-kube-api-access-p2trd\") pod \"multus-k4wr2\" (UID: \"0baf8575-26cf-4a72-8a6d-ae94c078eaad\") " pod="openshift-multus/multus-k4wr2" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.453316 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/0baf8575-26cf-4a72-8a6d-ae94c078eaad-multus-conf-dir\") pod \"multus-k4wr2\" (UID: \"0baf8575-26cf-4a72-8a6d-ae94c078eaad\") " pod="openshift-multus/multus-k4wr2" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.453353 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/0baf8575-26cf-4a72-8a6d-ae94c078eaad-hostroot\") pod \"multus-k4wr2\" (UID: \"0baf8575-26cf-4a72-8a6d-ae94c078eaad\") " pod="openshift-multus/multus-k4wr2" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.453383 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0baf8575-26cf-4a72-8a6d-ae94c078eaad-host-var-lib-cni-bin\") pod \"multus-k4wr2\" (UID: \"0baf8575-26cf-4a72-8a6d-ae94c078eaad\") " pod="openshift-multus/multus-k4wr2" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.453507 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/0baf8575-26cf-4a72-8a6d-ae94c078eaad-system-cni-dir\") pod \"multus-k4wr2\" (UID: \"0baf8575-26cf-4a72-8a6d-ae94c078eaad\") " pod="openshift-multus/multus-k4wr2" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.453556 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/98015cea-d753-4e7a-b2b5-5ae7733a81e3-os-release\") pod \"multus-additional-cni-plugins-8ps9j\" (UID: \"98015cea-d753-4e7a-b2b5-5ae7733a81e3\") " pod="openshift-multus/multus-additional-cni-plugins-8ps9j" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.453581 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/0baf8575-26cf-4a72-8a6d-ae94c078eaad-host-run-k8s-cni-cncf-io\") pod \"multus-k4wr2\" (UID: \"0baf8575-26cf-4a72-8a6d-ae94c078eaad\") " pod="openshift-multus/multus-k4wr2" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.453005 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/0baf8575-26cf-4a72-8a6d-ae94c078eaad-host-var-lib-kubelet\") pod \"multus-k4wr2\" (UID: \"0baf8575-26cf-4a72-8a6d-ae94c078eaad\") " pod="openshift-multus/multus-k4wr2" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.453590 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/98015cea-d753-4e7a-b2b5-5ae7733a81e3-cni-binary-copy\") pod \"multus-additional-cni-plugins-8ps9j\" (UID: \"98015cea-d753-4e7a-b2b5-5ae7733a81e3\") " pod="openshift-multus/multus-additional-cni-plugins-8ps9j" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.453621 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0baf8575-26cf-4a72-8a6d-ae94c078eaad-etc-kubernetes\") pod \"multus-k4wr2\" (UID: \"0baf8575-26cf-4a72-8a6d-ae94c078eaad\") " pod="openshift-multus/multus-k4wr2" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.453620 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/0baf8575-26cf-4a72-8a6d-ae94c078eaad-multus-cni-dir\") pod \"multus-k4wr2\" (UID: \"0baf8575-26cf-4a72-8a6d-ae94c078eaad\") " pod="openshift-multus/multus-k4wr2" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.453660 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/0baf8575-26cf-4a72-8a6d-ae94c078eaad-multus-daemon-config\") pod \"multus-k4wr2\" (UID: \"0baf8575-26cf-4a72-8a6d-ae94c078eaad\") " pod="openshift-multus/multus-k4wr2" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.455334 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.455383 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.455404 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.455433 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.455453 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:40Z","lastTransitionTime":"2025-12-07T16:02:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.455657 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef3335ae3875a75ab64685fc7e3590177280d651e8f8292c6f5ea6f472e55789\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:40Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.468469 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p2trd\" (UniqueName: \"kubernetes.io/projected/0baf8575-26cf-4a72-8a6d-ae94c078eaad-kube-api-access-p2trd\") pod \"multus-k4wr2\" (UID: \"0baf8575-26cf-4a72-8a6d-ae94c078eaad\") " pod="openshift-multus/multus-k4wr2" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.470430 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-llw85\" (UniqueName: \"kubernetes.io/projected/98015cea-d753-4e7a-b2b5-5ae7733a81e3-kube-api-access-llw85\") pod \"multus-additional-cni-plugins-8ps9j\" (UID: \"98015cea-d753-4e7a-b2b5-5ae7733a81e3\") " pod="openshift-multus/multus-additional-cni-plugins-8ps9j" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.472151 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4742ea84b3a2c9461d9b900beee7354a2ded24c39ad23e007cbf270ae12377c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:40Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.483995 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:40Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.494136 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c15b59eb-565d-4556-a4ce-75afdf159dc8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jks8z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jks8z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zcxxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:40Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.509289 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8ps9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98015cea-d753-4e7a-b2b5-5ae7733a81e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8ps9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:40Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.520937 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2jjld" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dee6f5c-861c-4819-9309-7791574c7b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmdns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2jjld\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:40Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.532990 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4wr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0baf8575-26cf-4a72-8a6d-ae94c078eaad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p2trd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4wr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:40Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.535963 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-k4wr2" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.540336 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-8ps9j" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.547800 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdee9957-48b9-458f-ba71-e0cd39f80596\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73f85248b4d68a9beecdacb3a01f433e97a0ffea2247ce92a41ef96855509377\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6276df652efe6ca935e745f24ef20f87a16465475139230c63b9999234a4d12a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3b1dd202586ee755ca89fbe12b30bbb12f95a62514f6e4d93cbbdbe55181e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd318647816e4a27849666b0ec4fc2ca4905295661c557dc381e1144f22cf284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d31de0326a05216fa10442bdb47151c6e073c97e9a16c34e793028e5e253756\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"le observer\\\\nW1207 16:02:34.645464 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1207 16:02:34.645587 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 16:02:34.646501 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2082283479/tls.crt::/tmp/serving-cert-2082283479/tls.key\\\\\\\"\\\\nI1207 16:02:34.972041 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 16:02:34.973702 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 16:02:34.973717 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 16:02:34.973738 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 16:02:34.973743 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 16:02:34.977185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1207 16:02:34.977204 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1207 16:02:34.977209 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977217 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 16:02:34.977225 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 16:02:34.977228 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 16:02:34.977232 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1207 16:02:34.979033 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27bd4a54f920b2725987c1c7fd2d9fdf16cd08738c246c333a4e9164a3a582a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:40Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:40 crc kubenswrapper[4716]: W1207 16:02:40.549650 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0baf8575_26cf_4a72_8a6d_ae94c078eaad.slice/crio-dadb075180d7076d20bf401a8631e46249aece3db6b134f2947ab868e12cb712 WatchSource:0}: Error finding container dadb075180d7076d20bf401a8631e46249aece3db6b134f2947ab868e12cb712: Status 404 returned error can't find the container with id dadb075180d7076d20bf401a8631e46249aece3db6b134f2947ab868e12cb712 Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.560099 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.560129 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.560139 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.560153 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.560163 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:40Z","lastTransitionTime":"2025-12-07T16:02:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.562252 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:40Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.618885 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-qvc7p"] Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.626691 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.628225 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.629873 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.632148 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.632207 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.632207 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.632410 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.632794 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.637570 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.642144 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.647260 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.647792 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdee9957-48b9-458f-ba71-e0cd39f80596\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73f85248b4d68a9beecdacb3a01f433e97a0ffea2247ce92a41ef96855509377\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6276df652efe6ca935e745f24ef20f87a16465475139230c63b9999234a4d12a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3b1dd202586ee755ca89fbe12b30bbb12f95a62514f6e4d93cbbdbe55181e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd318647816e4a27849666b0ec4fc2ca4905295661c557dc381e1144f22cf284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d31de0326a05216fa10442bdb47151c6e073c97e9a16c34e793028e5e253756\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"le observer\\\\nW1207 16:02:34.645464 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1207 16:02:34.645587 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 16:02:34.646501 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2082283479/tls.crt::/tmp/serving-cert-2082283479/tls.key\\\\\\\"\\\\nI1207 16:02:34.972041 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 16:02:34.973702 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 16:02:34.973717 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 16:02:34.973738 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 16:02:34.973743 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 16:02:34.977185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1207 16:02:34.977204 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1207 16:02:34.977209 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977217 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 16:02:34.977225 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 16:02:34.977228 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 16:02:34.977232 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1207 16:02:34.979033 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27bd4a54f920b2725987c1c7fd2d9fdf16cd08738c246c333a4e9164a3a582a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:40Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.653051 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.658614 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 16:02:40 crc kubenswrapper[4716]: E1207 16:02:40.658761 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.660416 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.662105 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:40Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.662428 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.662450 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.662461 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.662473 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.662483 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:40Z","lastTransitionTime":"2025-12-07T16:02:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.672224 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2jjld" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dee6f5c-861c-4819-9309-7791574c7b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmdns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2jjld\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:40Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.696794 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4wr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0baf8575-26cf-4a72-8a6d-ae94c078eaad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p2trd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4wr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:40Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.710479 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef3335ae3875a75ab64685fc7e3590177280d651e8f8292c6f5ea6f472e55789\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:40Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.724183 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:40Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.736855 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f95ef61779507288a03bda8eddf745e105f705998a117a30789675bfa452d88f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b955b23163e7c8c458a3bbe2f681633e6a663ce76748febcc0b3a3c37f3ca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:40Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.740589 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.749739 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8ps9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98015cea-d753-4e7a-b2b5-5ae7733a81e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8ps9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:40Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.756437 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-host-run-netns\") pod \"ovnkube-node-qvc7p\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.756472 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-host-run-ovn-kubernetes\") pod \"ovnkube-node-qvc7p\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.756490 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-etc-openvswitch\") pod \"ovnkube-node-qvc7p\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.756508 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c54d618a-dfec-4f2a-a288-f5188c8a128a-ovn-node-metrics-cert\") pod \"ovnkube-node-qvc7p\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.756524 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-run-systemd\") pod \"ovnkube-node-qvc7p\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.756539 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-log-socket\") pod \"ovnkube-node-qvc7p\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.756619 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c54d618a-dfec-4f2a-a288-f5188c8a128a-ovnkube-script-lib\") pod \"ovnkube-node-qvc7p\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.756679 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-host-kubelet\") pod \"ovnkube-node-qvc7p\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.756732 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-var-lib-openvswitch\") pod \"ovnkube-node-qvc7p\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.756757 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-qvc7p\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.756781 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-run-ovn\") pod \"ovnkube-node-qvc7p\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.756796 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-host-slash\") pod \"ovnkube-node-qvc7p\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.756810 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-run-openvswitch\") pod \"ovnkube-node-qvc7p\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.756851 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-systemd-units\") pod \"ovnkube-node-qvc7p\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.756868 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c54d618a-dfec-4f2a-a288-f5188c8a128a-env-overrides\") pod \"ovnkube-node-qvc7p\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.756885 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2qphz\" (UniqueName: \"kubernetes.io/projected/c54d618a-dfec-4f2a-a288-f5188c8a128a-kube-api-access-2qphz\") pod \"ovnkube-node-qvc7p\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.756907 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-node-log\") pod \"ovnkube-node-qvc7p\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.756923 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-host-cni-netd\") pod \"ovnkube-node-qvc7p\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.756949 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-host-cni-bin\") pod \"ovnkube-node-qvc7p\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.756965 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c54d618a-dfec-4f2a-a288-f5188c8a128a-ovnkube-config\") pod \"ovnkube-node-qvc7p\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.761579 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4742ea84b3a2c9461d9b900beee7354a2ded24c39ad23e007cbf270ae12377c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:40Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.765712 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.765738 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.765747 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.765760 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.765773 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:40Z","lastTransitionTime":"2025-12-07T16:02:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.775213 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:40Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.784875 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c15b59eb-565d-4556-a4ce-75afdf159dc8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jks8z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jks8z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zcxxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:40Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.803426 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c54d618a-dfec-4f2a-a288-f5188c8a128a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qvc7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:40Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.815304 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-8ps9j" event={"ID":"98015cea-d753-4e7a-b2b5-5ae7733a81e3","Type":"ContainerStarted","Data":"aa3520377e9653f6de23668f92a23f7e4cc7a818cba3c93760e23c04ad8e76e8"} Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.817941 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-k4wr2" event={"ID":"0baf8575-26cf-4a72-8a6d-ae94c078eaad","Type":"ContainerStarted","Data":"dadb075180d7076d20bf401a8631e46249aece3db6b134f2947ab868e12cb712"} Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.820748 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-2jjld" event={"ID":"3dee6f5c-861c-4819-9309-7791574c7b80","Type":"ContainerStarted","Data":"51c1d4159ab11087839a2c2d192566fb9322447c26df560bf09517c906ececb1"} Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.823748 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" event={"ID":"c15b59eb-565d-4556-a4ce-75afdf159dc8","Type":"ContainerStarted","Data":"530a2935826940767ea33fb4f37fa11f7dee4950d867fbcd6af72836d80f7db3"} Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.823778 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" event={"ID":"c15b59eb-565d-4556-a4ce-75afdf159dc8","Type":"ContainerStarted","Data":"8ccaf6c1e6a978eb978a1c207b6c4c00b23737f78b743f966579aca781e22e0b"} Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.823795 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" event={"ID":"c15b59eb-565d-4556-a4ce-75afdf159dc8","Type":"ContainerStarted","Data":"390c0fff32d3270bc2d944de6b3abd6b6150de4c91e750d472c5ded6f05ae3ea"} Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.823953 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c54d618a-dfec-4f2a-a288-f5188c8a128a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qvc7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:40Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:40 crc kubenswrapper[4716]: E1207 16:02:40.828388 4716 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.841375 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdee9957-48b9-458f-ba71-e0cd39f80596\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73f85248b4d68a9beecdacb3a01f433e97a0ffea2247ce92a41ef96855509377\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6276df652efe6ca935e745f24ef20f87a16465475139230c63b9999234a4d12a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3b1dd202586ee755ca89fbe12b30bbb12f95a62514f6e4d93cbbdbe55181e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd318647816e4a27849666b0ec4fc2ca4905295661c557dc381e1144f22cf284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d31de0326a05216fa10442bdb47151c6e073c97e9a16c34e793028e5e253756\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"le observer\\\\nW1207 16:02:34.645464 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1207 16:02:34.645587 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 16:02:34.646501 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2082283479/tls.crt::/tmp/serving-cert-2082283479/tls.key\\\\\\\"\\\\nI1207 16:02:34.972041 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 16:02:34.973702 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 16:02:34.973717 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 16:02:34.973738 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 16:02:34.973743 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 16:02:34.977185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1207 16:02:34.977204 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1207 16:02:34.977209 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977217 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 16:02:34.977225 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 16:02:34.977228 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 16:02:34.977232 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1207 16:02:34.979033 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27bd4a54f920b2725987c1c7fd2d9fdf16cd08738c246c333a4e9164a3a582a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:40Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.853906 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:40Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.858466 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-etc-openvswitch\") pod \"ovnkube-node-qvc7p\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.858507 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c54d618a-dfec-4f2a-a288-f5188c8a128a-ovn-node-metrics-cert\") pod \"ovnkube-node-qvc7p\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.858528 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-run-systemd\") pod \"ovnkube-node-qvc7p\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.858545 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-log-socket\") pod \"ovnkube-node-qvc7p\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.858564 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c54d618a-dfec-4f2a-a288-f5188c8a128a-ovnkube-script-lib\") pod \"ovnkube-node-qvc7p\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.858579 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-host-kubelet\") pod \"ovnkube-node-qvc7p\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.858586 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-etc-openvswitch\") pod \"ovnkube-node-qvc7p\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.858595 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-var-lib-openvswitch\") pod \"ovnkube-node-qvc7p\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.858643 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-var-lib-openvswitch\") pod \"ovnkube-node-qvc7p\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.858650 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-qvc7p\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.858686 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-log-socket\") pod \"ovnkube-node-qvc7p\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.858707 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-run-systemd\") pod \"ovnkube-node-qvc7p\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.858728 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-run-ovn\") pod \"ovnkube-node-qvc7p\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.858768 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-host-slash\") pod \"ovnkube-node-qvc7p\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.858763 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-host-kubelet\") pod \"ovnkube-node-qvc7p\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.858808 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-run-openvswitch\") pod \"ovnkube-node-qvc7p\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.858836 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-host-slash\") pod \"ovnkube-node-qvc7p\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.858703 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-qvc7p\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.858839 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2qphz\" (UniqueName: \"kubernetes.io/projected/c54d618a-dfec-4f2a-a288-f5188c8a128a-kube-api-access-2qphz\") pod \"ovnkube-node-qvc7p\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.858905 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-systemd-units\") pod \"ovnkube-node-qvc7p\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.858922 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c54d618a-dfec-4f2a-a288-f5188c8a128a-env-overrides\") pod \"ovnkube-node-qvc7p\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.858933 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-run-openvswitch\") pod \"ovnkube-node-qvc7p\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.858942 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-node-log\") pod \"ovnkube-node-qvc7p\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.858962 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-node-log\") pod \"ovnkube-node-qvc7p\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.858979 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-host-cni-netd\") pod \"ovnkube-node-qvc7p\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.859012 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-host-cni-bin\") pod \"ovnkube-node-qvc7p\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.858809 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-run-ovn\") pod \"ovnkube-node-qvc7p\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.859029 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c54d618a-dfec-4f2a-a288-f5188c8a128a-ovnkube-config\") pod \"ovnkube-node-qvc7p\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.859059 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-host-run-netns\") pod \"ovnkube-node-qvc7p\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.859095 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-host-run-ovn-kubernetes\") pod \"ovnkube-node-qvc7p\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.859173 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-host-run-ovn-kubernetes\") pod \"ovnkube-node-qvc7p\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.859198 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-host-run-netns\") pod \"ovnkube-node-qvc7p\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.858984 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-systemd-units\") pod \"ovnkube-node-qvc7p\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.859235 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-host-cni-netd\") pod \"ovnkube-node-qvc7p\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.859294 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-host-cni-bin\") pod \"ovnkube-node-qvc7p\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.859578 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c54d618a-dfec-4f2a-a288-f5188c8a128a-env-overrides\") pod \"ovnkube-node-qvc7p\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.859759 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c54d618a-dfec-4f2a-a288-f5188c8a128a-ovnkube-config\") pod \"ovnkube-node-qvc7p\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.859801 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c54d618a-dfec-4f2a-a288-f5188c8a128a-ovnkube-script-lib\") pod \"ovnkube-node-qvc7p\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.863474 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2jjld" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dee6f5c-861c-4819-9309-7791574c7b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmdns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2jjld\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:40Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.864981 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c54d618a-dfec-4f2a-a288-f5188c8a128a-ovn-node-metrics-cert\") pod \"ovnkube-node-qvc7p\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.868387 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.868428 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.868442 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.868465 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.868476 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:40Z","lastTransitionTime":"2025-12-07T16:02:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.876063 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2qphz\" (UniqueName: \"kubernetes.io/projected/c54d618a-dfec-4f2a-a288-f5188c8a128a-kube-api-access-2qphz\") pod \"ovnkube-node-qvc7p\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.877018 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4wr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0baf8575-26cf-4a72-8a6d-ae94c078eaad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p2trd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4wr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:40Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.890169 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef3335ae3875a75ab64685fc7e3590177280d651e8f8292c6f5ea6f472e55789\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:40Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.900924 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:40Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.912425 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f95ef61779507288a03bda8eddf745e105f705998a117a30789675bfa452d88f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b955b23163e7c8c458a3bbe2f681633e6a663ce76748febcc0b3a3c37f3ca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:40Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.926355 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8ps9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98015cea-d753-4e7a-b2b5-5ae7733a81e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8ps9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:40Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.945247 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7961cc9a-734e-42c2-b1ff-7075f3f20091\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84a36b76f20618c7885d611ecad9db29bac5101c27572be31d3acbb90c581291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0a0eca3c2087a457c0e448e649345ec4595a9c8a019aed2cddc5f30c77871e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://404426a171b8676eddd4a00996c76afab7e269643fa9608f1e1b6bdba385071d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fd3c2941782921f99a6cdc3de4e892234548bde527fec15a8ac92e513566e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e23abfdca314ed43d594333c9f9e401cc8ef7dad39b8ebac84a836b06ee5c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09b38cd4b389ea477c802a842ee4f2297a64d64684624117163e819539c257c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09b38cd4b389ea477c802a842ee4f2297a64d64684624117163e819539c257c6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c00f7a8e024e0e0dfd7b73e16d1d5437602e79ea95b73c21f24b209edef60e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c00f7a8e024e0e0dfd7b73e16d1d5437602e79ea95b73c21f24b209edef60e98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3f6392d426fa2cd788ec3a5c3bff4ac39483a54dfb2624a1c4370522413a350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3f6392d426fa2cd788ec3a5c3bff4ac39483a54dfb2624a1c4370522413a350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:40Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.967286 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3067de7c-c11c-4be1-bd94-d6572c266795\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d14d018fa044e9e80fe81e7e38df935992d14c773ed2b2193fd5cb873e5d0947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a9e9cf048220ca14c996847e9ced9f2367b9331dc6c2836dd023658855cd79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6d3c911874095b3c7cb062d03070e5090a7bb2456cb20cb4ce078f502515cda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ac6ce7701f1c6aa1ea8fcf7a3035e3b0eaa28ae3cb6afc442372bd9117bdde4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:40Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.971710 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.971752 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.971763 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.971780 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.971793 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:40Z","lastTransitionTime":"2025-12-07T16:02:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.978860 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4742ea84b3a2c9461d9b900beee7354a2ded24c39ad23e007cbf270ae12377c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:40Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:40 crc kubenswrapper[4716]: I1207 16:02:40.996184 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:40Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.007729 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c15b59eb-565d-4556-a4ce-75afdf159dc8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jks8z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jks8z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zcxxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:41Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.018100 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4742ea84b3a2c9461d9b900beee7354a2ded24c39ad23e007cbf270ae12377c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:41Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.029630 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:41Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.042730 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c15b59eb-565d-4556-a4ce-75afdf159dc8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://530a2935826940767ea33fb4f37fa11f7dee4950d867fbcd6af72836d80f7db3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jks8z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ccaf6c1e6a978eb978a1c207b6c4c00b23737f78b743f966579aca781e22e0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jks8z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zcxxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:41Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.054959 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8ps9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98015cea-d753-4e7a-b2b5-5ae7733a81e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8ps9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:41Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.075119 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.075166 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.075182 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.075204 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.075217 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:41Z","lastTransitionTime":"2025-12-07T16:02:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.079573 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7961cc9a-734e-42c2-b1ff-7075f3f20091\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84a36b76f20618c7885d611ecad9db29bac5101c27572be31d3acbb90c581291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0a0eca3c2087a457c0e448e649345ec4595a9c8a019aed2cddc5f30c77871e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://404426a171b8676eddd4a00996c76afab7e269643fa9608f1e1b6bdba385071d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fd3c2941782921f99a6cdc3de4e892234548bde527fec15a8ac92e513566e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e23abfdca314ed43d594333c9f9e401cc8ef7dad39b8ebac84a836b06ee5c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09b38cd4b389ea477c802a842ee4f2297a64d64684624117163e819539c257c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09b38cd4b389ea477c802a842ee4f2297a64d64684624117163e819539c257c6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c00f7a8e024e0e0dfd7b73e16d1d5437602e79ea95b73c21f24b209edef60e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c00f7a8e024e0e0dfd7b73e16d1d5437602e79ea95b73c21f24b209edef60e98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3f6392d426fa2cd788ec3a5c3bff4ac39483a54dfb2624a1c4370522413a350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3f6392d426fa2cd788ec3a5c3bff4ac39483a54dfb2624a1c4370522413a350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:41Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.093612 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3067de7c-c11c-4be1-bd94-d6572c266795\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d14d018fa044e9e80fe81e7e38df935992d14c773ed2b2193fd5cb873e5d0947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a9e9cf048220ca14c996847e9ced9f2367b9331dc6c2836dd023658855cd79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6d3c911874095b3c7cb062d03070e5090a7bb2456cb20cb4ce078f502515cda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ac6ce7701f1c6aa1ea8fcf7a3035e3b0eaa28ae3cb6afc442372bd9117bdde4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:41Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.114748 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.117988 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c54d618a-dfec-4f2a-a288-f5188c8a128a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qvc7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:41Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.133017 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2jjld" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dee6f5c-861c-4819-9309-7791574c7b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51c1d4159ab11087839a2c2d192566fb9322447c26df560bf09517c906ececb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmdns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2jjld\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:41Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.148466 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4wr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0baf8575-26cf-4a72-8a6d-ae94c078eaad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p2trd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4wr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:41Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.167171 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdee9957-48b9-458f-ba71-e0cd39f80596\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73f85248b4d68a9beecdacb3a01f433e97a0ffea2247ce92a41ef96855509377\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6276df652efe6ca935e745f24ef20f87a16465475139230c63b9999234a4d12a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3b1dd202586ee755ca89fbe12b30bbb12f95a62514f6e4d93cbbdbe55181e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd318647816e4a27849666b0ec4fc2ca4905295661c557dc381e1144f22cf284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d31de0326a05216fa10442bdb47151c6e073c97e9a16c34e793028e5e253756\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"le observer\\\\nW1207 16:02:34.645464 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1207 16:02:34.645587 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 16:02:34.646501 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2082283479/tls.crt::/tmp/serving-cert-2082283479/tls.key\\\\\\\"\\\\nI1207 16:02:34.972041 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 16:02:34.973702 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 16:02:34.973717 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 16:02:34.973738 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 16:02:34.973743 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 16:02:34.977185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1207 16:02:34.977204 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1207 16:02:34.977209 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977217 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 16:02:34.977225 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 16:02:34.977228 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 16:02:34.977232 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1207 16:02:34.979033 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27bd4a54f920b2725987c1c7fd2d9fdf16cd08738c246c333a4e9164a3a582a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:41Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.177886 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.177917 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.177932 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.177950 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.177964 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:41Z","lastTransitionTime":"2025-12-07T16:02:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.183969 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:41Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.197007 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:41Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.231980 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f95ef61779507288a03bda8eddf745e105f705998a117a30789675bfa452d88f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b955b23163e7c8c458a3bbe2f681633e6a663ce76748febcc0b3a3c37f3ca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:41Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.267885 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef3335ae3875a75ab64685fc7e3590177280d651e8f8292c6f5ea6f472e55789\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:41Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.287128 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.287156 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.287165 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.287178 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.287187 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:41Z","lastTransitionTime":"2025-12-07T16:02:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.389798 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.389834 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.389843 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.389855 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.389863 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:41Z","lastTransitionTime":"2025-12-07T16:02:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.493153 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.493210 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.493229 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.493250 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.493266 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:41Z","lastTransitionTime":"2025-12-07T16:02:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.595341 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.595378 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.595387 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.595401 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.595412 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:41Z","lastTransitionTime":"2025-12-07T16:02:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.657696 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.657911 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:02:41 crc kubenswrapper[4716]: E1207 16:02:41.658330 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 16:02:41 crc kubenswrapper[4716]: E1207 16:02:41.658492 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.697476 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.697527 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.697540 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.697557 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.697571 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:41Z","lastTransitionTime":"2025-12-07T16:02:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.737380 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-m5jl5"] Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.737753 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-m5jl5" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.747646 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.748543 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.748785 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.748925 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.759455 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f95ef61779507288a03bda8eddf745e105f705998a117a30789675bfa452d88f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b955b23163e7c8c458a3bbe2f681633e6a663ce76748febcc0b3a3c37f3ca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:41Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.772472 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef3335ae3875a75ab64685fc7e3590177280d651e8f8292c6f5ea6f472e55789\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:41Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.785044 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:41Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.800727 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.800763 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.800773 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.800790 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.800800 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:41Z","lastTransitionTime":"2025-12-07T16:02:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.801516 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:41Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.812601 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c15b59eb-565d-4556-a4ce-75afdf159dc8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://530a2935826940767ea33fb4f37fa11f7dee4950d867fbcd6af72836d80f7db3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jks8z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ccaf6c1e6a978eb978a1c207b6c4c00b23737f78b743f966579aca781e22e0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jks8z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zcxxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:41Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.826598 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8ps9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98015cea-d753-4e7a-b2b5-5ae7733a81e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8ps9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:41Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.828263 4716 generic.go:334] "Generic (PLEG): container finished" podID="98015cea-d753-4e7a-b2b5-5ae7733a81e3" containerID="f84f52a073437543721010dc440339654ffb5e04eeb6cc7afbcc62f00f64c8e1" exitCode=0 Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.828674 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-8ps9j" event={"ID":"98015cea-d753-4e7a-b2b5-5ae7733a81e3","Type":"ContainerDied","Data":"f84f52a073437543721010dc440339654ffb5e04eeb6cc7afbcc62f00f64c8e1"} Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.830284 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-k4wr2" event={"ID":"0baf8575-26cf-4a72-8a6d-ae94c078eaad","Type":"ContainerStarted","Data":"f0ca2b3cd5a11033afd219b0125be7ab5acd750935d7eb89a470eff3d93a0838"} Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.832912 4716 generic.go:334] "Generic (PLEG): container finished" podID="c54d618a-dfec-4f2a-a288-f5188c8a128a" containerID="9bfe2ffce7a7b7c320474117734fc05246cc2b4fd744833e896e8e7c83f5e4d8" exitCode=0 Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.833384 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" event={"ID":"c54d618a-dfec-4f2a-a288-f5188c8a128a","Type":"ContainerDied","Data":"9bfe2ffce7a7b7c320474117734fc05246cc2b4fd744833e896e8e7c83f5e4d8"} Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.833415 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" event={"ID":"c54d618a-dfec-4f2a-a288-f5188c8a128a","Type":"ContainerStarted","Data":"ca8d06419e72db2db45ffd4d586e3eac926332b14273bb82e37118f39d5ebd3f"} Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.858409 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7961cc9a-734e-42c2-b1ff-7075f3f20091\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84a36b76f20618c7885d611ecad9db29bac5101c27572be31d3acbb90c581291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0a0eca3c2087a457c0e448e649345ec4595a9c8a019aed2cddc5f30c77871e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://404426a171b8676eddd4a00996c76afab7e269643fa9608f1e1b6bdba385071d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fd3c2941782921f99a6cdc3de4e892234548bde527fec15a8ac92e513566e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e23abfdca314ed43d594333c9f9e401cc8ef7dad39b8ebac84a836b06ee5c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09b38cd4b389ea477c802a842ee4f2297a64d64684624117163e819539c257c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09b38cd4b389ea477c802a842ee4f2297a64d64684624117163e819539c257c6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c00f7a8e024e0e0dfd7b73e16d1d5437602e79ea95b73c21f24b209edef60e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c00f7a8e024e0e0dfd7b73e16d1d5437602e79ea95b73c21f24b209edef60e98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3f6392d426fa2cd788ec3a5c3bff4ac39483a54dfb2624a1c4370522413a350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3f6392d426fa2cd788ec3a5c3bff4ac39483a54dfb2624a1c4370522413a350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:41Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.873668 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhwzn\" (UniqueName: \"kubernetes.io/projected/824ee59f-1474-4994-a70e-462878521514-kube-api-access-zhwzn\") pod \"node-ca-m5jl5\" (UID: \"824ee59f-1474-4994-a70e-462878521514\") " pod="openshift-image-registry/node-ca-m5jl5" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.873728 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/824ee59f-1474-4994-a70e-462878521514-serviceca\") pod \"node-ca-m5jl5\" (UID: \"824ee59f-1474-4994-a70e-462878521514\") " pod="openshift-image-registry/node-ca-m5jl5" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.873808 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/824ee59f-1474-4994-a70e-462878521514-host\") pod \"node-ca-m5jl5\" (UID: \"824ee59f-1474-4994-a70e-462878521514\") " pod="openshift-image-registry/node-ca-m5jl5" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.874731 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3067de7c-c11c-4be1-bd94-d6572c266795\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d14d018fa044e9e80fe81e7e38df935992d14c773ed2b2193fd5cb873e5d0947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a9e9cf048220ca14c996847e9ced9f2367b9331dc6c2836dd023658855cd79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6d3c911874095b3c7cb062d03070e5090a7bb2456cb20cb4ce078f502515cda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ac6ce7701f1c6aa1ea8fcf7a3035e3b0eaa28ae3cb6afc442372bd9117bdde4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:41Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.887903 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4742ea84b3a2c9461d9b900beee7354a2ded24c39ad23e007cbf270ae12377c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:41Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.903018 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.903064 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.903101 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.903120 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.903132 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:41Z","lastTransitionTime":"2025-12-07T16:02:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.906361 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c54d618a-dfec-4f2a-a288-f5188c8a128a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qvc7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:41Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.923274 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4wr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0baf8575-26cf-4a72-8a6d-ae94c078eaad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p2trd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4wr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:41Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.939523 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-m5jl5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"824ee59f-1474-4994-a70e-462878521514\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhwzn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-m5jl5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:41Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.956938 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdee9957-48b9-458f-ba71-e0cd39f80596\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73f85248b4d68a9beecdacb3a01f433e97a0ffea2247ce92a41ef96855509377\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6276df652efe6ca935e745f24ef20f87a16465475139230c63b9999234a4d12a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3b1dd202586ee755ca89fbe12b30bbb12f95a62514f6e4d93cbbdbe55181e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd318647816e4a27849666b0ec4fc2ca4905295661c557dc381e1144f22cf284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d31de0326a05216fa10442bdb47151c6e073c97e9a16c34e793028e5e253756\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"le observer\\\\nW1207 16:02:34.645464 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1207 16:02:34.645587 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 16:02:34.646501 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2082283479/tls.crt::/tmp/serving-cert-2082283479/tls.key\\\\\\\"\\\\nI1207 16:02:34.972041 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 16:02:34.973702 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 16:02:34.973717 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 16:02:34.973738 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 16:02:34.973743 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 16:02:34.977185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1207 16:02:34.977204 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1207 16:02:34.977209 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977217 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 16:02:34.977225 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 16:02:34.977228 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 16:02:34.977232 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1207 16:02:34.979033 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27bd4a54f920b2725987c1c7fd2d9fdf16cd08738c246c333a4e9164a3a582a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:41Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.969064 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:41Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.974390 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/824ee59f-1474-4994-a70e-462878521514-host\") pod \"node-ca-m5jl5\" (UID: \"824ee59f-1474-4994-a70e-462878521514\") " pod="openshift-image-registry/node-ca-m5jl5" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.974495 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhwzn\" (UniqueName: \"kubernetes.io/projected/824ee59f-1474-4994-a70e-462878521514-kube-api-access-zhwzn\") pod \"node-ca-m5jl5\" (UID: \"824ee59f-1474-4994-a70e-462878521514\") " pod="openshift-image-registry/node-ca-m5jl5" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.974519 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/824ee59f-1474-4994-a70e-462878521514-serviceca\") pod \"node-ca-m5jl5\" (UID: \"824ee59f-1474-4994-a70e-462878521514\") " pod="openshift-image-registry/node-ca-m5jl5" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.974580 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/824ee59f-1474-4994-a70e-462878521514-host\") pod \"node-ca-m5jl5\" (UID: \"824ee59f-1474-4994-a70e-462878521514\") " pod="openshift-image-registry/node-ca-m5jl5" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.976494 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/824ee59f-1474-4994-a70e-462878521514-serviceca\") pod \"node-ca-m5jl5\" (UID: \"824ee59f-1474-4994-a70e-462878521514\") " pod="openshift-image-registry/node-ca-m5jl5" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.980494 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2jjld" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dee6f5c-861c-4819-9309-7791574c7b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51c1d4159ab11087839a2c2d192566fb9322447c26df560bf09517c906ececb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmdns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2jjld\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:41Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:41 crc kubenswrapper[4716]: I1207 16:02:41.992883 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2jjld" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dee6f5c-861c-4819-9309-7791574c7b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51c1d4159ab11087839a2c2d192566fb9322447c26df560bf09517c906ececb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmdns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2jjld\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:41Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.008327 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4wr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0baf8575-26cf-4a72-8a6d-ae94c078eaad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0ca2b3cd5a11033afd219b0125be7ab5acd750935d7eb89a470eff3d93a0838\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p2trd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4wr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:42Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.008870 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.008895 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.008905 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.008921 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.008930 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:42Z","lastTransitionTime":"2025-12-07T16:02:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.013449 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhwzn\" (UniqueName: \"kubernetes.io/projected/824ee59f-1474-4994-a70e-462878521514-kube-api-access-zhwzn\") pod \"node-ca-m5jl5\" (UID: \"824ee59f-1474-4994-a70e-462878521514\") " pod="openshift-image-registry/node-ca-m5jl5" Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.018236 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-m5jl5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"824ee59f-1474-4994-a70e-462878521514\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhwzn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-m5jl5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:42Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.060658 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-m5jl5" Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.061287 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdee9957-48b9-458f-ba71-e0cd39f80596\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73f85248b4d68a9beecdacb3a01f433e97a0ffea2247ce92a41ef96855509377\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6276df652efe6ca935e745f24ef20f87a16465475139230c63b9999234a4d12a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3b1dd202586ee755ca89fbe12b30bbb12f95a62514f6e4d93cbbdbe55181e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd318647816e4a27849666b0ec4fc2ca4905295661c557dc381e1144f22cf284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d31de0326a05216fa10442bdb47151c6e073c97e9a16c34e793028e5e253756\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"le observer\\\\nW1207 16:02:34.645464 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1207 16:02:34.645587 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 16:02:34.646501 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2082283479/tls.crt::/tmp/serving-cert-2082283479/tls.key\\\\\\\"\\\\nI1207 16:02:34.972041 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 16:02:34.973702 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 16:02:34.973717 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 16:02:34.973738 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 16:02:34.973743 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 16:02:34.977185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1207 16:02:34.977204 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1207 16:02:34.977209 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977217 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 16:02:34.977225 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 16:02:34.977228 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 16:02:34.977232 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1207 16:02:34.979033 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27bd4a54f920b2725987c1c7fd2d9fdf16cd08738c246c333a4e9164a3a582a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:42Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:42 crc kubenswrapper[4716]: W1207 16:02:42.075589 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod824ee59f_1474_4994_a70e_462878521514.slice/crio-88baeaac655d47a7d65a898afe839f6f1152a393a8ee470b33c6c39bec9aba1e WatchSource:0}: Error finding container 88baeaac655d47a7d65a898afe839f6f1152a393a8ee470b33c6c39bec9aba1e: Status 404 returned error can't find the container with id 88baeaac655d47a7d65a898afe839f6f1152a393a8ee470b33c6c39bec9aba1e Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.105373 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:42Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.117284 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.117420 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.117453 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.117484 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.117505 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:42Z","lastTransitionTime":"2025-12-07T16:02:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.141356 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:42Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.181115 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f95ef61779507288a03bda8eddf745e105f705998a117a30789675bfa452d88f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b955b23163e7c8c458a3bbe2f681633e6a663ce76748febcc0b3a3c37f3ca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:42Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.219464 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef3335ae3875a75ab64685fc7e3590177280d651e8f8292c6f5ea6f472e55789\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:42Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.222983 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.223013 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.223024 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.223041 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.223054 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:42Z","lastTransitionTime":"2025-12-07T16:02:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.259990 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4742ea84b3a2c9461d9b900beee7354a2ded24c39ad23e007cbf270ae12377c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:42Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.301370 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:42Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.326792 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.326834 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.326847 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.326864 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.326876 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:42Z","lastTransitionTime":"2025-12-07T16:02:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.340017 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c15b59eb-565d-4556-a4ce-75afdf159dc8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://530a2935826940767ea33fb4f37fa11f7dee4950d867fbcd6af72836d80f7db3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jks8z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ccaf6c1e6a978eb978a1c207b6c4c00b23737f78b743f966579aca781e22e0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jks8z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zcxxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:42Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.381409 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8ps9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98015cea-d753-4e7a-b2b5-5ae7733a81e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f84f52a073437543721010dc440339654ffb5e04eeb6cc7afbcc62f00f64c8e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f84f52a073437543721010dc440339654ffb5e04eeb6cc7afbcc62f00f64c8e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8ps9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:42Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.427791 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7961cc9a-734e-42c2-b1ff-7075f3f20091\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84a36b76f20618c7885d611ecad9db29bac5101c27572be31d3acbb90c581291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0a0eca3c2087a457c0e448e649345ec4595a9c8a019aed2cddc5f30c77871e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://404426a171b8676eddd4a00996c76afab7e269643fa9608f1e1b6bdba385071d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fd3c2941782921f99a6cdc3de4e892234548bde527fec15a8ac92e513566e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e23abfdca314ed43d594333c9f9e401cc8ef7dad39b8ebac84a836b06ee5c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09b38cd4b389ea477c802a842ee4f2297a64d64684624117163e819539c257c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09b38cd4b389ea477c802a842ee4f2297a64d64684624117163e819539c257c6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c00f7a8e024e0e0dfd7b73e16d1d5437602e79ea95b73c21f24b209edef60e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c00f7a8e024e0e0dfd7b73e16d1d5437602e79ea95b73c21f24b209edef60e98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3f6392d426fa2cd788ec3a5c3bff4ac39483a54dfb2624a1c4370522413a350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3f6392d426fa2cd788ec3a5c3bff4ac39483a54dfb2624a1c4370522413a350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:42Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.429252 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.429368 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.429436 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.429495 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.429560 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:42Z","lastTransitionTime":"2025-12-07T16:02:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.462515 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3067de7c-c11c-4be1-bd94-d6572c266795\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d14d018fa044e9e80fe81e7e38df935992d14c773ed2b2193fd5cb873e5d0947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a9e9cf048220ca14c996847e9ced9f2367b9331dc6c2836dd023658855cd79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6d3c911874095b3c7cb062d03070e5090a7bb2456cb20cb4ce078f502515cda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ac6ce7701f1c6aa1ea8fcf7a3035e3b0eaa28ae3cb6afc442372bd9117bdde4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:42Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.515716 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c54d618a-dfec-4f2a-a288-f5188c8a128a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9bfe2ffce7a7b7c320474117734fc05246cc2b4fd744833e896e8e7c83f5e4d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9bfe2ffce7a7b7c320474117734fc05246cc2b4fd744833e896e8e7c83f5e4d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qvc7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:42Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.532622 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.532682 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.532696 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.532716 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.532732 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:42Z","lastTransitionTime":"2025-12-07T16:02:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.636046 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.636106 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.636116 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.636156 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.636167 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:42Z","lastTransitionTime":"2025-12-07T16:02:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.657187 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 16:02:42 crc kubenswrapper[4716]: E1207 16:02:42.657324 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.738955 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.739009 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.739023 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.739043 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.739056 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:42Z","lastTransitionTime":"2025-12-07T16:02:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.840437 4716 generic.go:334] "Generic (PLEG): container finished" podID="98015cea-d753-4e7a-b2b5-5ae7733a81e3" containerID="3f6fe1cb25218cb3ac20488dc47efa4dc6a26a43dbc8306b5ee38cbe1a186c67" exitCode=0 Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.840528 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-8ps9j" event={"ID":"98015cea-d753-4e7a-b2b5-5ae7733a81e3","Type":"ContainerDied","Data":"3f6fe1cb25218cb3ac20488dc47efa4dc6a26a43dbc8306b5ee38cbe1a186c67"} Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.841723 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.841748 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.841757 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.841770 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.841782 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:42Z","lastTransitionTime":"2025-12-07T16:02:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.842890 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-m5jl5" event={"ID":"824ee59f-1474-4994-a70e-462878521514","Type":"ContainerStarted","Data":"3a15fe50cc75515704779646cc19aa50004f0db9aa26fe2535110df749bbfb3d"} Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.842920 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-m5jl5" event={"ID":"824ee59f-1474-4994-a70e-462878521514","Type":"ContainerStarted","Data":"88baeaac655d47a7d65a898afe839f6f1152a393a8ee470b33c6c39bec9aba1e"} Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.850938 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" event={"ID":"c54d618a-dfec-4f2a-a288-f5188c8a128a","Type":"ContainerStarted","Data":"def9ac840fb8ecf2778bed88ea980eddca7c5b2aba4bd1afb967d62fec3f7bae"} Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.850964 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" event={"ID":"c54d618a-dfec-4f2a-a288-f5188c8a128a","Type":"ContainerStarted","Data":"b73be664f5412c98b89d971a2cac37c494151c293fdb44a98b390978658a6643"} Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.850977 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" event={"ID":"c54d618a-dfec-4f2a-a288-f5188c8a128a","Type":"ContainerStarted","Data":"2ed4f4d3564674bf5cb4e7a76dc0fd95e203ecc8950798e59a1fb6c9ab4a20f7"} Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.850988 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" event={"ID":"c54d618a-dfec-4f2a-a288-f5188c8a128a","Type":"ContainerStarted","Data":"60324dc7d701bc064e14ae4f474bce10f6ee4ff3c3c942eab749e7166082cf9d"} Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.850996 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" event={"ID":"c54d618a-dfec-4f2a-a288-f5188c8a128a","Type":"ContainerStarted","Data":"a847d04e22e98091c0037a89054b9236228b701b8c7ba31940a56df0f37b95c6"} Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.851007 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" event={"ID":"c54d618a-dfec-4f2a-a288-f5188c8a128a","Type":"ContainerStarted","Data":"dcae78a6133d97f6a32dd7421ccec4333b970a237404e056a5cc601f8eefdc8d"} Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.869071 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c54d618a-dfec-4f2a-a288-f5188c8a128a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9bfe2ffce7a7b7c320474117734fc05246cc2b4fd744833e896e8e7c83f5e4d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9bfe2ffce7a7b7c320474117734fc05246cc2b4fd744833e896e8e7c83f5e4d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qvc7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:42Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.887818 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4wr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0baf8575-26cf-4a72-8a6d-ae94c078eaad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0ca2b3cd5a11033afd219b0125be7ab5acd750935d7eb89a470eff3d93a0838\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p2trd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4wr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:42Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.918314 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-m5jl5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"824ee59f-1474-4994-a70e-462878521514\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhwzn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-m5jl5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:42Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.939009 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdee9957-48b9-458f-ba71-e0cd39f80596\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73f85248b4d68a9beecdacb3a01f433e97a0ffea2247ce92a41ef96855509377\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6276df652efe6ca935e745f24ef20f87a16465475139230c63b9999234a4d12a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3b1dd202586ee755ca89fbe12b30bbb12f95a62514f6e4d93cbbdbe55181e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd318647816e4a27849666b0ec4fc2ca4905295661c557dc381e1144f22cf284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d31de0326a05216fa10442bdb47151c6e073c97e9a16c34e793028e5e253756\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"le observer\\\\nW1207 16:02:34.645464 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1207 16:02:34.645587 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 16:02:34.646501 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2082283479/tls.crt::/tmp/serving-cert-2082283479/tls.key\\\\\\\"\\\\nI1207 16:02:34.972041 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 16:02:34.973702 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 16:02:34.973717 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 16:02:34.973738 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 16:02:34.973743 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 16:02:34.977185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1207 16:02:34.977204 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1207 16:02:34.977209 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977217 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 16:02:34.977225 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 16:02:34.977228 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 16:02:34.977232 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1207 16:02:34.979033 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27bd4a54f920b2725987c1c7fd2d9fdf16cd08738c246c333a4e9164a3a582a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:42Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.949829 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.949889 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.949909 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.949939 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.949961 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:42Z","lastTransitionTime":"2025-12-07T16:02:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.958755 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:42Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.971017 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2jjld" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dee6f5c-861c-4819-9309-7791574c7b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51c1d4159ab11087839a2c2d192566fb9322447c26df560bf09517c906ececb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmdns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2jjld\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:42Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:42 crc kubenswrapper[4716]: I1207 16:02:42.985965 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f95ef61779507288a03bda8eddf745e105f705998a117a30789675bfa452d88f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b955b23163e7c8c458a3bbe2f681633e6a663ce76748febcc0b3a3c37f3ca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:42Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.004908 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef3335ae3875a75ab64685fc7e3590177280d651e8f8292c6f5ea6f472e55789\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:43Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.021215 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:43Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.038659 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:43Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.052013 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.052052 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.052064 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.052096 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.052109 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:43Z","lastTransitionTime":"2025-12-07T16:02:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.056698 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c15b59eb-565d-4556-a4ce-75afdf159dc8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://530a2935826940767ea33fb4f37fa11f7dee4950d867fbcd6af72836d80f7db3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jks8z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ccaf6c1e6a978eb978a1c207b6c4c00b23737f78b743f966579aca781e22e0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jks8z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zcxxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:43Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.078447 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8ps9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98015cea-d753-4e7a-b2b5-5ae7733a81e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f84f52a073437543721010dc440339654ffb5e04eeb6cc7afbcc62f00f64c8e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f84f52a073437543721010dc440339654ffb5e04eeb6cc7afbcc62f00f64c8e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f6fe1cb25218cb3ac20488dc47efa4dc6a26a43dbc8306b5ee38cbe1a186c67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f6fe1cb25218cb3ac20488dc47efa4dc6a26a43dbc8306b5ee38cbe1a186c67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8ps9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:43Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.108228 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7961cc9a-734e-42c2-b1ff-7075f3f20091\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84a36b76f20618c7885d611ecad9db29bac5101c27572be31d3acbb90c581291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0a0eca3c2087a457c0e448e649345ec4595a9c8a019aed2cddc5f30c77871e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://404426a171b8676eddd4a00996c76afab7e269643fa9608f1e1b6bdba385071d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fd3c2941782921f99a6cdc3de4e892234548bde527fec15a8ac92e513566e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e23abfdca314ed43d594333c9f9e401cc8ef7dad39b8ebac84a836b06ee5c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09b38cd4b389ea477c802a842ee4f2297a64d64684624117163e819539c257c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09b38cd4b389ea477c802a842ee4f2297a64d64684624117163e819539c257c6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c00f7a8e024e0e0dfd7b73e16d1d5437602e79ea95b73c21f24b209edef60e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c00f7a8e024e0e0dfd7b73e16d1d5437602e79ea95b73c21f24b209edef60e98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3f6392d426fa2cd788ec3a5c3bff4ac39483a54dfb2624a1c4370522413a350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3f6392d426fa2cd788ec3a5c3bff4ac39483a54dfb2624a1c4370522413a350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:43Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.126853 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3067de7c-c11c-4be1-bd94-d6572c266795\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d14d018fa044e9e80fe81e7e38df935992d14c773ed2b2193fd5cb873e5d0947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a9e9cf048220ca14c996847e9ced9f2367b9331dc6c2836dd023658855cd79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6d3c911874095b3c7cb062d03070e5090a7bb2456cb20cb4ce078f502515cda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ac6ce7701f1c6aa1ea8fcf7a3035e3b0eaa28ae3cb6afc442372bd9117bdde4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:43Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.142502 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4742ea84b3a2c9461d9b900beee7354a2ded24c39ad23e007cbf270ae12377c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:43Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.154396 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.154435 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.154447 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.154464 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.154476 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:43Z","lastTransitionTime":"2025-12-07T16:02:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.160248 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4742ea84b3a2c9461d9b900beee7354a2ded24c39ad23e007cbf270ae12377c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:43Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.183889 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:43Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.223757 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c15b59eb-565d-4556-a4ce-75afdf159dc8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://530a2935826940767ea33fb4f37fa11f7dee4950d867fbcd6af72836d80f7db3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jks8z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ccaf6c1e6a978eb978a1c207b6c4c00b23737f78b743f966579aca781e22e0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jks8z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zcxxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:43Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.258287 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.258366 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.258386 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.258417 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.258440 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:43Z","lastTransitionTime":"2025-12-07T16:02:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.266927 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8ps9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98015cea-d753-4e7a-b2b5-5ae7733a81e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f84f52a073437543721010dc440339654ffb5e04eeb6cc7afbcc62f00f64c8e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f84f52a073437543721010dc440339654ffb5e04eeb6cc7afbcc62f00f64c8e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f6fe1cb25218cb3ac20488dc47efa4dc6a26a43dbc8306b5ee38cbe1a186c67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f6fe1cb25218cb3ac20488dc47efa4dc6a26a43dbc8306b5ee38cbe1a186c67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8ps9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:43Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.289282 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.289382 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 16:02:43 crc kubenswrapper[4716]: E1207 16:02:43.289588 4716 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 07 16:02:43 crc kubenswrapper[4716]: E1207 16:02:43.289634 4716 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 07 16:02:43 crc kubenswrapper[4716]: E1207 16:02:43.289640 4716 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 07 16:02:43 crc kubenswrapper[4716]: E1207 16:02:43.289656 4716 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 07 16:02:43 crc kubenswrapper[4716]: E1207 16:02:43.289676 4716 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 07 16:02:43 crc kubenswrapper[4716]: E1207 16:02:43.289699 4716 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 07 16:02:43 crc kubenswrapper[4716]: E1207 16:02:43.289798 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-07 16:02:51.289763252 +0000 UTC m=+33.980048194 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 07 16:02:43 crc kubenswrapper[4716]: E1207 16:02:43.289847 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-07 16:02:51.289825404 +0000 UTC m=+33.980110616 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.310857 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7961cc9a-734e-42c2-b1ff-7075f3f20091\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84a36b76f20618c7885d611ecad9db29bac5101c27572be31d3acbb90c581291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0a0eca3c2087a457c0e448e649345ec4595a9c8a019aed2cddc5f30c77871e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://404426a171b8676eddd4a00996c76afab7e269643fa9608f1e1b6bdba385071d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fd3c2941782921f99a6cdc3de4e892234548bde527fec15a8ac92e513566e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e23abfdca314ed43d594333c9f9e401cc8ef7dad39b8ebac84a836b06ee5c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09b38cd4b389ea477c802a842ee4f2297a64d64684624117163e819539c257c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09b38cd4b389ea477c802a842ee4f2297a64d64684624117163e819539c257c6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c00f7a8e024e0e0dfd7b73e16d1d5437602e79ea95b73c21f24b209edef60e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c00f7a8e024e0e0dfd7b73e16d1d5437602e79ea95b73c21f24b209edef60e98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3f6392d426fa2cd788ec3a5c3bff4ac39483a54dfb2624a1c4370522413a350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3f6392d426fa2cd788ec3a5c3bff4ac39483a54dfb2624a1c4370522413a350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:43Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.341924 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3067de7c-c11c-4be1-bd94-d6572c266795\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d14d018fa044e9e80fe81e7e38df935992d14c773ed2b2193fd5cb873e5d0947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a9e9cf048220ca14c996847e9ced9f2367b9331dc6c2836dd023658855cd79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6d3c911874095b3c7cb062d03070e5090a7bb2456cb20cb4ce078f502515cda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ac6ce7701f1c6aa1ea8fcf7a3035e3b0eaa28ae3cb6afc442372bd9117bdde4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:43Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.362065 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.362154 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.362166 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.362186 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.362199 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:43Z","lastTransitionTime":"2025-12-07T16:02:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.389461 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c54d618a-dfec-4f2a-a288-f5188c8a128a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9bfe2ffce7a7b7c320474117734fc05246cc2b4fd744833e896e8e7c83f5e4d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9bfe2ffce7a7b7c320474117734fc05246cc2b4fd744833e896e8e7c83f5e4d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qvc7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:43Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.390181 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.390333 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:02:43 crc kubenswrapper[4716]: E1207 16:02:43.390468 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 16:02:51.390441384 +0000 UTC m=+34.080726296 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.390590 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:02:43 crc kubenswrapper[4716]: E1207 16:02:43.390522 4716 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 07 16:02:43 crc kubenswrapper[4716]: E1207 16:02:43.390671 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-07 16:02:51.39064647 +0000 UTC m=+34.080931422 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 07 16:02:43 crc kubenswrapper[4716]: E1207 16:02:43.390715 4716 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 07 16:02:43 crc kubenswrapper[4716]: E1207 16:02:43.390761 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-07 16:02:51.390754843 +0000 UTC m=+34.081039755 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.420154 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2jjld" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dee6f5c-861c-4819-9309-7791574c7b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51c1d4159ab11087839a2c2d192566fb9322447c26df560bf09517c906ececb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmdns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2jjld\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:43Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.462842 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4wr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0baf8575-26cf-4a72-8a6d-ae94c078eaad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0ca2b3cd5a11033afd219b0125be7ab5acd750935d7eb89a470eff3d93a0838\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p2trd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4wr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:43Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.464990 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.465106 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.465171 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.465232 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.465288 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:43Z","lastTransitionTime":"2025-12-07T16:02:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.500608 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-m5jl5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"824ee59f-1474-4994-a70e-462878521514\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a15fe50cc75515704779646cc19aa50004f0db9aa26fe2535110df749bbfb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhwzn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-m5jl5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:43Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.543039 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdee9957-48b9-458f-ba71-e0cd39f80596\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73f85248b4d68a9beecdacb3a01f433e97a0ffea2247ce92a41ef96855509377\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6276df652efe6ca935e745f24ef20f87a16465475139230c63b9999234a4d12a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3b1dd202586ee755ca89fbe12b30bbb12f95a62514f6e4d93cbbdbe55181e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd318647816e4a27849666b0ec4fc2ca4905295661c557dc381e1144f22cf284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d31de0326a05216fa10442bdb47151c6e073c97e9a16c34e793028e5e253756\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"le observer\\\\nW1207 16:02:34.645464 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1207 16:02:34.645587 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 16:02:34.646501 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2082283479/tls.crt::/tmp/serving-cert-2082283479/tls.key\\\\\\\"\\\\nI1207 16:02:34.972041 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 16:02:34.973702 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 16:02:34.973717 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 16:02:34.973738 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 16:02:34.973743 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 16:02:34.977185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1207 16:02:34.977204 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1207 16:02:34.977209 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977217 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 16:02:34.977225 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 16:02:34.977228 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 16:02:34.977232 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1207 16:02:34.979033 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27bd4a54f920b2725987c1c7fd2d9fdf16cd08738c246c333a4e9164a3a582a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:43Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.567708 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.567921 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.567981 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.568044 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.568159 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:43Z","lastTransitionTime":"2025-12-07T16:02:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.585413 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:43Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.627333 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:43Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.657479 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 16:02:43 crc kubenswrapper[4716]: E1207 16:02:43.657823 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.658341 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:02:43 crc kubenswrapper[4716]: E1207 16:02:43.658560 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.668919 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f95ef61779507288a03bda8eddf745e105f705998a117a30789675bfa452d88f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b955b23163e7c8c458a3bbe2f681633e6a663ce76748febcc0b3a3c37f3ca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:43Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.671492 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.671557 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.671574 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.671599 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.671620 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:43Z","lastTransitionTime":"2025-12-07T16:02:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.709833 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef3335ae3875a75ab64685fc7e3590177280d651e8f8292c6f5ea6f472e55789\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:43Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.773779 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.773815 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.773828 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.773845 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.773859 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:43Z","lastTransitionTime":"2025-12-07T16:02:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.857178 4716 generic.go:334] "Generic (PLEG): container finished" podID="98015cea-d753-4e7a-b2b5-5ae7733a81e3" containerID="f14b748a8247f6863e8be61e90057a06626dfd3cd4f01bd3a11629f700f56378" exitCode=0 Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.857226 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-8ps9j" event={"ID":"98015cea-d753-4e7a-b2b5-5ae7733a81e3","Type":"ContainerDied","Data":"f14b748a8247f6863e8be61e90057a06626dfd3cd4f01bd3a11629f700f56378"} Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.873759 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4742ea84b3a2c9461d9b900beee7354a2ded24c39ad23e007cbf270ae12377c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:43Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.880135 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.880186 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.880202 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.880222 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.880237 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:43Z","lastTransitionTime":"2025-12-07T16:02:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.891476 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:43Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.908820 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c15b59eb-565d-4556-a4ce-75afdf159dc8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://530a2935826940767ea33fb4f37fa11f7dee4950d867fbcd6af72836d80f7db3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jks8z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ccaf6c1e6a978eb978a1c207b6c4c00b23737f78b743f966579aca781e22e0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jks8z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zcxxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:43Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.929912 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8ps9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98015cea-d753-4e7a-b2b5-5ae7733a81e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f84f52a073437543721010dc440339654ffb5e04eeb6cc7afbcc62f00f64c8e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f84f52a073437543721010dc440339654ffb5e04eeb6cc7afbcc62f00f64c8e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f6fe1cb25218cb3ac20488dc47efa4dc6a26a43dbc8306b5ee38cbe1a186c67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f6fe1cb25218cb3ac20488dc47efa4dc6a26a43dbc8306b5ee38cbe1a186c67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14b748a8247f6863e8be61e90057a06626dfd3cd4f01bd3a11629f700f56378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f14b748a8247f6863e8be61e90057a06626dfd3cd4f01bd3a11629f700f56378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8ps9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:43Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.961013 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7961cc9a-734e-42c2-b1ff-7075f3f20091\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84a36b76f20618c7885d611ecad9db29bac5101c27572be31d3acbb90c581291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0a0eca3c2087a457c0e448e649345ec4595a9c8a019aed2cddc5f30c77871e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://404426a171b8676eddd4a00996c76afab7e269643fa9608f1e1b6bdba385071d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fd3c2941782921f99a6cdc3de4e892234548bde527fec15a8ac92e513566e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e23abfdca314ed43d594333c9f9e401cc8ef7dad39b8ebac84a836b06ee5c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09b38cd4b389ea477c802a842ee4f2297a64d64684624117163e819539c257c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09b38cd4b389ea477c802a842ee4f2297a64d64684624117163e819539c257c6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c00f7a8e024e0e0dfd7b73e16d1d5437602e79ea95b73c21f24b209edef60e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c00f7a8e024e0e0dfd7b73e16d1d5437602e79ea95b73c21f24b209edef60e98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3f6392d426fa2cd788ec3a5c3bff4ac39483a54dfb2624a1c4370522413a350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3f6392d426fa2cd788ec3a5c3bff4ac39483a54dfb2624a1c4370522413a350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:43Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.976038 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3067de7c-c11c-4be1-bd94-d6572c266795\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d14d018fa044e9e80fe81e7e38df935992d14c773ed2b2193fd5cb873e5d0947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a9e9cf048220ca14c996847e9ced9f2367b9331dc6c2836dd023658855cd79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6d3c911874095b3c7cb062d03070e5090a7bb2456cb20cb4ce078f502515cda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ac6ce7701f1c6aa1ea8fcf7a3035e3b0eaa28ae3cb6afc442372bd9117bdde4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:43Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.981700 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.981729 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.981738 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.981751 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:43 crc kubenswrapper[4716]: I1207 16:02:43.981761 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:43Z","lastTransitionTime":"2025-12-07T16:02:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:44 crc kubenswrapper[4716]: I1207 16:02:44.002649 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c54d618a-dfec-4f2a-a288-f5188c8a128a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9bfe2ffce7a7b7c320474117734fc05246cc2b4fd744833e896e8e7c83f5e4d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9bfe2ffce7a7b7c320474117734fc05246cc2b4fd744833e896e8e7c83f5e4d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qvc7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:44Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:44 crc kubenswrapper[4716]: I1207 16:02:44.017642 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2jjld" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dee6f5c-861c-4819-9309-7791574c7b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51c1d4159ab11087839a2c2d192566fb9322447c26df560bf09517c906ececb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmdns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2jjld\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:44Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:44 crc kubenswrapper[4716]: I1207 16:02:44.060902 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4wr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0baf8575-26cf-4a72-8a6d-ae94c078eaad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0ca2b3cd5a11033afd219b0125be7ab5acd750935d7eb89a470eff3d93a0838\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p2trd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4wr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:44Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:44 crc kubenswrapper[4716]: I1207 16:02:44.084343 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:44 crc kubenswrapper[4716]: I1207 16:02:44.084372 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:44 crc kubenswrapper[4716]: I1207 16:02:44.084380 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:44 crc kubenswrapper[4716]: I1207 16:02:44.084394 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:44 crc kubenswrapper[4716]: I1207 16:02:44.084402 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:44Z","lastTransitionTime":"2025-12-07T16:02:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:44 crc kubenswrapper[4716]: I1207 16:02:44.097866 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-m5jl5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"824ee59f-1474-4994-a70e-462878521514\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a15fe50cc75515704779646cc19aa50004f0db9aa26fe2535110df749bbfb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhwzn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-m5jl5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:44Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:44 crc kubenswrapper[4716]: I1207 16:02:44.153371 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdee9957-48b9-458f-ba71-e0cd39f80596\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73f85248b4d68a9beecdacb3a01f433e97a0ffea2247ce92a41ef96855509377\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6276df652efe6ca935e745f24ef20f87a16465475139230c63b9999234a4d12a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3b1dd202586ee755ca89fbe12b30bbb12f95a62514f6e4d93cbbdbe55181e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd318647816e4a27849666b0ec4fc2ca4905295661c557dc381e1144f22cf284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d31de0326a05216fa10442bdb47151c6e073c97e9a16c34e793028e5e253756\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"le observer\\\\nW1207 16:02:34.645464 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1207 16:02:34.645587 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 16:02:34.646501 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2082283479/tls.crt::/tmp/serving-cert-2082283479/tls.key\\\\\\\"\\\\nI1207 16:02:34.972041 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 16:02:34.973702 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 16:02:34.973717 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 16:02:34.973738 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 16:02:34.973743 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 16:02:34.977185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1207 16:02:34.977204 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1207 16:02:34.977209 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977217 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 16:02:34.977225 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 16:02:34.977228 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 16:02:34.977232 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1207 16:02:34.979033 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27bd4a54f920b2725987c1c7fd2d9fdf16cd08738c246c333a4e9164a3a582a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:44Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:44 crc kubenswrapper[4716]: I1207 16:02:44.186840 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:44 crc kubenswrapper[4716]: I1207 16:02:44.186874 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:44 crc kubenswrapper[4716]: I1207 16:02:44.186884 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:44 crc kubenswrapper[4716]: I1207 16:02:44.186899 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:44 crc kubenswrapper[4716]: I1207 16:02:44.186909 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:44Z","lastTransitionTime":"2025-12-07T16:02:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:44 crc kubenswrapper[4716]: I1207 16:02:44.195568 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:44Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:44 crc kubenswrapper[4716]: I1207 16:02:44.218788 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:44Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:44 crc kubenswrapper[4716]: I1207 16:02:44.259479 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f95ef61779507288a03bda8eddf745e105f705998a117a30789675bfa452d88f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b955b23163e7c8c458a3bbe2f681633e6a663ce76748febcc0b3a3c37f3ca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:44Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:44 crc kubenswrapper[4716]: I1207 16:02:44.289935 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:44 crc kubenswrapper[4716]: I1207 16:02:44.289975 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:44 crc kubenswrapper[4716]: I1207 16:02:44.289983 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:44 crc kubenswrapper[4716]: I1207 16:02:44.289997 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:44 crc kubenswrapper[4716]: I1207 16:02:44.290005 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:44Z","lastTransitionTime":"2025-12-07T16:02:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:44 crc kubenswrapper[4716]: I1207 16:02:44.302466 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef3335ae3875a75ab64685fc7e3590177280d651e8f8292c6f5ea6f472e55789\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:44Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:44 crc kubenswrapper[4716]: I1207 16:02:44.393293 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:44 crc kubenswrapper[4716]: I1207 16:02:44.393354 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:44 crc kubenswrapper[4716]: I1207 16:02:44.393373 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:44 crc kubenswrapper[4716]: I1207 16:02:44.393398 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:44 crc kubenswrapper[4716]: I1207 16:02:44.393417 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:44Z","lastTransitionTime":"2025-12-07T16:02:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:44 crc kubenswrapper[4716]: I1207 16:02:44.497192 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:44 crc kubenswrapper[4716]: I1207 16:02:44.497256 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:44 crc kubenswrapper[4716]: I1207 16:02:44.497276 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:44 crc kubenswrapper[4716]: I1207 16:02:44.497301 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:44 crc kubenswrapper[4716]: I1207 16:02:44.497317 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:44Z","lastTransitionTime":"2025-12-07T16:02:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:44 crc kubenswrapper[4716]: I1207 16:02:44.599237 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:44 crc kubenswrapper[4716]: I1207 16:02:44.599334 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:44 crc kubenswrapper[4716]: I1207 16:02:44.599347 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:44 crc kubenswrapper[4716]: I1207 16:02:44.599365 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:44 crc kubenswrapper[4716]: I1207 16:02:44.599383 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:44Z","lastTransitionTime":"2025-12-07T16:02:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:44 crc kubenswrapper[4716]: I1207 16:02:44.656719 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 16:02:44 crc kubenswrapper[4716]: E1207 16:02:44.656935 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 16:02:44 crc kubenswrapper[4716]: I1207 16:02:44.701809 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:44 crc kubenswrapper[4716]: I1207 16:02:44.701852 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:44 crc kubenswrapper[4716]: I1207 16:02:44.701865 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:44 crc kubenswrapper[4716]: I1207 16:02:44.701882 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:44 crc kubenswrapper[4716]: I1207 16:02:44.701897 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:44Z","lastTransitionTime":"2025-12-07T16:02:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:44 crc kubenswrapper[4716]: I1207 16:02:44.805369 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:44 crc kubenswrapper[4716]: I1207 16:02:44.805398 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:44 crc kubenswrapper[4716]: I1207 16:02:44.805437 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:44 crc kubenswrapper[4716]: I1207 16:02:44.805452 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:44 crc kubenswrapper[4716]: I1207 16:02:44.805461 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:44Z","lastTransitionTime":"2025-12-07T16:02:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:44 crc kubenswrapper[4716]: I1207 16:02:44.865155 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" event={"ID":"c54d618a-dfec-4f2a-a288-f5188c8a128a","Type":"ContainerStarted","Data":"bcfdffe822635b0ae13b253e6133449fe00461b2792be43d913c4a77219b8255"} Dec 07 16:02:44 crc kubenswrapper[4716]: I1207 16:02:44.868095 4716 generic.go:334] "Generic (PLEG): container finished" podID="98015cea-d753-4e7a-b2b5-5ae7733a81e3" containerID="9ec379c0d573cb9b5ffcfc5ca7c653ba60e2fbdd7e3d37836102e26618ce43e8" exitCode=0 Dec 07 16:02:44 crc kubenswrapper[4716]: I1207 16:02:44.868127 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-8ps9j" event={"ID":"98015cea-d753-4e7a-b2b5-5ae7733a81e3","Type":"ContainerDied","Data":"9ec379c0d573cb9b5ffcfc5ca7c653ba60e2fbdd7e3d37836102e26618ce43e8"} Dec 07 16:02:44 crc kubenswrapper[4716]: I1207 16:02:44.882436 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef3335ae3875a75ab64685fc7e3590177280d651e8f8292c6f5ea6f472e55789\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:44Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:44 crc kubenswrapper[4716]: I1207 16:02:44.908512 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:44Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:44 crc kubenswrapper[4716]: I1207 16:02:44.909691 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:44 crc kubenswrapper[4716]: I1207 16:02:44.909723 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:44 crc kubenswrapper[4716]: I1207 16:02:44.909732 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:44 crc kubenswrapper[4716]: I1207 16:02:44.909745 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:44 crc kubenswrapper[4716]: I1207 16:02:44.909755 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:44Z","lastTransitionTime":"2025-12-07T16:02:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:44 crc kubenswrapper[4716]: I1207 16:02:44.924726 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f95ef61779507288a03bda8eddf745e105f705998a117a30789675bfa452d88f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b955b23163e7c8c458a3bbe2f681633e6a663ce76748febcc0b3a3c37f3ca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:44Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:44 crc kubenswrapper[4716]: I1207 16:02:44.952386 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7961cc9a-734e-42c2-b1ff-7075f3f20091\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84a36b76f20618c7885d611ecad9db29bac5101c27572be31d3acbb90c581291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0a0eca3c2087a457c0e448e649345ec4595a9c8a019aed2cddc5f30c77871e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://404426a171b8676eddd4a00996c76afab7e269643fa9608f1e1b6bdba385071d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fd3c2941782921f99a6cdc3de4e892234548bde527fec15a8ac92e513566e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e23abfdca314ed43d594333c9f9e401cc8ef7dad39b8ebac84a836b06ee5c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09b38cd4b389ea477c802a842ee4f2297a64d64684624117163e819539c257c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09b38cd4b389ea477c802a842ee4f2297a64d64684624117163e819539c257c6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c00f7a8e024e0e0dfd7b73e16d1d5437602e79ea95b73c21f24b209edef60e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c00f7a8e024e0e0dfd7b73e16d1d5437602e79ea95b73c21f24b209edef60e98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3f6392d426fa2cd788ec3a5c3bff4ac39483a54dfb2624a1c4370522413a350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3f6392d426fa2cd788ec3a5c3bff4ac39483a54dfb2624a1c4370522413a350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:44Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:44 crc kubenswrapper[4716]: I1207 16:02:44.967166 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3067de7c-c11c-4be1-bd94-d6572c266795\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d14d018fa044e9e80fe81e7e38df935992d14c773ed2b2193fd5cb873e5d0947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a9e9cf048220ca14c996847e9ced9f2367b9331dc6c2836dd023658855cd79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6d3c911874095b3c7cb062d03070e5090a7bb2456cb20cb4ce078f502515cda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ac6ce7701f1c6aa1ea8fcf7a3035e3b0eaa28ae3cb6afc442372bd9117bdde4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:44Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:44 crc kubenswrapper[4716]: I1207 16:02:44.978718 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4742ea84b3a2c9461d9b900beee7354a2ded24c39ad23e007cbf270ae12377c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:44Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:44 crc kubenswrapper[4716]: I1207 16:02:44.991813 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:44Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:45 crc kubenswrapper[4716]: I1207 16:02:45.006392 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c15b59eb-565d-4556-a4ce-75afdf159dc8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://530a2935826940767ea33fb4f37fa11f7dee4950d867fbcd6af72836d80f7db3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jks8z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ccaf6c1e6a978eb978a1c207b6c4c00b23737f78b743f966579aca781e22e0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jks8z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zcxxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:45Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:45 crc kubenswrapper[4716]: I1207 16:02:45.013801 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:45 crc kubenswrapper[4716]: I1207 16:02:45.013849 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:45 crc kubenswrapper[4716]: I1207 16:02:45.013862 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:45 crc kubenswrapper[4716]: I1207 16:02:45.013884 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:45 crc kubenswrapper[4716]: I1207 16:02:45.013899 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:45Z","lastTransitionTime":"2025-12-07T16:02:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:45 crc kubenswrapper[4716]: I1207 16:02:45.026375 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8ps9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98015cea-d753-4e7a-b2b5-5ae7733a81e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f84f52a073437543721010dc440339654ffb5e04eeb6cc7afbcc62f00f64c8e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f84f52a073437543721010dc440339654ffb5e04eeb6cc7afbcc62f00f64c8e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f6fe1cb25218cb3ac20488dc47efa4dc6a26a43dbc8306b5ee38cbe1a186c67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f6fe1cb25218cb3ac20488dc47efa4dc6a26a43dbc8306b5ee38cbe1a186c67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14b748a8247f6863e8be61e90057a06626dfd3cd4f01bd3a11629f700f56378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f14b748a8247f6863e8be61e90057a06626dfd3cd4f01bd3a11629f700f56378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ec379c0d573cb9b5ffcfc5ca7c653ba60e2fbdd7e3d37836102e26618ce43e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ec379c0d573cb9b5ffcfc5ca7c653ba60e2fbdd7e3d37836102e26618ce43e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8ps9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:45Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:45 crc kubenswrapper[4716]: I1207 16:02:45.053558 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c54d618a-dfec-4f2a-a288-f5188c8a128a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9bfe2ffce7a7b7c320474117734fc05246cc2b4fd744833e896e8e7c83f5e4d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9bfe2ffce7a7b7c320474117734fc05246cc2b4fd744833e896e8e7c83f5e4d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qvc7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:45Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:45 crc kubenswrapper[4716]: I1207 16:02:45.070733 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdee9957-48b9-458f-ba71-e0cd39f80596\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73f85248b4d68a9beecdacb3a01f433e97a0ffea2247ce92a41ef96855509377\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6276df652efe6ca935e745f24ef20f87a16465475139230c63b9999234a4d12a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3b1dd202586ee755ca89fbe12b30bbb12f95a62514f6e4d93cbbdbe55181e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd318647816e4a27849666b0ec4fc2ca4905295661c557dc381e1144f22cf284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d31de0326a05216fa10442bdb47151c6e073c97e9a16c34e793028e5e253756\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"le observer\\\\nW1207 16:02:34.645464 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1207 16:02:34.645587 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 16:02:34.646501 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2082283479/tls.crt::/tmp/serving-cert-2082283479/tls.key\\\\\\\"\\\\nI1207 16:02:34.972041 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 16:02:34.973702 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 16:02:34.973717 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 16:02:34.973738 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 16:02:34.973743 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 16:02:34.977185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1207 16:02:34.977204 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1207 16:02:34.977209 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977217 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 16:02:34.977225 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 16:02:34.977228 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 16:02:34.977232 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1207 16:02:34.979033 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27bd4a54f920b2725987c1c7fd2d9fdf16cd08738c246c333a4e9164a3a582a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:45Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:45 crc kubenswrapper[4716]: I1207 16:02:45.082834 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:45Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:45 crc kubenswrapper[4716]: I1207 16:02:45.093397 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2jjld" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dee6f5c-861c-4819-9309-7791574c7b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51c1d4159ab11087839a2c2d192566fb9322447c26df560bf09517c906ececb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmdns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2jjld\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:45Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:45 crc kubenswrapper[4716]: I1207 16:02:45.105206 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4wr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0baf8575-26cf-4a72-8a6d-ae94c078eaad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0ca2b3cd5a11033afd219b0125be7ab5acd750935d7eb89a470eff3d93a0838\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p2trd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4wr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:45Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:45 crc kubenswrapper[4716]: I1207 16:02:45.115118 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-m5jl5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"824ee59f-1474-4994-a70e-462878521514\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a15fe50cc75515704779646cc19aa50004f0db9aa26fe2535110df749bbfb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhwzn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-m5jl5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:45Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:45 crc kubenswrapper[4716]: I1207 16:02:45.116665 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:45 crc kubenswrapper[4716]: I1207 16:02:45.116690 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:45 crc kubenswrapper[4716]: I1207 16:02:45.116701 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:45 crc kubenswrapper[4716]: I1207 16:02:45.116718 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:45 crc kubenswrapper[4716]: I1207 16:02:45.116730 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:45Z","lastTransitionTime":"2025-12-07T16:02:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:45 crc kubenswrapper[4716]: I1207 16:02:45.218971 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:45 crc kubenswrapper[4716]: I1207 16:02:45.219001 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:45 crc kubenswrapper[4716]: I1207 16:02:45.219009 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:45 crc kubenswrapper[4716]: I1207 16:02:45.219022 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:45 crc kubenswrapper[4716]: I1207 16:02:45.219030 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:45Z","lastTransitionTime":"2025-12-07T16:02:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:45 crc kubenswrapper[4716]: I1207 16:02:45.321173 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:45 crc kubenswrapper[4716]: I1207 16:02:45.321213 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:45 crc kubenswrapper[4716]: I1207 16:02:45.321224 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:45 crc kubenswrapper[4716]: I1207 16:02:45.321241 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:45 crc kubenswrapper[4716]: I1207 16:02:45.321254 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:45Z","lastTransitionTime":"2025-12-07T16:02:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:45 crc kubenswrapper[4716]: I1207 16:02:45.425189 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:45 crc kubenswrapper[4716]: I1207 16:02:45.425238 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:45 crc kubenswrapper[4716]: I1207 16:02:45.425264 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:45 crc kubenswrapper[4716]: I1207 16:02:45.425325 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:45 crc kubenswrapper[4716]: I1207 16:02:45.425343 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:45Z","lastTransitionTime":"2025-12-07T16:02:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:45 crc kubenswrapper[4716]: I1207 16:02:45.529007 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:45 crc kubenswrapper[4716]: I1207 16:02:45.529062 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:45 crc kubenswrapper[4716]: I1207 16:02:45.529120 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:45 crc kubenswrapper[4716]: I1207 16:02:45.529154 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:45 crc kubenswrapper[4716]: I1207 16:02:45.529176 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:45Z","lastTransitionTime":"2025-12-07T16:02:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:45 crc kubenswrapper[4716]: I1207 16:02:45.632272 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:45 crc kubenswrapper[4716]: I1207 16:02:45.632326 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:45 crc kubenswrapper[4716]: I1207 16:02:45.632348 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:45 crc kubenswrapper[4716]: I1207 16:02:45.632375 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:45 crc kubenswrapper[4716]: I1207 16:02:45.632397 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:45Z","lastTransitionTime":"2025-12-07T16:02:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:45 crc kubenswrapper[4716]: I1207 16:02:45.656889 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 16:02:45 crc kubenswrapper[4716]: E1207 16:02:45.657163 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 16:02:45 crc kubenswrapper[4716]: I1207 16:02:45.657826 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:02:45 crc kubenswrapper[4716]: E1207 16:02:45.657933 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 16:02:45 crc kubenswrapper[4716]: I1207 16:02:45.735578 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:45 crc kubenswrapper[4716]: I1207 16:02:45.735615 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:45 crc kubenswrapper[4716]: I1207 16:02:45.735636 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:45 crc kubenswrapper[4716]: I1207 16:02:45.735659 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:45 crc kubenswrapper[4716]: I1207 16:02:45.735675 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:45Z","lastTransitionTime":"2025-12-07T16:02:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:45 crc kubenswrapper[4716]: I1207 16:02:45.838839 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:45 crc kubenswrapper[4716]: I1207 16:02:45.838894 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:45 crc kubenswrapper[4716]: I1207 16:02:45.838912 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:45 crc kubenswrapper[4716]: I1207 16:02:45.838937 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:45 crc kubenswrapper[4716]: I1207 16:02:45.838956 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:45Z","lastTransitionTime":"2025-12-07T16:02:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:45 crc kubenswrapper[4716]: I1207 16:02:45.879179 4716 generic.go:334] "Generic (PLEG): container finished" podID="98015cea-d753-4e7a-b2b5-5ae7733a81e3" containerID="4c75bf2a6703cf99865d4ed5fa000e1e3b1330d4946176cf585fec1a2f226df0" exitCode=0 Dec 07 16:02:45 crc kubenswrapper[4716]: I1207 16:02:45.879244 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-8ps9j" event={"ID":"98015cea-d753-4e7a-b2b5-5ae7733a81e3","Type":"ContainerDied","Data":"4c75bf2a6703cf99865d4ed5fa000e1e3b1330d4946176cf585fec1a2f226df0"} Dec 07 16:02:45 crc kubenswrapper[4716]: I1207 16:02:45.925357 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7961cc9a-734e-42c2-b1ff-7075f3f20091\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84a36b76f20618c7885d611ecad9db29bac5101c27572be31d3acbb90c581291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0a0eca3c2087a457c0e448e649345ec4595a9c8a019aed2cddc5f30c77871e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://404426a171b8676eddd4a00996c76afab7e269643fa9608f1e1b6bdba385071d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fd3c2941782921f99a6cdc3de4e892234548bde527fec15a8ac92e513566e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e23abfdca314ed43d594333c9f9e401cc8ef7dad39b8ebac84a836b06ee5c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09b38cd4b389ea477c802a842ee4f2297a64d64684624117163e819539c257c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09b38cd4b389ea477c802a842ee4f2297a64d64684624117163e819539c257c6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c00f7a8e024e0e0dfd7b73e16d1d5437602e79ea95b73c21f24b209edef60e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c00f7a8e024e0e0dfd7b73e16d1d5437602e79ea95b73c21f24b209edef60e98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3f6392d426fa2cd788ec3a5c3bff4ac39483a54dfb2624a1c4370522413a350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3f6392d426fa2cd788ec3a5c3bff4ac39483a54dfb2624a1c4370522413a350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:45Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:45 crc kubenswrapper[4716]: I1207 16:02:45.942355 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:45 crc kubenswrapper[4716]: I1207 16:02:45.942409 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:45 crc kubenswrapper[4716]: I1207 16:02:45.942424 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:45 crc kubenswrapper[4716]: I1207 16:02:45.942472 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:45 crc kubenswrapper[4716]: I1207 16:02:45.942489 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:45Z","lastTransitionTime":"2025-12-07T16:02:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:45 crc kubenswrapper[4716]: I1207 16:02:45.951026 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3067de7c-c11c-4be1-bd94-d6572c266795\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d14d018fa044e9e80fe81e7e38df935992d14c773ed2b2193fd5cb873e5d0947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a9e9cf048220ca14c996847e9ced9f2367b9331dc6c2836dd023658855cd79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6d3c911874095b3c7cb062d03070e5090a7bb2456cb20cb4ce078f502515cda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ac6ce7701f1c6aa1ea8fcf7a3035e3b0eaa28ae3cb6afc442372bd9117bdde4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:45Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:45 crc kubenswrapper[4716]: I1207 16:02:45.966590 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4742ea84b3a2c9461d9b900beee7354a2ded24c39ad23e007cbf270ae12377c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:45Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:45 crc kubenswrapper[4716]: I1207 16:02:45.981257 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:45Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:45 crc kubenswrapper[4716]: I1207 16:02:45.997859 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c15b59eb-565d-4556-a4ce-75afdf159dc8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://530a2935826940767ea33fb4f37fa11f7dee4950d867fbcd6af72836d80f7db3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jks8z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ccaf6c1e6a978eb978a1c207b6c4c00b23737f78b743f966579aca781e22e0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jks8z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zcxxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:45Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.019051 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8ps9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98015cea-d753-4e7a-b2b5-5ae7733a81e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f84f52a073437543721010dc440339654ffb5e04eeb6cc7afbcc62f00f64c8e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f84f52a073437543721010dc440339654ffb5e04eeb6cc7afbcc62f00f64c8e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f6fe1cb25218cb3ac20488dc47efa4dc6a26a43dbc8306b5ee38cbe1a186c67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f6fe1cb25218cb3ac20488dc47efa4dc6a26a43dbc8306b5ee38cbe1a186c67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14b748a8247f6863e8be61e90057a06626dfd3cd4f01bd3a11629f700f56378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f14b748a8247f6863e8be61e90057a06626dfd3cd4f01bd3a11629f700f56378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ec379c0d573cb9b5ffcfc5ca7c653ba60e2fbdd7e3d37836102e26618ce43e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ec379c0d573cb9b5ffcfc5ca7c653ba60e2fbdd7e3d37836102e26618ce43e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c75bf2a6703cf99865d4ed5fa000e1e3b1330d4946176cf585fec1a2f226df0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c75bf2a6703cf99865d4ed5fa000e1e3b1330d4946176cf585fec1a2f226df0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8ps9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:46Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.041613 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c54d618a-dfec-4f2a-a288-f5188c8a128a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9bfe2ffce7a7b7c320474117734fc05246cc2b4fd744833e896e8e7c83f5e4d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9bfe2ffce7a7b7c320474117734fc05246cc2b4fd744833e896e8e7c83f5e4d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qvc7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:46Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.045568 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.045610 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.045620 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.045637 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.045793 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:46Z","lastTransitionTime":"2025-12-07T16:02:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.060989 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdee9957-48b9-458f-ba71-e0cd39f80596\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73f85248b4d68a9beecdacb3a01f433e97a0ffea2247ce92a41ef96855509377\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6276df652efe6ca935e745f24ef20f87a16465475139230c63b9999234a4d12a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3b1dd202586ee755ca89fbe12b30bbb12f95a62514f6e4d93cbbdbe55181e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd318647816e4a27849666b0ec4fc2ca4905295661c557dc381e1144f22cf284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d31de0326a05216fa10442bdb47151c6e073c97e9a16c34e793028e5e253756\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"le observer\\\\nW1207 16:02:34.645464 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1207 16:02:34.645587 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 16:02:34.646501 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2082283479/tls.crt::/tmp/serving-cert-2082283479/tls.key\\\\\\\"\\\\nI1207 16:02:34.972041 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 16:02:34.973702 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 16:02:34.973717 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 16:02:34.973738 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 16:02:34.973743 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 16:02:34.977185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1207 16:02:34.977204 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1207 16:02:34.977209 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977217 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 16:02:34.977225 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 16:02:34.977228 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 16:02:34.977232 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1207 16:02:34.979033 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27bd4a54f920b2725987c1c7fd2d9fdf16cd08738c246c333a4e9164a3a582a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:46Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.078056 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:46Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.087971 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2jjld" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dee6f5c-861c-4819-9309-7791574c7b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51c1d4159ab11087839a2c2d192566fb9322447c26df560bf09517c906ececb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmdns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2jjld\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:46Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.105340 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4wr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0baf8575-26cf-4a72-8a6d-ae94c078eaad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0ca2b3cd5a11033afd219b0125be7ab5acd750935d7eb89a470eff3d93a0838\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p2trd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4wr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:46Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.117795 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-m5jl5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"824ee59f-1474-4994-a70e-462878521514\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a15fe50cc75515704779646cc19aa50004f0db9aa26fe2535110df749bbfb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhwzn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-m5jl5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:46Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.135243 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef3335ae3875a75ab64685fc7e3590177280d651e8f8292c6f5ea6f472e55789\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:46Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.147196 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.147230 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.147239 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.147252 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.147262 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:46Z","lastTransitionTime":"2025-12-07T16:02:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.151594 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:46Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.163340 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.163381 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.163393 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.163411 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.163426 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:46Z","lastTransitionTime":"2025-12-07T16:02:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.166331 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f95ef61779507288a03bda8eddf745e105f705998a117a30789675bfa452d88f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b955b23163e7c8c458a3bbe2f681633e6a663ce76748febcc0b3a3c37f3ca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:46Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:46 crc kubenswrapper[4716]: E1207 16:02:46.179041 4716 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a8d1b7d4-7ec4-40c7-86b0-291db1f38894\\\",\\\"systemUUID\\\":\\\"a57d09b0-0224-46eb-a3ac-a2b60dad2250\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:46Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.182812 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.182865 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.182880 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.182902 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.182918 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:46Z","lastTransitionTime":"2025-12-07T16:02:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:46 crc kubenswrapper[4716]: E1207 16:02:46.196378 4716 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a8d1b7d4-7ec4-40c7-86b0-291db1f38894\\\",\\\"systemUUID\\\":\\\"a57d09b0-0224-46eb-a3ac-a2b60dad2250\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:46Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.200002 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.200026 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.200035 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.200049 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.200058 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:46Z","lastTransitionTime":"2025-12-07T16:02:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:46 crc kubenswrapper[4716]: E1207 16:02:46.215486 4716 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a8d1b7d4-7ec4-40c7-86b0-291db1f38894\\\",\\\"systemUUID\\\":\\\"a57d09b0-0224-46eb-a3ac-a2b60dad2250\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:46Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.218822 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.218852 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.218861 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.218875 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.218887 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:46Z","lastTransitionTime":"2025-12-07T16:02:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:46 crc kubenswrapper[4716]: E1207 16:02:46.229719 4716 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a8d1b7d4-7ec4-40c7-86b0-291db1f38894\\\",\\\"systemUUID\\\":\\\"a57d09b0-0224-46eb-a3ac-a2b60dad2250\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:46Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.233060 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.233104 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.233112 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.233146 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.233157 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:46Z","lastTransitionTime":"2025-12-07T16:02:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:46 crc kubenswrapper[4716]: E1207 16:02:46.249450 4716 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a8d1b7d4-7ec4-40c7-86b0-291db1f38894\\\",\\\"systemUUID\\\":\\\"a57d09b0-0224-46eb-a3ac-a2b60dad2250\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:46Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:46 crc kubenswrapper[4716]: E1207 16:02:46.249605 4716 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.250943 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.250967 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.250975 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.250985 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.250992 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:46Z","lastTransitionTime":"2025-12-07T16:02:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.353361 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.353401 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.353412 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.353431 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.353443 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:46Z","lastTransitionTime":"2025-12-07T16:02:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.457054 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.457190 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.457217 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.457262 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.457290 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:46Z","lastTransitionTime":"2025-12-07T16:02:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.559576 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.559619 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.559636 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.559655 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.559699 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:46Z","lastTransitionTime":"2025-12-07T16:02:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.657436 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 16:02:46 crc kubenswrapper[4716]: E1207 16:02:46.657558 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.662354 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.662386 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.662398 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.662419 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.662434 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:46Z","lastTransitionTime":"2025-12-07T16:02:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.767738 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.767788 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.767798 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.767813 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.767822 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:46Z","lastTransitionTime":"2025-12-07T16:02:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.872058 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.872503 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.872528 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.872558 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.872575 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:46Z","lastTransitionTime":"2025-12-07T16:02:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.887463 4716 generic.go:334] "Generic (PLEG): container finished" podID="98015cea-d753-4e7a-b2b5-5ae7733a81e3" containerID="de45812c9a5a5bfe6b0146828d03fab28362b03d04ef861c150b9ffbf181d345" exitCode=0 Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.887507 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-8ps9j" event={"ID":"98015cea-d753-4e7a-b2b5-5ae7733a81e3","Type":"ContainerDied","Data":"de45812c9a5a5bfe6b0146828d03fab28362b03d04ef861c150b9ffbf181d345"} Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.914639 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:46Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.933024 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f95ef61779507288a03bda8eddf745e105f705998a117a30789675bfa452d88f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b955b23163e7c8c458a3bbe2f681633e6a663ce76748febcc0b3a3c37f3ca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:46Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.948474 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef3335ae3875a75ab64685fc7e3590177280d651e8f8292c6f5ea6f472e55789\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:46Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.959249 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4742ea84b3a2c9461d9b900beee7354a2ded24c39ad23e007cbf270ae12377c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:46Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.971850 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:46Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.975280 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.975329 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.975345 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.975367 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.975382 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:46Z","lastTransitionTime":"2025-12-07T16:02:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:46 crc kubenswrapper[4716]: I1207 16:02:46.989680 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c15b59eb-565d-4556-a4ce-75afdf159dc8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://530a2935826940767ea33fb4f37fa11f7dee4950d867fbcd6af72836d80f7db3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jks8z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ccaf6c1e6a978eb978a1c207b6c4c00b23737f78b743f966579aca781e22e0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jks8z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zcxxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:46Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.014686 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8ps9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98015cea-d753-4e7a-b2b5-5ae7733a81e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f84f52a073437543721010dc440339654ffb5e04eeb6cc7afbcc62f00f64c8e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f84f52a073437543721010dc440339654ffb5e04eeb6cc7afbcc62f00f64c8e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f6fe1cb25218cb3ac20488dc47efa4dc6a26a43dbc8306b5ee38cbe1a186c67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f6fe1cb25218cb3ac20488dc47efa4dc6a26a43dbc8306b5ee38cbe1a186c67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14b748a8247f6863e8be61e90057a06626dfd3cd4f01bd3a11629f700f56378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f14b748a8247f6863e8be61e90057a06626dfd3cd4f01bd3a11629f700f56378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ec379c0d573cb9b5ffcfc5ca7c653ba60e2fbdd7e3d37836102e26618ce43e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ec379c0d573cb9b5ffcfc5ca7c653ba60e2fbdd7e3d37836102e26618ce43e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c75bf2a6703cf99865d4ed5fa000e1e3b1330d4946176cf585fec1a2f226df0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c75bf2a6703cf99865d4ed5fa000e1e3b1330d4946176cf585fec1a2f226df0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de45812c9a5a5bfe6b0146828d03fab28362b03d04ef861c150b9ffbf181d345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de45812c9a5a5bfe6b0146828d03fab28362b03d04ef861c150b9ffbf181d345\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8ps9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:47Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.044285 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7961cc9a-734e-42c2-b1ff-7075f3f20091\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84a36b76f20618c7885d611ecad9db29bac5101c27572be31d3acbb90c581291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0a0eca3c2087a457c0e448e649345ec4595a9c8a019aed2cddc5f30c77871e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://404426a171b8676eddd4a00996c76afab7e269643fa9608f1e1b6bdba385071d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fd3c2941782921f99a6cdc3de4e892234548bde527fec15a8ac92e513566e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e23abfdca314ed43d594333c9f9e401cc8ef7dad39b8ebac84a836b06ee5c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09b38cd4b389ea477c802a842ee4f2297a64d64684624117163e819539c257c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09b38cd4b389ea477c802a842ee4f2297a64d64684624117163e819539c257c6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c00f7a8e024e0e0dfd7b73e16d1d5437602e79ea95b73c21f24b209edef60e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c00f7a8e024e0e0dfd7b73e16d1d5437602e79ea95b73c21f24b209edef60e98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3f6392d426fa2cd788ec3a5c3bff4ac39483a54dfb2624a1c4370522413a350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3f6392d426fa2cd788ec3a5c3bff4ac39483a54dfb2624a1c4370522413a350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:47Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.059255 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3067de7c-c11c-4be1-bd94-d6572c266795\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d14d018fa044e9e80fe81e7e38df935992d14c773ed2b2193fd5cb873e5d0947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a9e9cf048220ca14c996847e9ced9f2367b9331dc6c2836dd023658855cd79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6d3c911874095b3c7cb062d03070e5090a7bb2456cb20cb4ce078f502515cda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ac6ce7701f1c6aa1ea8fcf7a3035e3b0eaa28ae3cb6afc442372bd9117bdde4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:47Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.078550 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.078648 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.078662 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.079276 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.079344 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:47Z","lastTransitionTime":"2025-12-07T16:02:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.089413 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c54d618a-dfec-4f2a-a288-f5188c8a128a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9bfe2ffce7a7b7c320474117734fc05246cc2b4fd744833e896e8e7c83f5e4d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9bfe2ffce7a7b7c320474117734fc05246cc2b4fd744833e896e8e7c83f5e4d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qvc7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:47Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.109932 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2jjld" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dee6f5c-861c-4819-9309-7791574c7b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51c1d4159ab11087839a2c2d192566fb9322447c26df560bf09517c906ececb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmdns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2jjld\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:47Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.129748 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4wr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0baf8575-26cf-4a72-8a6d-ae94c078eaad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0ca2b3cd5a11033afd219b0125be7ab5acd750935d7eb89a470eff3d93a0838\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p2trd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4wr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:47Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.138852 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-m5jl5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"824ee59f-1474-4994-a70e-462878521514\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a15fe50cc75515704779646cc19aa50004f0db9aa26fe2535110df749bbfb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhwzn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-m5jl5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:47Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.156393 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdee9957-48b9-458f-ba71-e0cd39f80596\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73f85248b4d68a9beecdacb3a01f433e97a0ffea2247ce92a41ef96855509377\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6276df652efe6ca935e745f24ef20f87a16465475139230c63b9999234a4d12a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3b1dd202586ee755ca89fbe12b30bbb12f95a62514f6e4d93cbbdbe55181e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd318647816e4a27849666b0ec4fc2ca4905295661c557dc381e1144f22cf284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d31de0326a05216fa10442bdb47151c6e073c97e9a16c34e793028e5e253756\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"le observer\\\\nW1207 16:02:34.645464 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1207 16:02:34.645587 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 16:02:34.646501 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2082283479/tls.crt::/tmp/serving-cert-2082283479/tls.key\\\\\\\"\\\\nI1207 16:02:34.972041 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 16:02:34.973702 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 16:02:34.973717 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 16:02:34.973738 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 16:02:34.973743 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 16:02:34.977185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1207 16:02:34.977204 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1207 16:02:34.977209 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977217 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 16:02:34.977225 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 16:02:34.977228 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 16:02:34.977232 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1207 16:02:34.979033 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27bd4a54f920b2725987c1c7fd2d9fdf16cd08738c246c333a4e9164a3a582a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:47Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.169500 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:47Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.183273 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.183341 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.183360 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.183387 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.183408 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:47Z","lastTransitionTime":"2025-12-07T16:02:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.285494 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.285538 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.285550 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.285567 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.285578 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:47Z","lastTransitionTime":"2025-12-07T16:02:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.388634 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.388684 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.388696 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.388714 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.388728 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:47Z","lastTransitionTime":"2025-12-07T16:02:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.446658 4716 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.496274 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.496338 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.496369 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.496397 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.496416 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:47Z","lastTransitionTime":"2025-12-07T16:02:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.600346 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.600410 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.600424 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.600444 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.600456 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:47Z","lastTransitionTime":"2025-12-07T16:02:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.657252 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.657267 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:02:47 crc kubenswrapper[4716]: E1207 16:02:47.657475 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 16:02:47 crc kubenswrapper[4716]: E1207 16:02:47.657615 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.676741 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdee9957-48b9-458f-ba71-e0cd39f80596\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73f85248b4d68a9beecdacb3a01f433e97a0ffea2247ce92a41ef96855509377\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6276df652efe6ca935e745f24ef20f87a16465475139230c63b9999234a4d12a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3b1dd202586ee755ca89fbe12b30bbb12f95a62514f6e4d93cbbdbe55181e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd318647816e4a27849666b0ec4fc2ca4905295661c557dc381e1144f22cf284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d31de0326a05216fa10442bdb47151c6e073c97e9a16c34e793028e5e253756\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"le observer\\\\nW1207 16:02:34.645464 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1207 16:02:34.645587 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 16:02:34.646501 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2082283479/tls.crt::/tmp/serving-cert-2082283479/tls.key\\\\\\\"\\\\nI1207 16:02:34.972041 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 16:02:34.973702 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 16:02:34.973717 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 16:02:34.973738 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 16:02:34.973743 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 16:02:34.977185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1207 16:02:34.977204 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1207 16:02:34.977209 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977217 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 16:02:34.977225 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 16:02:34.977228 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 16:02:34.977232 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1207 16:02:34.979033 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27bd4a54f920b2725987c1c7fd2d9fdf16cd08738c246c333a4e9164a3a582a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:47Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.693852 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:47Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.703609 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.703650 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.703661 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.703710 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.703722 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:47Z","lastTransitionTime":"2025-12-07T16:02:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.708401 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2jjld" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dee6f5c-861c-4819-9309-7791574c7b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51c1d4159ab11087839a2c2d192566fb9322447c26df560bf09517c906ececb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmdns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2jjld\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:47Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.734371 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4wr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0baf8575-26cf-4a72-8a6d-ae94c078eaad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0ca2b3cd5a11033afd219b0125be7ab5acd750935d7eb89a470eff3d93a0838\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p2trd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4wr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:47Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.751838 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-m5jl5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"824ee59f-1474-4994-a70e-462878521514\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a15fe50cc75515704779646cc19aa50004f0db9aa26fe2535110df749bbfb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhwzn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-m5jl5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:47Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.771792 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef3335ae3875a75ab64685fc7e3590177280d651e8f8292c6f5ea6f472e55789\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:47Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.786459 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:47Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.800105 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f95ef61779507288a03bda8eddf745e105f705998a117a30789675bfa452d88f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b955b23163e7c8c458a3bbe2f681633e6a663ce76748febcc0b3a3c37f3ca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:47Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.805584 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.805618 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.805627 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.805642 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.805651 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:47Z","lastTransitionTime":"2025-12-07T16:02:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.821625 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7961cc9a-734e-42c2-b1ff-7075f3f20091\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84a36b76f20618c7885d611ecad9db29bac5101c27572be31d3acbb90c581291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0a0eca3c2087a457c0e448e649345ec4595a9c8a019aed2cddc5f30c77871e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://404426a171b8676eddd4a00996c76afab7e269643fa9608f1e1b6bdba385071d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fd3c2941782921f99a6cdc3de4e892234548bde527fec15a8ac92e513566e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e23abfdca314ed43d594333c9f9e401cc8ef7dad39b8ebac84a836b06ee5c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09b38cd4b389ea477c802a842ee4f2297a64d64684624117163e819539c257c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09b38cd4b389ea477c802a842ee4f2297a64d64684624117163e819539c257c6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c00f7a8e024e0e0dfd7b73e16d1d5437602e79ea95b73c21f24b209edef60e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c00f7a8e024e0e0dfd7b73e16d1d5437602e79ea95b73c21f24b209edef60e98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3f6392d426fa2cd788ec3a5c3bff4ac39483a54dfb2624a1c4370522413a350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3f6392d426fa2cd788ec3a5c3bff4ac39483a54dfb2624a1c4370522413a350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:47Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.835322 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3067de7c-c11c-4be1-bd94-d6572c266795\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d14d018fa044e9e80fe81e7e38df935992d14c773ed2b2193fd5cb873e5d0947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a9e9cf048220ca14c996847e9ced9f2367b9331dc6c2836dd023658855cd79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6d3c911874095b3c7cb062d03070e5090a7bb2456cb20cb4ce078f502515cda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ac6ce7701f1c6aa1ea8fcf7a3035e3b0eaa28ae3cb6afc442372bd9117bdde4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:47Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.901064 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" event={"ID":"c54d618a-dfec-4f2a-a288-f5188c8a128a","Type":"ContainerStarted","Data":"53c40176ba965341b13456a99ba4fad80dc6115939e3c4c884e74f4e026a7535"} Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.901288 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.901345 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.901354 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.904255 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4742ea84b3a2c9461d9b900beee7354a2ded24c39ad23e007cbf270ae12377c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:47Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.906549 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-8ps9j" event={"ID":"98015cea-d753-4e7a-b2b5-5ae7733a81e3","Type":"ContainerStarted","Data":"d68509b673eecd37725cdcc48e3d0612c3b9c7601433f2b72c21d662140daa59"} Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.907541 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.907574 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.907587 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.907609 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.907621 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:47Z","lastTransitionTime":"2025-12-07T16:02:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.915723 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:47Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.928046 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.928304 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c15b59eb-565d-4556-a4ce-75afdf159dc8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://530a2935826940767ea33fb4f37fa11f7dee4950d867fbcd6af72836d80f7db3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jks8z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ccaf6c1e6a978eb978a1c207b6c4c00b23737f78b743f966579aca781e22e0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jks8z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zcxxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:47Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.928671 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.946046 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8ps9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98015cea-d753-4e7a-b2b5-5ae7733a81e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f84f52a073437543721010dc440339654ffb5e04eeb6cc7afbcc62f00f64c8e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f84f52a073437543721010dc440339654ffb5e04eeb6cc7afbcc62f00f64c8e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f6fe1cb25218cb3ac20488dc47efa4dc6a26a43dbc8306b5ee38cbe1a186c67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f6fe1cb25218cb3ac20488dc47efa4dc6a26a43dbc8306b5ee38cbe1a186c67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14b748a8247f6863e8be61e90057a06626dfd3cd4f01bd3a11629f700f56378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f14b748a8247f6863e8be61e90057a06626dfd3cd4f01bd3a11629f700f56378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ec379c0d573cb9b5ffcfc5ca7c653ba60e2fbdd7e3d37836102e26618ce43e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ec379c0d573cb9b5ffcfc5ca7c653ba60e2fbdd7e3d37836102e26618ce43e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c75bf2a6703cf99865d4ed5fa000e1e3b1330d4946176cf585fec1a2f226df0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c75bf2a6703cf99865d4ed5fa000e1e3b1330d4946176cf585fec1a2f226df0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de45812c9a5a5bfe6b0146828d03fab28362b03d04ef861c150b9ffbf181d345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de45812c9a5a5bfe6b0146828d03fab28362b03d04ef861c150b9ffbf181d345\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8ps9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:47Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.967060 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c54d618a-dfec-4f2a-a288-f5188c8a128a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9bfe2ffce7a7b7c320474117734fc05246cc2b4fd744833e896e8e7c83f5e4d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9bfe2ffce7a7b7c320474117734fc05246cc2b4fd744833e896e8e7c83f5e4d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qvc7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:47Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.975200 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.981451 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdee9957-48b9-458f-ba71-e0cd39f80596\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73f85248b4d68a9beecdacb3a01f433e97a0ffea2247ce92a41ef96855509377\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6276df652efe6ca935e745f24ef20f87a16465475139230c63b9999234a4d12a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3b1dd202586ee755ca89fbe12b30bbb12f95a62514f6e4d93cbbdbe55181e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd318647816e4a27849666b0ec4fc2ca4905295661c557dc381e1144f22cf284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d31de0326a05216fa10442bdb47151c6e073c97e9a16c34e793028e5e253756\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"le observer\\\\nW1207 16:02:34.645464 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1207 16:02:34.645587 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 16:02:34.646501 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2082283479/tls.crt::/tmp/serving-cert-2082283479/tls.key\\\\\\\"\\\\nI1207 16:02:34.972041 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 16:02:34.973702 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 16:02:34.973717 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 16:02:34.973738 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 16:02:34.973743 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 16:02:34.977185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1207 16:02:34.977204 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1207 16:02:34.977209 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977217 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 16:02:34.977225 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 16:02:34.977228 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 16:02:34.977232 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1207 16:02:34.979033 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27bd4a54f920b2725987c1c7fd2d9fdf16cd08738c246c333a4e9164a3a582a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:47Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:47 crc kubenswrapper[4716]: I1207 16:02:47.991740 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:47Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.002974 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2jjld" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dee6f5c-861c-4819-9309-7791574c7b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51c1d4159ab11087839a2c2d192566fb9322447c26df560bf09517c906ececb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmdns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2jjld\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:48Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.009945 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.009985 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.009999 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.010018 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.010030 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:48Z","lastTransitionTime":"2025-12-07T16:02:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.014781 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4wr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0baf8575-26cf-4a72-8a6d-ae94c078eaad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0ca2b3cd5a11033afd219b0125be7ab5acd750935d7eb89a470eff3d93a0838\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p2trd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4wr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:48Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.027450 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-m5jl5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"824ee59f-1474-4994-a70e-462878521514\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a15fe50cc75515704779646cc19aa50004f0db9aa26fe2535110df749bbfb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhwzn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-m5jl5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:48Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.038771 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef3335ae3875a75ab64685fc7e3590177280d651e8f8292c6f5ea6f472e55789\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:48Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.054684 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:48Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.070516 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f95ef61779507288a03bda8eddf745e105f705998a117a30789675bfa452d88f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b955b23163e7c8c458a3bbe2f681633e6a663ce76748febcc0b3a3c37f3ca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:48Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.089580 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7961cc9a-734e-42c2-b1ff-7075f3f20091\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84a36b76f20618c7885d611ecad9db29bac5101c27572be31d3acbb90c581291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0a0eca3c2087a457c0e448e649345ec4595a9c8a019aed2cddc5f30c77871e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://404426a171b8676eddd4a00996c76afab7e269643fa9608f1e1b6bdba385071d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fd3c2941782921f99a6cdc3de4e892234548bde527fec15a8ac92e513566e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e23abfdca314ed43d594333c9f9e401cc8ef7dad39b8ebac84a836b06ee5c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09b38cd4b389ea477c802a842ee4f2297a64d64684624117163e819539c257c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09b38cd4b389ea477c802a842ee4f2297a64d64684624117163e819539c257c6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c00f7a8e024e0e0dfd7b73e16d1d5437602e79ea95b73c21f24b209edef60e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c00f7a8e024e0e0dfd7b73e16d1d5437602e79ea95b73c21f24b209edef60e98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3f6392d426fa2cd788ec3a5c3bff4ac39483a54dfb2624a1c4370522413a350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3f6392d426fa2cd788ec3a5c3bff4ac39483a54dfb2624a1c4370522413a350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:48Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.106062 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3067de7c-c11c-4be1-bd94-d6572c266795\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d14d018fa044e9e80fe81e7e38df935992d14c773ed2b2193fd5cb873e5d0947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a9e9cf048220ca14c996847e9ced9f2367b9331dc6c2836dd023658855cd79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6d3c911874095b3c7cb062d03070e5090a7bb2456cb20cb4ce078f502515cda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ac6ce7701f1c6aa1ea8fcf7a3035e3b0eaa28ae3cb6afc442372bd9117bdde4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:48Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.113135 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.113215 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.113268 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.113302 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.113325 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:48Z","lastTransitionTime":"2025-12-07T16:02:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.121153 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4742ea84b3a2c9461d9b900beee7354a2ded24c39ad23e007cbf270ae12377c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:48Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.138052 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:48Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.154718 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c15b59eb-565d-4556-a4ce-75afdf159dc8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://530a2935826940767ea33fb4f37fa11f7dee4950d867fbcd6af72836d80f7db3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jks8z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ccaf6c1e6a978eb978a1c207b6c4c00b23737f78b743f966579aca781e22e0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jks8z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zcxxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:48Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.171928 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8ps9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98015cea-d753-4e7a-b2b5-5ae7733a81e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d68509b673eecd37725cdcc48e3d0612c3b9c7601433f2b72c21d662140daa59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f84f52a073437543721010dc440339654ffb5e04eeb6cc7afbcc62f00f64c8e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f84f52a073437543721010dc440339654ffb5e04eeb6cc7afbcc62f00f64c8e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f6fe1cb25218cb3ac20488dc47efa4dc6a26a43dbc8306b5ee38cbe1a186c67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f6fe1cb25218cb3ac20488dc47efa4dc6a26a43dbc8306b5ee38cbe1a186c67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14b748a8247f6863e8be61e90057a06626dfd3cd4f01bd3a11629f700f56378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f14b748a8247f6863e8be61e90057a06626dfd3cd4f01bd3a11629f700f56378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ec379c0d573cb9b5ffcfc5ca7c653ba60e2fbdd7e3d37836102e26618ce43e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ec379c0d573cb9b5ffcfc5ca7c653ba60e2fbdd7e3d37836102e26618ce43e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c75bf2a6703cf99865d4ed5fa000e1e3b1330d4946176cf585fec1a2f226df0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c75bf2a6703cf99865d4ed5fa000e1e3b1330d4946176cf585fec1a2f226df0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de45812c9a5a5bfe6b0146828d03fab28362b03d04ef861c150b9ffbf181d345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de45812c9a5a5bfe6b0146828d03fab28362b03d04ef861c150b9ffbf181d345\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8ps9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:48Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.202748 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c54d618a-dfec-4f2a-a288-f5188c8a128a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60324dc7d701bc064e14ae4f474bce10f6ee4ff3c3c942eab749e7166082cf9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ed4f4d3564674bf5cb4e7a76dc0fd95e203ecc8950798e59a1fb6c9ab4a20f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://def9ac840fb8ecf2778bed88ea980eddca7c5b2aba4bd1afb967d62fec3f7bae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b73be664f5412c98b89d971a2cac37c494151c293fdb44a98b390978658a6643\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a847d04e22e98091c0037a89054b9236228b701b8c7ba31940a56df0f37b95c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcae78a6133d97f6a32dd7421ccec4333b970a237404e056a5cc601f8eefdc8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://53c40176ba965341b13456a99ba4fad80dc6115939e3c4c884e74f4e026a7535\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcfdffe822635b0ae13b253e6133449fe00461b2792be43d913c4a77219b8255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9bfe2ffce7a7b7c320474117734fc05246cc2b4fd744833e896e8e7c83f5e4d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9bfe2ffce7a7b7c320474117734fc05246cc2b4fd744833e896e8e7c83f5e4d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qvc7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:48Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.216942 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.217004 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.217023 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.217054 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.217073 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:48Z","lastTransitionTime":"2025-12-07T16:02:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.218311 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdee9957-48b9-458f-ba71-e0cd39f80596\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73f85248b4d68a9beecdacb3a01f433e97a0ffea2247ce92a41ef96855509377\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6276df652efe6ca935e745f24ef20f87a16465475139230c63b9999234a4d12a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3b1dd202586ee755ca89fbe12b30bbb12f95a62514f6e4d93cbbdbe55181e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd318647816e4a27849666b0ec4fc2ca4905295661c557dc381e1144f22cf284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d31de0326a05216fa10442bdb47151c6e073c97e9a16c34e793028e5e253756\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"le observer\\\\nW1207 16:02:34.645464 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1207 16:02:34.645587 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 16:02:34.646501 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2082283479/tls.crt::/tmp/serving-cert-2082283479/tls.key\\\\\\\"\\\\nI1207 16:02:34.972041 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 16:02:34.973702 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 16:02:34.973717 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 16:02:34.973738 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 16:02:34.973743 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 16:02:34.977185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1207 16:02:34.977204 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1207 16:02:34.977209 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977217 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 16:02:34.977225 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 16:02:34.977228 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 16:02:34.977232 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1207 16:02:34.979033 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27bd4a54f920b2725987c1c7fd2d9fdf16cd08738c246c333a4e9164a3a582a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:48Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.232691 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:48Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.245393 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2jjld" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dee6f5c-861c-4819-9309-7791574c7b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51c1d4159ab11087839a2c2d192566fb9322447c26df560bf09517c906ececb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmdns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2jjld\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:48Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.259196 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4wr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0baf8575-26cf-4a72-8a6d-ae94c078eaad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0ca2b3cd5a11033afd219b0125be7ab5acd750935d7eb89a470eff3d93a0838\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p2trd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4wr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:48Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.278203 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-m5jl5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"824ee59f-1474-4994-a70e-462878521514\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a15fe50cc75515704779646cc19aa50004f0db9aa26fe2535110df749bbfb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhwzn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-m5jl5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:48Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.291473 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef3335ae3875a75ab64685fc7e3590177280d651e8f8292c6f5ea6f472e55789\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:48Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.304019 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:48Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.314685 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f95ef61779507288a03bda8eddf745e105f705998a117a30789675bfa452d88f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b955b23163e7c8c458a3bbe2f681633e6a663ce76748febcc0b3a3c37f3ca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:48Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.319609 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.319670 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.319692 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.319726 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.319750 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:48Z","lastTransitionTime":"2025-12-07T16:02:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.332775 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8ps9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98015cea-d753-4e7a-b2b5-5ae7733a81e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d68509b673eecd37725cdcc48e3d0612c3b9c7601433f2b72c21d662140daa59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f84f52a073437543721010dc440339654ffb5e04eeb6cc7afbcc62f00f64c8e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f84f52a073437543721010dc440339654ffb5e04eeb6cc7afbcc62f00f64c8e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f6fe1cb25218cb3ac20488dc47efa4dc6a26a43dbc8306b5ee38cbe1a186c67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f6fe1cb25218cb3ac20488dc47efa4dc6a26a43dbc8306b5ee38cbe1a186c67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14b748a8247f6863e8be61e90057a06626dfd3cd4f01bd3a11629f700f56378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f14b748a8247f6863e8be61e90057a06626dfd3cd4f01bd3a11629f700f56378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ec379c0d573cb9b5ffcfc5ca7c653ba60e2fbdd7e3d37836102e26618ce43e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ec379c0d573cb9b5ffcfc5ca7c653ba60e2fbdd7e3d37836102e26618ce43e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c75bf2a6703cf99865d4ed5fa000e1e3b1330d4946176cf585fec1a2f226df0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c75bf2a6703cf99865d4ed5fa000e1e3b1330d4946176cf585fec1a2f226df0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de45812c9a5a5bfe6b0146828d03fab28362b03d04ef861c150b9ffbf181d345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de45812c9a5a5bfe6b0146828d03fab28362b03d04ef861c150b9ffbf181d345\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8ps9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:48Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.353341 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7961cc9a-734e-42c2-b1ff-7075f3f20091\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84a36b76f20618c7885d611ecad9db29bac5101c27572be31d3acbb90c581291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0a0eca3c2087a457c0e448e649345ec4595a9c8a019aed2cddc5f30c77871e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://404426a171b8676eddd4a00996c76afab7e269643fa9608f1e1b6bdba385071d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fd3c2941782921f99a6cdc3de4e892234548bde527fec15a8ac92e513566e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e23abfdca314ed43d594333c9f9e401cc8ef7dad39b8ebac84a836b06ee5c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09b38cd4b389ea477c802a842ee4f2297a64d64684624117163e819539c257c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09b38cd4b389ea477c802a842ee4f2297a64d64684624117163e819539c257c6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c00f7a8e024e0e0dfd7b73e16d1d5437602e79ea95b73c21f24b209edef60e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c00f7a8e024e0e0dfd7b73e16d1d5437602e79ea95b73c21f24b209edef60e98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3f6392d426fa2cd788ec3a5c3bff4ac39483a54dfb2624a1c4370522413a350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3f6392d426fa2cd788ec3a5c3bff4ac39483a54dfb2624a1c4370522413a350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:48Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.365713 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3067de7c-c11c-4be1-bd94-d6572c266795\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d14d018fa044e9e80fe81e7e38df935992d14c773ed2b2193fd5cb873e5d0947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a9e9cf048220ca14c996847e9ced9f2367b9331dc6c2836dd023658855cd79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6d3c911874095b3c7cb062d03070e5090a7bb2456cb20cb4ce078f502515cda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ac6ce7701f1c6aa1ea8fcf7a3035e3b0eaa28ae3cb6afc442372bd9117bdde4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:48Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.376471 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4742ea84b3a2c9461d9b900beee7354a2ded24c39ad23e007cbf270ae12377c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:48Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.386593 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:48Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.395441 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c15b59eb-565d-4556-a4ce-75afdf159dc8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://530a2935826940767ea33fb4f37fa11f7dee4950d867fbcd6af72836d80f7db3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jks8z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ccaf6c1e6a978eb978a1c207b6c4c00b23737f78b743f966579aca781e22e0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jks8z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zcxxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:48Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.409680 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c54d618a-dfec-4f2a-a288-f5188c8a128a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60324dc7d701bc064e14ae4f474bce10f6ee4ff3c3c942eab749e7166082cf9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ed4f4d3564674bf5cb4e7a76dc0fd95e203ecc8950798e59a1fb6c9ab4a20f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://def9ac840fb8ecf2778bed88ea980eddca7c5b2aba4bd1afb967d62fec3f7bae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b73be664f5412c98b89d971a2cac37c494151c293fdb44a98b390978658a6643\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a847d04e22e98091c0037a89054b9236228b701b8c7ba31940a56df0f37b95c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcae78a6133d97f6a32dd7421ccec4333b970a237404e056a5cc601f8eefdc8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://53c40176ba965341b13456a99ba4fad80dc6115939e3c4c884e74f4e026a7535\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcfdffe822635b0ae13b253e6133449fe00461b2792be43d913c4a77219b8255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9bfe2ffce7a7b7c320474117734fc05246cc2b4fd744833e896e8e7c83f5e4d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9bfe2ffce7a7b7c320474117734fc05246cc2b4fd744833e896e8e7c83f5e4d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qvc7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:48Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.422315 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.422350 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.422359 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.422374 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.422384 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:48Z","lastTransitionTime":"2025-12-07T16:02:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.526416 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.526641 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.526777 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.526826 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.526876 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:48Z","lastTransitionTime":"2025-12-07T16:02:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.630398 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.630435 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.630444 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.630457 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.630468 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:48Z","lastTransitionTime":"2025-12-07T16:02:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.657259 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 16:02:48 crc kubenswrapper[4716]: E1207 16:02:48.657402 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.732427 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.732454 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.732477 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.732490 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.732498 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:48Z","lastTransitionTime":"2025-12-07T16:02:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.835829 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.836092 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.836182 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.836289 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.836421 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:48Z","lastTransitionTime":"2025-12-07T16:02:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.952008 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.952531 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.952644 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.952736 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:48 crc kubenswrapper[4716]: I1207 16:02:48.952814 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:48Z","lastTransitionTime":"2025-12-07T16:02:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:49 crc kubenswrapper[4716]: I1207 16:02:49.055972 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:49 crc kubenswrapper[4716]: I1207 16:02:49.056028 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:49 crc kubenswrapper[4716]: I1207 16:02:49.056046 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:49 crc kubenswrapper[4716]: I1207 16:02:49.056070 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:49 crc kubenswrapper[4716]: I1207 16:02:49.056123 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:49Z","lastTransitionTime":"2025-12-07T16:02:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:49 crc kubenswrapper[4716]: I1207 16:02:49.162768 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:49 crc kubenswrapper[4716]: I1207 16:02:49.163362 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:49 crc kubenswrapper[4716]: I1207 16:02:49.163375 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:49 crc kubenswrapper[4716]: I1207 16:02:49.163412 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:49 crc kubenswrapper[4716]: I1207 16:02:49.163422 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:49Z","lastTransitionTime":"2025-12-07T16:02:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:49 crc kubenswrapper[4716]: I1207 16:02:49.265866 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:49 crc kubenswrapper[4716]: I1207 16:02:49.266095 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:49 crc kubenswrapper[4716]: I1207 16:02:49.266106 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:49 crc kubenswrapper[4716]: I1207 16:02:49.266121 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:49 crc kubenswrapper[4716]: I1207 16:02:49.266132 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:49Z","lastTransitionTime":"2025-12-07T16:02:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:49 crc kubenswrapper[4716]: I1207 16:02:49.368275 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:49 crc kubenswrapper[4716]: I1207 16:02:49.368318 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:49 crc kubenswrapper[4716]: I1207 16:02:49.368330 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:49 crc kubenswrapper[4716]: I1207 16:02:49.368347 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:49 crc kubenswrapper[4716]: I1207 16:02:49.368361 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:49Z","lastTransitionTime":"2025-12-07T16:02:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:49 crc kubenswrapper[4716]: I1207 16:02:49.471132 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:49 crc kubenswrapper[4716]: I1207 16:02:49.471172 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:49 crc kubenswrapper[4716]: I1207 16:02:49.471182 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:49 crc kubenswrapper[4716]: I1207 16:02:49.471197 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:49 crc kubenswrapper[4716]: I1207 16:02:49.471208 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:49Z","lastTransitionTime":"2025-12-07T16:02:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:49 crc kubenswrapper[4716]: I1207 16:02:49.574323 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:49 crc kubenswrapper[4716]: I1207 16:02:49.574356 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:49 crc kubenswrapper[4716]: I1207 16:02:49.574364 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:49 crc kubenswrapper[4716]: I1207 16:02:49.574376 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:49 crc kubenswrapper[4716]: I1207 16:02:49.574385 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:49Z","lastTransitionTime":"2025-12-07T16:02:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:49 crc kubenswrapper[4716]: I1207 16:02:49.657325 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:02:49 crc kubenswrapper[4716]: E1207 16:02:49.657501 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 16:02:49 crc kubenswrapper[4716]: I1207 16:02:49.658046 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 16:02:49 crc kubenswrapper[4716]: E1207 16:02:49.658226 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 16:02:49 crc kubenswrapper[4716]: I1207 16:02:49.677263 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:49 crc kubenswrapper[4716]: I1207 16:02:49.677323 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:49 crc kubenswrapper[4716]: I1207 16:02:49.677343 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:49 crc kubenswrapper[4716]: I1207 16:02:49.677367 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:49 crc kubenswrapper[4716]: I1207 16:02:49.677386 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:49Z","lastTransitionTime":"2025-12-07T16:02:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:49 crc kubenswrapper[4716]: I1207 16:02:49.780632 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:49 crc kubenswrapper[4716]: I1207 16:02:49.780700 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:49 crc kubenswrapper[4716]: I1207 16:02:49.780717 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:49 crc kubenswrapper[4716]: I1207 16:02:49.780737 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:49 crc kubenswrapper[4716]: I1207 16:02:49.780754 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:49Z","lastTransitionTime":"2025-12-07T16:02:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:49 crc kubenswrapper[4716]: I1207 16:02:49.885153 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:49 crc kubenswrapper[4716]: I1207 16:02:49.885253 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:49 crc kubenswrapper[4716]: I1207 16:02:49.885274 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:49 crc kubenswrapper[4716]: I1207 16:02:49.885304 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:49 crc kubenswrapper[4716]: I1207 16:02:49.885330 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:49Z","lastTransitionTime":"2025-12-07T16:02:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:49 crc kubenswrapper[4716]: I1207 16:02:49.989548 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:49 crc kubenswrapper[4716]: I1207 16:02:49.989622 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:49 crc kubenswrapper[4716]: I1207 16:02:49.989642 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:49 crc kubenswrapper[4716]: I1207 16:02:49.989671 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:49 crc kubenswrapper[4716]: I1207 16:02:49.989757 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:49Z","lastTransitionTime":"2025-12-07T16:02:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:50 crc kubenswrapper[4716]: I1207 16:02:50.092802 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:50 crc kubenswrapper[4716]: I1207 16:02:50.092865 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:50 crc kubenswrapper[4716]: I1207 16:02:50.092877 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:50 crc kubenswrapper[4716]: I1207 16:02:50.092898 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:50 crc kubenswrapper[4716]: I1207 16:02:50.092913 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:50Z","lastTransitionTime":"2025-12-07T16:02:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:50 crc kubenswrapper[4716]: I1207 16:02:50.195925 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:50 crc kubenswrapper[4716]: I1207 16:02:50.196018 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:50 crc kubenswrapper[4716]: I1207 16:02:50.196037 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:50 crc kubenswrapper[4716]: I1207 16:02:50.196059 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:50 crc kubenswrapper[4716]: I1207 16:02:50.196098 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:50Z","lastTransitionTime":"2025-12-07T16:02:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:50 crc kubenswrapper[4716]: I1207 16:02:50.299189 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:50 crc kubenswrapper[4716]: I1207 16:02:50.299245 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:50 crc kubenswrapper[4716]: I1207 16:02:50.299267 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:50 crc kubenswrapper[4716]: I1207 16:02:50.299291 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:50 crc kubenswrapper[4716]: I1207 16:02:50.299309 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:50Z","lastTransitionTime":"2025-12-07T16:02:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:50 crc kubenswrapper[4716]: I1207 16:02:50.402867 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:50 crc kubenswrapper[4716]: I1207 16:02:50.402982 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:50 crc kubenswrapper[4716]: I1207 16:02:50.403004 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:50 crc kubenswrapper[4716]: I1207 16:02:50.403035 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:50 crc kubenswrapper[4716]: I1207 16:02:50.403057 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:50Z","lastTransitionTime":"2025-12-07T16:02:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:50 crc kubenswrapper[4716]: I1207 16:02:50.505527 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:50 crc kubenswrapper[4716]: I1207 16:02:50.505563 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:50 crc kubenswrapper[4716]: I1207 16:02:50.505571 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:50 crc kubenswrapper[4716]: I1207 16:02:50.505584 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:50 crc kubenswrapper[4716]: I1207 16:02:50.505593 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:50Z","lastTransitionTime":"2025-12-07T16:02:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:50 crc kubenswrapper[4716]: I1207 16:02:50.607839 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:50 crc kubenswrapper[4716]: I1207 16:02:50.607910 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:50 crc kubenswrapper[4716]: I1207 16:02:50.607924 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:50 crc kubenswrapper[4716]: I1207 16:02:50.607940 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:50 crc kubenswrapper[4716]: I1207 16:02:50.607954 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:50Z","lastTransitionTime":"2025-12-07T16:02:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:50 crc kubenswrapper[4716]: I1207 16:02:50.656859 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 16:02:50 crc kubenswrapper[4716]: E1207 16:02:50.656982 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 16:02:50 crc kubenswrapper[4716]: I1207 16:02:50.711446 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:50 crc kubenswrapper[4716]: I1207 16:02:50.711526 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:50 crc kubenswrapper[4716]: I1207 16:02:50.711547 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:50 crc kubenswrapper[4716]: I1207 16:02:50.711579 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:50 crc kubenswrapper[4716]: I1207 16:02:50.711599 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:50Z","lastTransitionTime":"2025-12-07T16:02:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:50 crc kubenswrapper[4716]: I1207 16:02:50.814923 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:50 crc kubenswrapper[4716]: I1207 16:02:50.814969 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:50 crc kubenswrapper[4716]: I1207 16:02:50.814978 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:50 crc kubenswrapper[4716]: I1207 16:02:50.814995 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:50 crc kubenswrapper[4716]: I1207 16:02:50.815006 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:50Z","lastTransitionTime":"2025-12-07T16:02:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:50 crc kubenswrapper[4716]: I1207 16:02:50.917485 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:50 crc kubenswrapper[4716]: I1207 16:02:50.917546 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:50 crc kubenswrapper[4716]: I1207 16:02:50.917565 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:50 crc kubenswrapper[4716]: I1207 16:02:50.917601 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:50 crc kubenswrapper[4716]: I1207 16:02:50.917617 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:50Z","lastTransitionTime":"2025-12-07T16:02:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:50 crc kubenswrapper[4716]: I1207 16:02:50.920507 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qvc7p_c54d618a-dfec-4f2a-a288-f5188c8a128a/ovnkube-controller/0.log" Dec 07 16:02:50 crc kubenswrapper[4716]: I1207 16:02:50.923308 4716 generic.go:334] "Generic (PLEG): container finished" podID="c54d618a-dfec-4f2a-a288-f5188c8a128a" containerID="53c40176ba965341b13456a99ba4fad80dc6115939e3c4c884e74f4e026a7535" exitCode=1 Dec 07 16:02:50 crc kubenswrapper[4716]: I1207 16:02:50.923345 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" event={"ID":"c54d618a-dfec-4f2a-a288-f5188c8a128a","Type":"ContainerDied","Data":"53c40176ba965341b13456a99ba4fad80dc6115939e3c4c884e74f4e026a7535"} Dec 07 16:02:50 crc kubenswrapper[4716]: I1207 16:02:50.923952 4716 scope.go:117] "RemoveContainer" containerID="53c40176ba965341b13456a99ba4fad80dc6115939e3c4c884e74f4e026a7535" Dec 07 16:02:50 crc kubenswrapper[4716]: I1207 16:02:50.936875 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2jjld" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dee6f5c-861c-4819-9309-7791574c7b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51c1d4159ab11087839a2c2d192566fb9322447c26df560bf09517c906ececb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmdns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2jjld\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:50Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:50 crc kubenswrapper[4716]: I1207 16:02:50.950937 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4wr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0baf8575-26cf-4a72-8a6d-ae94c078eaad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0ca2b3cd5a11033afd219b0125be7ab5acd750935d7eb89a470eff3d93a0838\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p2trd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4wr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:50Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:50 crc kubenswrapper[4716]: I1207 16:02:50.960510 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-m5jl5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"824ee59f-1474-4994-a70e-462878521514\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a15fe50cc75515704779646cc19aa50004f0db9aa26fe2535110df749bbfb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhwzn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-m5jl5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:50Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:50 crc kubenswrapper[4716]: I1207 16:02:50.978058 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdee9957-48b9-458f-ba71-e0cd39f80596\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73f85248b4d68a9beecdacb3a01f433e97a0ffea2247ce92a41ef96855509377\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6276df652efe6ca935e745f24ef20f87a16465475139230c63b9999234a4d12a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3b1dd202586ee755ca89fbe12b30bbb12f95a62514f6e4d93cbbdbe55181e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd318647816e4a27849666b0ec4fc2ca4905295661c557dc381e1144f22cf284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d31de0326a05216fa10442bdb47151c6e073c97e9a16c34e793028e5e253756\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"le observer\\\\nW1207 16:02:34.645464 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1207 16:02:34.645587 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 16:02:34.646501 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2082283479/tls.crt::/tmp/serving-cert-2082283479/tls.key\\\\\\\"\\\\nI1207 16:02:34.972041 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 16:02:34.973702 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 16:02:34.973717 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 16:02:34.973738 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 16:02:34.973743 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 16:02:34.977185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1207 16:02:34.977204 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1207 16:02:34.977209 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977217 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 16:02:34.977225 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 16:02:34.977228 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 16:02:34.977232 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1207 16:02:34.979033 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27bd4a54f920b2725987c1c7fd2d9fdf16cd08738c246c333a4e9164a3a582a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:50Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:50 crc kubenswrapper[4716]: I1207 16:02:50.990727 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:50Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:51 crc kubenswrapper[4716]: I1207 16:02:51.002054 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:51Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:51 crc kubenswrapper[4716]: I1207 16:02:51.019842 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f95ef61779507288a03bda8eddf745e105f705998a117a30789675bfa452d88f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b955b23163e7c8c458a3bbe2f681633e6a663ce76748febcc0b3a3c37f3ca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:51Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:51 crc kubenswrapper[4716]: I1207 16:02:51.021031 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:51 crc kubenswrapper[4716]: I1207 16:02:51.021064 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:51 crc kubenswrapper[4716]: I1207 16:02:51.021087 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:51 crc kubenswrapper[4716]: I1207 16:02:51.021106 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:51 crc kubenswrapper[4716]: I1207 16:02:51.021118 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:51Z","lastTransitionTime":"2025-12-07T16:02:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:51 crc kubenswrapper[4716]: I1207 16:02:51.035831 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef3335ae3875a75ab64685fc7e3590177280d651e8f8292c6f5ea6f472e55789\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:51Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:51 crc kubenswrapper[4716]: I1207 16:02:51.048798 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4742ea84b3a2c9461d9b900beee7354a2ded24c39ad23e007cbf270ae12377c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:51Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:51 crc kubenswrapper[4716]: I1207 16:02:51.059486 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:51Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:51 crc kubenswrapper[4716]: I1207 16:02:51.074313 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c15b59eb-565d-4556-a4ce-75afdf159dc8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://530a2935826940767ea33fb4f37fa11f7dee4950d867fbcd6af72836d80f7db3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jks8z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ccaf6c1e6a978eb978a1c207b6c4c00b23737f78b743f966579aca781e22e0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jks8z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zcxxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:51Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:51 crc kubenswrapper[4716]: I1207 16:02:51.094176 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8ps9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98015cea-d753-4e7a-b2b5-5ae7733a81e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d68509b673eecd37725cdcc48e3d0612c3b9c7601433f2b72c21d662140daa59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f84f52a073437543721010dc440339654ffb5e04eeb6cc7afbcc62f00f64c8e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f84f52a073437543721010dc440339654ffb5e04eeb6cc7afbcc62f00f64c8e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f6fe1cb25218cb3ac20488dc47efa4dc6a26a43dbc8306b5ee38cbe1a186c67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f6fe1cb25218cb3ac20488dc47efa4dc6a26a43dbc8306b5ee38cbe1a186c67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14b748a8247f6863e8be61e90057a06626dfd3cd4f01bd3a11629f700f56378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f14b748a8247f6863e8be61e90057a06626dfd3cd4f01bd3a11629f700f56378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ec379c0d573cb9b5ffcfc5ca7c653ba60e2fbdd7e3d37836102e26618ce43e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ec379c0d573cb9b5ffcfc5ca7c653ba60e2fbdd7e3d37836102e26618ce43e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c75bf2a6703cf99865d4ed5fa000e1e3b1330d4946176cf585fec1a2f226df0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c75bf2a6703cf99865d4ed5fa000e1e3b1330d4946176cf585fec1a2f226df0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de45812c9a5a5bfe6b0146828d03fab28362b03d04ef861c150b9ffbf181d345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de45812c9a5a5bfe6b0146828d03fab28362b03d04ef861c150b9ffbf181d345\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8ps9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:51Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:51 crc kubenswrapper[4716]: I1207 16:02:51.123155 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:51 crc kubenswrapper[4716]: I1207 16:02:51.123199 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:51 crc kubenswrapper[4716]: I1207 16:02:51.123208 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:51 crc kubenswrapper[4716]: I1207 16:02:51.123227 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:51 crc kubenswrapper[4716]: I1207 16:02:51.123239 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:51Z","lastTransitionTime":"2025-12-07T16:02:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:51 crc kubenswrapper[4716]: I1207 16:02:51.130347 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7961cc9a-734e-42c2-b1ff-7075f3f20091\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84a36b76f20618c7885d611ecad9db29bac5101c27572be31d3acbb90c581291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0a0eca3c2087a457c0e448e649345ec4595a9c8a019aed2cddc5f30c77871e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://404426a171b8676eddd4a00996c76afab7e269643fa9608f1e1b6bdba385071d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fd3c2941782921f99a6cdc3de4e892234548bde527fec15a8ac92e513566e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e23abfdca314ed43d594333c9f9e401cc8ef7dad39b8ebac84a836b06ee5c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09b38cd4b389ea477c802a842ee4f2297a64d64684624117163e819539c257c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09b38cd4b389ea477c802a842ee4f2297a64d64684624117163e819539c257c6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c00f7a8e024e0e0dfd7b73e16d1d5437602e79ea95b73c21f24b209edef60e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c00f7a8e024e0e0dfd7b73e16d1d5437602e79ea95b73c21f24b209edef60e98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3f6392d426fa2cd788ec3a5c3bff4ac39483a54dfb2624a1c4370522413a350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3f6392d426fa2cd788ec3a5c3bff4ac39483a54dfb2624a1c4370522413a350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:51Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:51 crc kubenswrapper[4716]: I1207 16:02:51.148001 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3067de7c-c11c-4be1-bd94-d6572c266795\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d14d018fa044e9e80fe81e7e38df935992d14c773ed2b2193fd5cb873e5d0947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a9e9cf048220ca14c996847e9ced9f2367b9331dc6c2836dd023658855cd79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6d3c911874095b3c7cb062d03070e5090a7bb2456cb20cb4ce078f502515cda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ac6ce7701f1c6aa1ea8fcf7a3035e3b0eaa28ae3cb6afc442372bd9117bdde4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:51Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:51 crc kubenswrapper[4716]: I1207 16:02:51.171933 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c54d618a-dfec-4f2a-a288-f5188c8a128a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60324dc7d701bc064e14ae4f474bce10f6ee4ff3c3c942eab749e7166082cf9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ed4f4d3564674bf5cb4e7a76dc0fd95e203ecc8950798e59a1fb6c9ab4a20f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://def9ac840fb8ecf2778bed88ea980eddca7c5b2aba4bd1afb967d62fec3f7bae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b73be664f5412c98b89d971a2cac37c494151c293fdb44a98b390978658a6643\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a847d04e22e98091c0037a89054b9236228b701b8c7ba31940a56df0f37b95c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcae78a6133d97f6a32dd7421ccec4333b970a237404e056a5cc601f8eefdc8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://53c40176ba965341b13456a99ba4fad80dc6115939e3c4c884e74f4e026a7535\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53c40176ba965341b13456a99ba4fad80dc6115939e3c4c884e74f4e026a7535\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-07T16:02:50Z\\\",\\\"message\\\":\\\"rvice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1207 16:02:49.273594 6061 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1207 16:02:49.273647 6061 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1207 16:02:49.273655 6061 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1207 16:02:49.273699 6061 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1207 16:02:49.273719 6061 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1207 16:02:49.273725 6061 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1207 16:02:49.273749 6061 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1207 16:02:49.273770 6061 factory.go:656] Stopping watch factory\\\\nI1207 16:02:49.273826 6061 ovnkube.go:599] Stopped ovnkube\\\\nI1207 16:02:49.273858 6061 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1207 16:02:49.273867 6061 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1207 16:02:49.273875 6061 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1207 16:02:49.273882 6061 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1207 16:02:49.273890 6061 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1207 16:02:49.273911 6061 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1207 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcfdffe822635b0ae13b253e6133449fe00461b2792be43d913c4a77219b8255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9bfe2ffce7a7b7c320474117734fc05246cc2b4fd744833e896e8e7c83f5e4d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9bfe2ffce7a7b7c320474117734fc05246cc2b4fd744833e896e8e7c83f5e4d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qvc7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:51Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:51 crc kubenswrapper[4716]: I1207 16:02:51.226361 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:51 crc kubenswrapper[4716]: I1207 16:02:51.226421 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:51 crc kubenswrapper[4716]: I1207 16:02:51.226444 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:51 crc kubenswrapper[4716]: I1207 16:02:51.226474 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:51 crc kubenswrapper[4716]: I1207 16:02:51.226494 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:51Z","lastTransitionTime":"2025-12-07T16:02:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:51 crc kubenswrapper[4716]: I1207 16:02:51.328926 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:51 crc kubenswrapper[4716]: I1207 16:02:51.328977 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:51 crc kubenswrapper[4716]: I1207 16:02:51.328996 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:51 crc kubenswrapper[4716]: I1207 16:02:51.329020 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:51 crc kubenswrapper[4716]: I1207 16:02:51.329054 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:51Z","lastTransitionTime":"2025-12-07T16:02:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:51 crc kubenswrapper[4716]: I1207 16:02:51.338686 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 16:02:51 crc kubenswrapper[4716]: I1207 16:02:51.338809 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 16:02:51 crc kubenswrapper[4716]: E1207 16:02:51.339015 4716 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 07 16:02:51 crc kubenswrapper[4716]: E1207 16:02:51.339220 4716 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 07 16:02:51 crc kubenswrapper[4716]: E1207 16:02:51.339249 4716 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 07 16:02:51 crc kubenswrapper[4716]: E1207 16:02:51.339273 4716 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 07 16:02:51 crc kubenswrapper[4716]: E1207 16:02:51.339358 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-07 16:03:07.339326731 +0000 UTC m=+50.029611673 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 07 16:02:51 crc kubenswrapper[4716]: E1207 16:02:51.339850 4716 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 07 16:02:51 crc kubenswrapper[4716]: E1207 16:02:51.339878 4716 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 07 16:02:51 crc kubenswrapper[4716]: E1207 16:02:51.339939 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-07 16:03:07.339918757 +0000 UTC m=+50.030203709 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 07 16:02:51 crc kubenswrapper[4716]: I1207 16:02:51.431406 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:51 crc kubenswrapper[4716]: I1207 16:02:51.431464 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:51 crc kubenswrapper[4716]: I1207 16:02:51.431481 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:51 crc kubenswrapper[4716]: I1207 16:02:51.431503 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:51 crc kubenswrapper[4716]: I1207 16:02:51.431521 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:51Z","lastTransitionTime":"2025-12-07T16:02:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:51 crc kubenswrapper[4716]: I1207 16:02:51.439983 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 16:02:51 crc kubenswrapper[4716]: E1207 16:02:51.440142 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 16:03:07.440124766 +0000 UTC m=+50.130409678 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:02:51 crc kubenswrapper[4716]: I1207 16:02:51.440181 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:02:51 crc kubenswrapper[4716]: E1207 16:02:51.440289 4716 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 07 16:02:51 crc kubenswrapper[4716]: I1207 16:02:51.440301 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:02:51 crc kubenswrapper[4716]: E1207 16:02:51.440322 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-07 16:03:07.440315331 +0000 UTC m=+50.130600243 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 07 16:02:51 crc kubenswrapper[4716]: E1207 16:02:51.440459 4716 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 07 16:02:51 crc kubenswrapper[4716]: E1207 16:02:51.440525 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-07 16:03:07.440508236 +0000 UTC m=+50.130793188 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 07 16:02:51 crc kubenswrapper[4716]: I1207 16:02:51.533624 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:51 crc kubenswrapper[4716]: I1207 16:02:51.533660 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:51 crc kubenswrapper[4716]: I1207 16:02:51.533670 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:51 crc kubenswrapper[4716]: I1207 16:02:51.533685 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:51 crc kubenswrapper[4716]: I1207 16:02:51.533696 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:51Z","lastTransitionTime":"2025-12-07T16:02:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:51 crc kubenswrapper[4716]: I1207 16:02:51.636457 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:51 crc kubenswrapper[4716]: I1207 16:02:51.636496 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:51 crc kubenswrapper[4716]: I1207 16:02:51.636507 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:51 crc kubenswrapper[4716]: I1207 16:02:51.636523 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:51 crc kubenswrapper[4716]: I1207 16:02:51.636532 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:51Z","lastTransitionTime":"2025-12-07T16:02:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:51 crc kubenswrapper[4716]: I1207 16:02:51.657017 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 16:02:51 crc kubenswrapper[4716]: I1207 16:02:51.657075 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:02:51 crc kubenswrapper[4716]: E1207 16:02:51.657124 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 16:02:51 crc kubenswrapper[4716]: E1207 16:02:51.657337 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 16:02:51 crc kubenswrapper[4716]: I1207 16:02:51.738932 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:51 crc kubenswrapper[4716]: I1207 16:02:51.738980 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:51 crc kubenswrapper[4716]: I1207 16:02:51.738992 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:51 crc kubenswrapper[4716]: I1207 16:02:51.739009 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:51 crc kubenswrapper[4716]: I1207 16:02:51.739023 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:51Z","lastTransitionTime":"2025-12-07T16:02:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:51 crc kubenswrapper[4716]: I1207 16:02:51.841350 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:51 crc kubenswrapper[4716]: I1207 16:02:51.841394 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:51 crc kubenswrapper[4716]: I1207 16:02:51.841408 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:51 crc kubenswrapper[4716]: I1207 16:02:51.841426 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:51 crc kubenswrapper[4716]: I1207 16:02:51.841438 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:51Z","lastTransitionTime":"2025-12-07T16:02:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:51 crc kubenswrapper[4716]: I1207 16:02:51.929384 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qvc7p_c54d618a-dfec-4f2a-a288-f5188c8a128a/ovnkube-controller/0.log" Dec 07 16:02:51 crc kubenswrapper[4716]: I1207 16:02:51.932124 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" event={"ID":"c54d618a-dfec-4f2a-a288-f5188c8a128a","Type":"ContainerStarted","Data":"a8baee0f3d1e686bc2c17b0777f4eacc31c7f1fe44720e68f94612e6c806bc8b"} Dec 07 16:02:51 crc kubenswrapper[4716]: I1207 16:02:51.932674 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" Dec 07 16:02:51 crc kubenswrapper[4716]: I1207 16:02:51.943771 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:51 crc kubenswrapper[4716]: I1207 16:02:51.943880 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:51 crc kubenswrapper[4716]: I1207 16:02:51.943910 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:51 crc kubenswrapper[4716]: I1207 16:02:51.943939 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:51 crc kubenswrapper[4716]: I1207 16:02:51.943960 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:51Z","lastTransitionTime":"2025-12-07T16:02:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:51 crc kubenswrapper[4716]: I1207 16:02:51.950768 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef3335ae3875a75ab64685fc7e3590177280d651e8f8292c6f5ea6f472e55789\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:51Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:51 crc kubenswrapper[4716]: I1207 16:02:51.963903 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:51Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:51 crc kubenswrapper[4716]: I1207 16:02:51.976155 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f95ef61779507288a03bda8eddf745e105f705998a117a30789675bfa452d88f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b955b23163e7c8c458a3bbe2f681633e6a663ce76748febcc0b3a3c37f3ca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:51Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:51 crc kubenswrapper[4716]: I1207 16:02:51.988175 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c15b59eb-565d-4556-a4ce-75afdf159dc8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://530a2935826940767ea33fb4f37fa11f7dee4950d867fbcd6af72836d80f7db3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jks8z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ccaf6c1e6a978eb978a1c207b6c4c00b23737f78b743f966579aca781e22e0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jks8z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zcxxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:51Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:52 crc kubenswrapper[4716]: I1207 16:02:52.005779 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8ps9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98015cea-d753-4e7a-b2b5-5ae7733a81e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d68509b673eecd37725cdcc48e3d0612c3b9c7601433f2b72c21d662140daa59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f84f52a073437543721010dc440339654ffb5e04eeb6cc7afbcc62f00f64c8e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f84f52a073437543721010dc440339654ffb5e04eeb6cc7afbcc62f00f64c8e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f6fe1cb25218cb3ac20488dc47efa4dc6a26a43dbc8306b5ee38cbe1a186c67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f6fe1cb25218cb3ac20488dc47efa4dc6a26a43dbc8306b5ee38cbe1a186c67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14b748a8247f6863e8be61e90057a06626dfd3cd4f01bd3a11629f700f56378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f14b748a8247f6863e8be61e90057a06626dfd3cd4f01bd3a11629f700f56378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ec379c0d573cb9b5ffcfc5ca7c653ba60e2fbdd7e3d37836102e26618ce43e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ec379c0d573cb9b5ffcfc5ca7c653ba60e2fbdd7e3d37836102e26618ce43e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c75bf2a6703cf99865d4ed5fa000e1e3b1330d4946176cf585fec1a2f226df0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c75bf2a6703cf99865d4ed5fa000e1e3b1330d4946176cf585fec1a2f226df0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de45812c9a5a5bfe6b0146828d03fab28362b03d04ef861c150b9ffbf181d345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de45812c9a5a5bfe6b0146828d03fab28362b03d04ef861c150b9ffbf181d345\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8ps9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:52Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:52 crc kubenswrapper[4716]: I1207 16:02:52.022183 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7961cc9a-734e-42c2-b1ff-7075f3f20091\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84a36b76f20618c7885d611ecad9db29bac5101c27572be31d3acbb90c581291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0a0eca3c2087a457c0e448e649345ec4595a9c8a019aed2cddc5f30c77871e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://404426a171b8676eddd4a00996c76afab7e269643fa9608f1e1b6bdba385071d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fd3c2941782921f99a6cdc3de4e892234548bde527fec15a8ac92e513566e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e23abfdca314ed43d594333c9f9e401cc8ef7dad39b8ebac84a836b06ee5c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09b38cd4b389ea477c802a842ee4f2297a64d64684624117163e819539c257c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09b38cd4b389ea477c802a842ee4f2297a64d64684624117163e819539c257c6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c00f7a8e024e0e0dfd7b73e16d1d5437602e79ea95b73c21f24b209edef60e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c00f7a8e024e0e0dfd7b73e16d1d5437602e79ea95b73c21f24b209edef60e98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3f6392d426fa2cd788ec3a5c3bff4ac39483a54dfb2624a1c4370522413a350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3f6392d426fa2cd788ec3a5c3bff4ac39483a54dfb2624a1c4370522413a350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:52Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:52 crc kubenswrapper[4716]: I1207 16:02:52.032301 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3067de7c-c11c-4be1-bd94-d6572c266795\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d14d018fa044e9e80fe81e7e38df935992d14c773ed2b2193fd5cb873e5d0947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a9e9cf048220ca14c996847e9ced9f2367b9331dc6c2836dd023658855cd79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6d3c911874095b3c7cb062d03070e5090a7bb2456cb20cb4ce078f502515cda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ac6ce7701f1c6aa1ea8fcf7a3035e3b0eaa28ae3cb6afc442372bd9117bdde4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:52Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:52 crc kubenswrapper[4716]: I1207 16:02:52.041680 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4742ea84b3a2c9461d9b900beee7354a2ded24c39ad23e007cbf270ae12377c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:52Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:52 crc kubenswrapper[4716]: I1207 16:02:52.045987 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:52 crc kubenswrapper[4716]: I1207 16:02:52.046115 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:52 crc kubenswrapper[4716]: I1207 16:02:52.046180 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:52 crc kubenswrapper[4716]: I1207 16:02:52.046252 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:52 crc kubenswrapper[4716]: I1207 16:02:52.046317 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:52Z","lastTransitionTime":"2025-12-07T16:02:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:52 crc kubenswrapper[4716]: I1207 16:02:52.053273 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:52Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:52 crc kubenswrapper[4716]: I1207 16:02:52.075796 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c54d618a-dfec-4f2a-a288-f5188c8a128a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60324dc7d701bc064e14ae4f474bce10f6ee4ff3c3c942eab749e7166082cf9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ed4f4d3564674bf5cb4e7a76dc0fd95e203ecc8950798e59a1fb6c9ab4a20f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://def9ac840fb8ecf2778bed88ea980eddca7c5b2aba4bd1afb967d62fec3f7bae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b73be664f5412c98b89d971a2cac37c494151c293fdb44a98b390978658a6643\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a847d04e22e98091c0037a89054b9236228b701b8c7ba31940a56df0f37b95c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcae78a6133d97f6a32dd7421ccec4333b970a237404e056a5cc601f8eefdc8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8baee0f3d1e686bc2c17b0777f4eacc31c7f1fe44720e68f94612e6c806bc8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53c40176ba965341b13456a99ba4fad80dc6115939e3c4c884e74f4e026a7535\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-07T16:02:50Z\\\",\\\"message\\\":\\\"rvice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1207 16:02:49.273594 6061 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1207 16:02:49.273647 6061 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1207 16:02:49.273655 6061 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1207 16:02:49.273699 6061 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1207 16:02:49.273719 6061 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1207 16:02:49.273725 6061 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1207 16:02:49.273749 6061 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1207 16:02:49.273770 6061 factory.go:656] Stopping watch factory\\\\nI1207 16:02:49.273826 6061 ovnkube.go:599] Stopped ovnkube\\\\nI1207 16:02:49.273858 6061 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1207 16:02:49.273867 6061 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1207 16:02:49.273875 6061 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1207 16:02:49.273882 6061 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1207 16:02:49.273890 6061 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1207 16:02:49.273911 6061 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1207 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:47Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcfdffe822635b0ae13b253e6133449fe00461b2792be43d913c4a77219b8255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9bfe2ffce7a7b7c320474117734fc05246cc2b4fd744833e896e8e7c83f5e4d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9bfe2ffce7a7b7c320474117734fc05246cc2b4fd744833e896e8e7c83f5e4d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qvc7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:52Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:52 crc kubenswrapper[4716]: I1207 16:02:52.085640 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-m5jl5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"824ee59f-1474-4994-a70e-462878521514\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a15fe50cc75515704779646cc19aa50004f0db9aa26fe2535110df749bbfb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhwzn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-m5jl5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:52Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:52 crc kubenswrapper[4716]: I1207 16:02:52.101949 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdee9957-48b9-458f-ba71-e0cd39f80596\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73f85248b4d68a9beecdacb3a01f433e97a0ffea2247ce92a41ef96855509377\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6276df652efe6ca935e745f24ef20f87a16465475139230c63b9999234a4d12a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3b1dd202586ee755ca89fbe12b30bbb12f95a62514f6e4d93cbbdbe55181e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd318647816e4a27849666b0ec4fc2ca4905295661c557dc381e1144f22cf284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d31de0326a05216fa10442bdb47151c6e073c97e9a16c34e793028e5e253756\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"le observer\\\\nW1207 16:02:34.645464 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1207 16:02:34.645587 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 16:02:34.646501 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2082283479/tls.crt::/tmp/serving-cert-2082283479/tls.key\\\\\\\"\\\\nI1207 16:02:34.972041 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 16:02:34.973702 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 16:02:34.973717 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 16:02:34.973738 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 16:02:34.973743 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 16:02:34.977185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1207 16:02:34.977204 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1207 16:02:34.977209 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977217 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 16:02:34.977225 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 16:02:34.977228 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 16:02:34.977232 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1207 16:02:34.979033 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27bd4a54f920b2725987c1c7fd2d9fdf16cd08738c246c333a4e9164a3a582a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:52Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:52 crc kubenswrapper[4716]: I1207 16:02:52.114951 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:52Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:52 crc kubenswrapper[4716]: I1207 16:02:52.124451 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2jjld" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dee6f5c-861c-4819-9309-7791574c7b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51c1d4159ab11087839a2c2d192566fb9322447c26df560bf09517c906ececb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmdns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2jjld\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:52Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:52 crc kubenswrapper[4716]: I1207 16:02:52.136886 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4wr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0baf8575-26cf-4a72-8a6d-ae94c078eaad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0ca2b3cd5a11033afd219b0125be7ab5acd750935d7eb89a470eff3d93a0838\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p2trd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4wr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:52Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:52 crc kubenswrapper[4716]: I1207 16:02:52.148839 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:52 crc kubenswrapper[4716]: I1207 16:02:52.148899 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:52 crc kubenswrapper[4716]: I1207 16:02:52.148916 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:52 crc kubenswrapper[4716]: I1207 16:02:52.148945 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:52 crc kubenswrapper[4716]: I1207 16:02:52.148958 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:52Z","lastTransitionTime":"2025-12-07T16:02:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:52 crc kubenswrapper[4716]: I1207 16:02:52.250916 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:52 crc kubenswrapper[4716]: I1207 16:02:52.250954 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:52 crc kubenswrapper[4716]: I1207 16:02:52.250965 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:52 crc kubenswrapper[4716]: I1207 16:02:52.250980 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:52 crc kubenswrapper[4716]: I1207 16:02:52.250992 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:52Z","lastTransitionTime":"2025-12-07T16:02:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:52 crc kubenswrapper[4716]: I1207 16:02:52.354112 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:52 crc kubenswrapper[4716]: I1207 16:02:52.354437 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:52 crc kubenswrapper[4716]: I1207 16:02:52.354567 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:52 crc kubenswrapper[4716]: I1207 16:02:52.354700 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:52 crc kubenswrapper[4716]: I1207 16:02:52.354819 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:52Z","lastTransitionTime":"2025-12-07T16:02:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:52 crc kubenswrapper[4716]: I1207 16:02:52.457515 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:52 crc kubenswrapper[4716]: I1207 16:02:52.457588 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:52 crc kubenswrapper[4716]: I1207 16:02:52.457606 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:52 crc kubenswrapper[4716]: I1207 16:02:52.457629 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:52 crc kubenswrapper[4716]: I1207 16:02:52.457650 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:52Z","lastTransitionTime":"2025-12-07T16:02:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:52 crc kubenswrapper[4716]: I1207 16:02:52.559583 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:52 crc kubenswrapper[4716]: I1207 16:02:52.559644 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:52 crc kubenswrapper[4716]: I1207 16:02:52.559668 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:52 crc kubenswrapper[4716]: I1207 16:02:52.559696 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:52 crc kubenswrapper[4716]: I1207 16:02:52.559718 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:52Z","lastTransitionTime":"2025-12-07T16:02:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:52 crc kubenswrapper[4716]: I1207 16:02:52.656960 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 16:02:52 crc kubenswrapper[4716]: E1207 16:02:52.657308 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 16:02:52 crc kubenswrapper[4716]: I1207 16:02:52.662375 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:52 crc kubenswrapper[4716]: I1207 16:02:52.662415 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:52 crc kubenswrapper[4716]: I1207 16:02:52.662426 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:52 crc kubenswrapper[4716]: I1207 16:02:52.662439 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:52 crc kubenswrapper[4716]: I1207 16:02:52.662448 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:52Z","lastTransitionTime":"2025-12-07T16:02:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:52 crc kubenswrapper[4716]: I1207 16:02:52.765411 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:52 crc kubenswrapper[4716]: I1207 16:02:52.765463 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:52 crc kubenswrapper[4716]: I1207 16:02:52.765478 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:52 crc kubenswrapper[4716]: I1207 16:02:52.765494 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:52 crc kubenswrapper[4716]: I1207 16:02:52.765506 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:52Z","lastTransitionTime":"2025-12-07T16:02:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:52 crc kubenswrapper[4716]: I1207 16:02:52.868170 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:52 crc kubenswrapper[4716]: I1207 16:02:52.868242 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:52 crc kubenswrapper[4716]: I1207 16:02:52.868257 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:52 crc kubenswrapper[4716]: I1207 16:02:52.868293 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:52 crc kubenswrapper[4716]: I1207 16:02:52.868307 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:52Z","lastTransitionTime":"2025-12-07T16:02:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:52 crc kubenswrapper[4716]: I1207 16:02:52.939104 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qvc7p_c54d618a-dfec-4f2a-a288-f5188c8a128a/ovnkube-controller/1.log" Dec 07 16:02:52 crc kubenswrapper[4716]: I1207 16:02:52.939750 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qvc7p_c54d618a-dfec-4f2a-a288-f5188c8a128a/ovnkube-controller/0.log" Dec 07 16:02:52 crc kubenswrapper[4716]: I1207 16:02:52.944129 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" event={"ID":"c54d618a-dfec-4f2a-a288-f5188c8a128a","Type":"ContainerDied","Data":"a8baee0f3d1e686bc2c17b0777f4eacc31c7f1fe44720e68f94612e6c806bc8b"} Dec 07 16:02:52 crc kubenswrapper[4716]: I1207 16:02:52.944259 4716 scope.go:117] "RemoveContainer" containerID="53c40176ba965341b13456a99ba4fad80dc6115939e3c4c884e74f4e026a7535" Dec 07 16:02:52 crc kubenswrapper[4716]: I1207 16:02:52.944013 4716 generic.go:334] "Generic (PLEG): container finished" podID="c54d618a-dfec-4f2a-a288-f5188c8a128a" containerID="a8baee0f3d1e686bc2c17b0777f4eacc31c7f1fe44720e68f94612e6c806bc8b" exitCode=1 Dec 07 16:02:52 crc kubenswrapper[4716]: I1207 16:02:52.945910 4716 scope.go:117] "RemoveContainer" containerID="a8baee0f3d1e686bc2c17b0777f4eacc31c7f1fe44720e68f94612e6c806bc8b" Dec 07 16:02:52 crc kubenswrapper[4716]: E1207 16:02:52.946319 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-qvc7p_openshift-ovn-kubernetes(c54d618a-dfec-4f2a-a288-f5188c8a128a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" podUID="c54d618a-dfec-4f2a-a288-f5188c8a128a" Dec 07 16:02:52 crc kubenswrapper[4716]: I1207 16:02:52.968346 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef3335ae3875a75ab64685fc7e3590177280d651e8f8292c6f5ea6f472e55789\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:52Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:52 crc kubenswrapper[4716]: I1207 16:02:52.971621 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:52 crc kubenswrapper[4716]: I1207 16:02:52.971683 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:52 crc kubenswrapper[4716]: I1207 16:02:52.971702 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:52 crc kubenswrapper[4716]: I1207 16:02:52.971726 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:52 crc kubenswrapper[4716]: I1207 16:02:52.971744 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:52Z","lastTransitionTime":"2025-12-07T16:02:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:52 crc kubenswrapper[4716]: I1207 16:02:52.987517 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:52Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.008902 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f95ef61779507288a03bda8eddf745e105f705998a117a30789675bfa452d88f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b955b23163e7c8c458a3bbe2f681633e6a663ce76748febcc0b3a3c37f3ca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:53Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.030625 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8ps9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98015cea-d753-4e7a-b2b5-5ae7733a81e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d68509b673eecd37725cdcc48e3d0612c3b9c7601433f2b72c21d662140daa59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f84f52a073437543721010dc440339654ffb5e04eeb6cc7afbcc62f00f64c8e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f84f52a073437543721010dc440339654ffb5e04eeb6cc7afbcc62f00f64c8e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f6fe1cb25218cb3ac20488dc47efa4dc6a26a43dbc8306b5ee38cbe1a186c67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f6fe1cb25218cb3ac20488dc47efa4dc6a26a43dbc8306b5ee38cbe1a186c67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14b748a8247f6863e8be61e90057a06626dfd3cd4f01bd3a11629f700f56378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f14b748a8247f6863e8be61e90057a06626dfd3cd4f01bd3a11629f700f56378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ec379c0d573cb9b5ffcfc5ca7c653ba60e2fbdd7e3d37836102e26618ce43e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ec379c0d573cb9b5ffcfc5ca7c653ba60e2fbdd7e3d37836102e26618ce43e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c75bf2a6703cf99865d4ed5fa000e1e3b1330d4946176cf585fec1a2f226df0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c75bf2a6703cf99865d4ed5fa000e1e3b1330d4946176cf585fec1a2f226df0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de45812c9a5a5bfe6b0146828d03fab28362b03d04ef861c150b9ffbf181d345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de45812c9a5a5bfe6b0146828d03fab28362b03d04ef861c150b9ffbf181d345\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8ps9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:53Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.061152 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mx85f"] Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.061511 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mx85f" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.062559 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7961cc9a-734e-42c2-b1ff-7075f3f20091\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84a36b76f20618c7885d611ecad9db29bac5101c27572be31d3acbb90c581291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0a0eca3c2087a457c0e448e649345ec4595a9c8a019aed2cddc5f30c77871e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://404426a171b8676eddd4a00996c76afab7e269643fa9608f1e1b6bdba385071d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fd3c2941782921f99a6cdc3de4e892234548bde527fec15a8ac92e513566e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e23abfdca314ed43d594333c9f9e401cc8ef7dad39b8ebac84a836b06ee5c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09b38cd4b389ea477c802a842ee4f2297a64d64684624117163e819539c257c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09b38cd4b389ea477c802a842ee4f2297a64d64684624117163e819539c257c6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c00f7a8e024e0e0dfd7b73e16d1d5437602e79ea95b73c21f24b209edef60e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c00f7a8e024e0e0dfd7b73e16d1d5437602e79ea95b73c21f24b209edef60e98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3f6392d426fa2cd788ec3a5c3bff4ac39483a54dfb2624a1c4370522413a350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3f6392d426fa2cd788ec3a5c3bff4ac39483a54dfb2624a1c4370522413a350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:53Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.063186 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.063747 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.074776 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.074823 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.074839 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.074862 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.074879 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:53Z","lastTransitionTime":"2025-12-07T16:02:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.078149 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3067de7c-c11c-4be1-bd94-d6572c266795\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d14d018fa044e9e80fe81e7e38df935992d14c773ed2b2193fd5cb873e5d0947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a9e9cf048220ca14c996847e9ced9f2367b9331dc6c2836dd023658855cd79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6d3c911874095b3c7cb062d03070e5090a7bb2456cb20cb4ce078f502515cda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ac6ce7701f1c6aa1ea8fcf7a3035e3b0eaa28ae3cb6afc442372bd9117bdde4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:53Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.089967 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4742ea84b3a2c9461d9b900beee7354a2ded24c39ad23e007cbf270ae12377c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:53Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.102548 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:53Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.114272 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c15b59eb-565d-4556-a4ce-75afdf159dc8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://530a2935826940767ea33fb4f37fa11f7dee4950d867fbcd6af72836d80f7db3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jks8z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ccaf6c1e6a978eb978a1c207b6c4c00b23737f78b743f966579aca781e22e0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jks8z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zcxxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:53Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.131650 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c54d618a-dfec-4f2a-a288-f5188c8a128a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60324dc7d701bc064e14ae4f474bce10f6ee4ff3c3c942eab749e7166082cf9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ed4f4d3564674bf5cb4e7a76dc0fd95e203ecc8950798e59a1fb6c9ab4a20f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://def9ac840fb8ecf2778bed88ea980eddca7c5b2aba4bd1afb967d62fec3f7bae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b73be664f5412c98b89d971a2cac37c494151c293fdb44a98b390978658a6643\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a847d04e22e98091c0037a89054b9236228b701b8c7ba31940a56df0f37b95c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcae78a6133d97f6a32dd7421ccec4333b970a237404e056a5cc601f8eefdc8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8baee0f3d1e686bc2c17b0777f4eacc31c7f1fe44720e68f94612e6c806bc8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53c40176ba965341b13456a99ba4fad80dc6115939e3c4c884e74f4e026a7535\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-07T16:02:50Z\\\",\\\"message\\\":\\\"rvice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1207 16:02:49.273594 6061 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1207 16:02:49.273647 6061 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1207 16:02:49.273655 6061 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1207 16:02:49.273699 6061 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1207 16:02:49.273719 6061 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1207 16:02:49.273725 6061 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1207 16:02:49.273749 6061 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1207 16:02:49.273770 6061 factory.go:656] Stopping watch factory\\\\nI1207 16:02:49.273826 6061 ovnkube.go:599] Stopped ovnkube\\\\nI1207 16:02:49.273858 6061 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1207 16:02:49.273867 6061 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1207 16:02:49.273875 6061 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1207 16:02:49.273882 6061 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1207 16:02:49.273890 6061 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1207 16:02:49.273911 6061 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1207 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:47Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8baee0f3d1e686bc2c17b0777f4eacc31c7f1fe44720e68f94612e6c806bc8b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-07T16:02:52Z\\\",\\\"message\\\":\\\"r occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:51Z is after 2025-08-24T17:21:41Z]\\\\nI1207 16:02:51.943802 6191 services_controller.go:444] Built service openshift-marketplace/redhat-marketplace LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1207 16:02:51.943764 6191 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-route-controller-manager/route-controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"18746a4d-8a63-458a-b7e3-8fb89ff95fc0\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-route-controller-manager/route-controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-route-controller-manager/route-controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcfdffe822635b0ae13b253e6133449fe00461b2792be43d913c4a77219b8255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9bfe2ffce7a7b7c320474117734fc05246cc2b4fd744833e896e8e7c83f5e4d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9bfe2ffce7a7b7c320474117734fc05246cc2b4fd744833e896e8e7c83f5e4d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qvc7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:53Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.146174 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdee9957-48b9-458f-ba71-e0cd39f80596\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73f85248b4d68a9beecdacb3a01f433e97a0ffea2247ce92a41ef96855509377\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6276df652efe6ca935e745f24ef20f87a16465475139230c63b9999234a4d12a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3b1dd202586ee755ca89fbe12b30bbb12f95a62514f6e4d93cbbdbe55181e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd318647816e4a27849666b0ec4fc2ca4905295661c557dc381e1144f22cf284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d31de0326a05216fa10442bdb47151c6e073c97e9a16c34e793028e5e253756\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"le observer\\\\nW1207 16:02:34.645464 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1207 16:02:34.645587 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 16:02:34.646501 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2082283479/tls.crt::/tmp/serving-cert-2082283479/tls.key\\\\\\\"\\\\nI1207 16:02:34.972041 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 16:02:34.973702 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 16:02:34.973717 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 16:02:34.973738 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 16:02:34.973743 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 16:02:34.977185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1207 16:02:34.977204 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1207 16:02:34.977209 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977217 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 16:02:34.977225 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 16:02:34.977228 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 16:02:34.977232 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1207 16:02:34.979033 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27bd4a54f920b2725987c1c7fd2d9fdf16cd08738c246c333a4e9164a3a582a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:53Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.157702 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gg5nb\" (UniqueName: \"kubernetes.io/projected/118dd9ba-437c-4e5c-b5ff-4239ae6d4a63-kube-api-access-gg5nb\") pod \"ovnkube-control-plane-749d76644c-mx85f\" (UID: \"118dd9ba-437c-4e5c-b5ff-4239ae6d4a63\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mx85f" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.158425 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/118dd9ba-437c-4e5c-b5ff-4239ae6d4a63-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-mx85f\" (UID: \"118dd9ba-437c-4e5c-b5ff-4239ae6d4a63\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mx85f" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.158583 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/118dd9ba-437c-4e5c-b5ff-4239ae6d4a63-env-overrides\") pod \"ovnkube-control-plane-749d76644c-mx85f\" (UID: \"118dd9ba-437c-4e5c-b5ff-4239ae6d4a63\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mx85f" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.158676 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/118dd9ba-437c-4e5c-b5ff-4239ae6d4a63-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-mx85f\" (UID: \"118dd9ba-437c-4e5c-b5ff-4239ae6d4a63\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mx85f" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.159056 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:53Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.167849 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2jjld" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dee6f5c-861c-4819-9309-7791574c7b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51c1d4159ab11087839a2c2d192566fb9322447c26df560bf09517c906ececb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmdns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2jjld\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:53Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.178999 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.179038 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.179051 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.179069 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.179100 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:53Z","lastTransitionTime":"2025-12-07T16:02:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.181521 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4wr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0baf8575-26cf-4a72-8a6d-ae94c078eaad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0ca2b3cd5a11033afd219b0125be7ab5acd750935d7eb89a470eff3d93a0838\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p2trd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4wr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:53Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.193169 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-m5jl5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"824ee59f-1474-4994-a70e-462878521514\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a15fe50cc75515704779646cc19aa50004f0db9aa26fe2535110df749bbfb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhwzn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-m5jl5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:53Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.204689 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4742ea84b3a2c9461d9b900beee7354a2ded24c39ad23e007cbf270ae12377c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:53Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.216273 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:53Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.227723 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c15b59eb-565d-4556-a4ce-75afdf159dc8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://530a2935826940767ea33fb4f37fa11f7dee4950d867fbcd6af72836d80f7db3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jks8z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ccaf6c1e6a978eb978a1c207b6c4c00b23737f78b743f966579aca781e22e0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jks8z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zcxxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:53Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.247414 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8ps9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98015cea-d753-4e7a-b2b5-5ae7733a81e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d68509b673eecd37725cdcc48e3d0612c3b9c7601433f2b72c21d662140daa59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f84f52a073437543721010dc440339654ffb5e04eeb6cc7afbcc62f00f64c8e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f84f52a073437543721010dc440339654ffb5e04eeb6cc7afbcc62f00f64c8e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f6fe1cb25218cb3ac20488dc47efa4dc6a26a43dbc8306b5ee38cbe1a186c67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f6fe1cb25218cb3ac20488dc47efa4dc6a26a43dbc8306b5ee38cbe1a186c67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14b748a8247f6863e8be61e90057a06626dfd3cd4f01bd3a11629f700f56378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f14b748a8247f6863e8be61e90057a06626dfd3cd4f01bd3a11629f700f56378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ec379c0d573cb9b5ffcfc5ca7c653ba60e2fbdd7e3d37836102e26618ce43e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ec379c0d573cb9b5ffcfc5ca7c653ba60e2fbdd7e3d37836102e26618ce43e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c75bf2a6703cf99865d4ed5fa000e1e3b1330d4946176cf585fec1a2f226df0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c75bf2a6703cf99865d4ed5fa000e1e3b1330d4946176cf585fec1a2f226df0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de45812c9a5a5bfe6b0146828d03fab28362b03d04ef861c150b9ffbf181d345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de45812c9a5a5bfe6b0146828d03fab28362b03d04ef861c150b9ffbf181d345\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8ps9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:53Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.259575 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/118dd9ba-437c-4e5c-b5ff-4239ae6d4a63-env-overrides\") pod \"ovnkube-control-plane-749d76644c-mx85f\" (UID: \"118dd9ba-437c-4e5c-b5ff-4239ae6d4a63\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mx85f" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.259646 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/118dd9ba-437c-4e5c-b5ff-4239ae6d4a63-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-mx85f\" (UID: \"118dd9ba-437c-4e5c-b5ff-4239ae6d4a63\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mx85f" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.259673 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gg5nb\" (UniqueName: \"kubernetes.io/projected/118dd9ba-437c-4e5c-b5ff-4239ae6d4a63-kube-api-access-gg5nb\") pod \"ovnkube-control-plane-749d76644c-mx85f\" (UID: \"118dd9ba-437c-4e5c-b5ff-4239ae6d4a63\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mx85f" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.259697 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/118dd9ba-437c-4e5c-b5ff-4239ae6d4a63-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-mx85f\" (UID: \"118dd9ba-437c-4e5c-b5ff-4239ae6d4a63\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mx85f" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.260137 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/118dd9ba-437c-4e5c-b5ff-4239ae6d4a63-env-overrides\") pod \"ovnkube-control-plane-749d76644c-mx85f\" (UID: \"118dd9ba-437c-4e5c-b5ff-4239ae6d4a63\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mx85f" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.260690 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/118dd9ba-437c-4e5c-b5ff-4239ae6d4a63-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-mx85f\" (UID: \"118dd9ba-437c-4e5c-b5ff-4239ae6d4a63\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mx85f" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.265407 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/118dd9ba-437c-4e5c-b5ff-4239ae6d4a63-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-mx85f\" (UID: \"118dd9ba-437c-4e5c-b5ff-4239ae6d4a63\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mx85f" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.266175 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7961cc9a-734e-42c2-b1ff-7075f3f20091\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84a36b76f20618c7885d611ecad9db29bac5101c27572be31d3acbb90c581291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0a0eca3c2087a457c0e448e649345ec4595a9c8a019aed2cddc5f30c77871e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://404426a171b8676eddd4a00996c76afab7e269643fa9608f1e1b6bdba385071d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fd3c2941782921f99a6cdc3de4e892234548bde527fec15a8ac92e513566e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e23abfdca314ed43d594333c9f9e401cc8ef7dad39b8ebac84a836b06ee5c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09b38cd4b389ea477c802a842ee4f2297a64d64684624117163e819539c257c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09b38cd4b389ea477c802a842ee4f2297a64d64684624117163e819539c257c6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c00f7a8e024e0e0dfd7b73e16d1d5437602e79ea95b73c21f24b209edef60e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c00f7a8e024e0e0dfd7b73e16d1d5437602e79ea95b73c21f24b209edef60e98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3f6392d426fa2cd788ec3a5c3bff4ac39483a54dfb2624a1c4370522413a350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3f6392d426fa2cd788ec3a5c3bff4ac39483a54dfb2624a1c4370522413a350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:53Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.275444 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gg5nb\" (UniqueName: \"kubernetes.io/projected/118dd9ba-437c-4e5c-b5ff-4239ae6d4a63-kube-api-access-gg5nb\") pod \"ovnkube-control-plane-749d76644c-mx85f\" (UID: \"118dd9ba-437c-4e5c-b5ff-4239ae6d4a63\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mx85f" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.282401 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.282464 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.282481 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.282503 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.282519 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:53Z","lastTransitionTime":"2025-12-07T16:02:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.283067 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3067de7c-c11c-4be1-bd94-d6572c266795\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d14d018fa044e9e80fe81e7e38df935992d14c773ed2b2193fd5cb873e5d0947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a9e9cf048220ca14c996847e9ced9f2367b9331dc6c2836dd023658855cd79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6d3c911874095b3c7cb062d03070e5090a7bb2456cb20cb4ce078f502515cda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ac6ce7701f1c6aa1ea8fcf7a3035e3b0eaa28ae3cb6afc442372bd9117bdde4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:53Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.293100 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mx85f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"118dd9ba-437c-4e5c-b5ff-4239ae6d4a63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gg5nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gg5nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mx85f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:53Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.311942 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c54d618a-dfec-4f2a-a288-f5188c8a128a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60324dc7d701bc064e14ae4f474bce10f6ee4ff3c3c942eab749e7166082cf9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ed4f4d3564674bf5cb4e7a76dc0fd95e203ecc8950798e59a1fb6c9ab4a20f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://def9ac840fb8ecf2778bed88ea980eddca7c5b2aba4bd1afb967d62fec3f7bae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b73be664f5412c98b89d971a2cac37c494151c293fdb44a98b390978658a6643\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a847d04e22e98091c0037a89054b9236228b701b8c7ba31940a56df0f37b95c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcae78a6133d97f6a32dd7421ccec4333b970a237404e056a5cc601f8eefdc8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8baee0f3d1e686bc2c17b0777f4eacc31c7f1fe44720e68f94612e6c806bc8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53c40176ba965341b13456a99ba4fad80dc6115939e3c4c884e74f4e026a7535\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-07T16:02:50Z\\\",\\\"message\\\":\\\"rvice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1207 16:02:49.273594 6061 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1207 16:02:49.273647 6061 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1207 16:02:49.273655 6061 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1207 16:02:49.273699 6061 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1207 16:02:49.273719 6061 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1207 16:02:49.273725 6061 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1207 16:02:49.273749 6061 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1207 16:02:49.273770 6061 factory.go:656] Stopping watch factory\\\\nI1207 16:02:49.273826 6061 ovnkube.go:599] Stopped ovnkube\\\\nI1207 16:02:49.273858 6061 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1207 16:02:49.273867 6061 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1207 16:02:49.273875 6061 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1207 16:02:49.273882 6061 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1207 16:02:49.273890 6061 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1207 16:02:49.273911 6061 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1207 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:47Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8baee0f3d1e686bc2c17b0777f4eacc31c7f1fe44720e68f94612e6c806bc8b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-07T16:02:52Z\\\",\\\"message\\\":\\\"r occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:51Z is after 2025-08-24T17:21:41Z]\\\\nI1207 16:02:51.943802 6191 services_controller.go:444] Built service openshift-marketplace/redhat-marketplace LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1207 16:02:51.943764 6191 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-route-controller-manager/route-controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"18746a4d-8a63-458a-b7e3-8fb89ff95fc0\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-route-controller-manager/route-controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-route-controller-manager/route-controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcfdffe822635b0ae13b253e6133449fe00461b2792be43d913c4a77219b8255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9bfe2ffce7a7b7c320474117734fc05246cc2b4fd744833e896e8e7c83f5e4d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9bfe2ffce7a7b7c320474117734fc05246cc2b4fd744833e896e8e7c83f5e4d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qvc7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:53Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.323119 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2jjld" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dee6f5c-861c-4819-9309-7791574c7b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51c1d4159ab11087839a2c2d192566fb9322447c26df560bf09517c906ececb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmdns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2jjld\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:53Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.334344 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4wr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0baf8575-26cf-4a72-8a6d-ae94c078eaad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0ca2b3cd5a11033afd219b0125be7ab5acd750935d7eb89a470eff3d93a0838\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p2trd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4wr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:53Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.343154 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-m5jl5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"824ee59f-1474-4994-a70e-462878521514\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a15fe50cc75515704779646cc19aa50004f0db9aa26fe2535110df749bbfb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhwzn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-m5jl5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:53Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.354819 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdee9957-48b9-458f-ba71-e0cd39f80596\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73f85248b4d68a9beecdacb3a01f433e97a0ffea2247ce92a41ef96855509377\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6276df652efe6ca935e745f24ef20f87a16465475139230c63b9999234a4d12a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3b1dd202586ee755ca89fbe12b30bbb12f95a62514f6e4d93cbbdbe55181e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd318647816e4a27849666b0ec4fc2ca4905295661c557dc381e1144f22cf284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d31de0326a05216fa10442bdb47151c6e073c97e9a16c34e793028e5e253756\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"le observer\\\\nW1207 16:02:34.645464 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1207 16:02:34.645587 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 16:02:34.646501 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2082283479/tls.crt::/tmp/serving-cert-2082283479/tls.key\\\\\\\"\\\\nI1207 16:02:34.972041 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 16:02:34.973702 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 16:02:34.973717 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 16:02:34.973738 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 16:02:34.973743 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 16:02:34.977185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1207 16:02:34.977204 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1207 16:02:34.977209 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977217 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 16:02:34.977225 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 16:02:34.977228 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 16:02:34.977232 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1207 16:02:34.979033 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27bd4a54f920b2725987c1c7fd2d9fdf16cd08738c246c333a4e9164a3a582a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:53Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.365063 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:53Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.377691 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:53Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.381250 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mx85f" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.387801 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.387856 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.387868 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.387885 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.387896 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:53Z","lastTransitionTime":"2025-12-07T16:02:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.391733 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f95ef61779507288a03bda8eddf745e105f705998a117a30789675bfa452d88f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b955b23163e7c8c458a3bbe2f681633e6a663ce76748febcc0b3a3c37f3ca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:53Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:53 crc kubenswrapper[4716]: W1207 16:02:53.393871 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod118dd9ba_437c_4e5c_b5ff_4239ae6d4a63.slice/crio-a015a0105713df559050ddefa63a97c55aecd6b40f4c7c7248ae989df36f3fb4 WatchSource:0}: Error finding container a015a0105713df559050ddefa63a97c55aecd6b40f4c7c7248ae989df36f3fb4: Status 404 returned error can't find the container with id a015a0105713df559050ddefa63a97c55aecd6b40f4c7c7248ae989df36f3fb4 Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.405548 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef3335ae3875a75ab64685fc7e3590177280d651e8f8292c6f5ea6f472e55789\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:53Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.490721 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.490777 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.490794 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.490818 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.490834 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:53Z","lastTransitionTime":"2025-12-07T16:02:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.592829 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.592891 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.592901 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.592916 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.592925 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:53Z","lastTransitionTime":"2025-12-07T16:02:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.657567 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.657575 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:02:53 crc kubenswrapper[4716]: E1207 16:02:53.657715 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 16:02:53 crc kubenswrapper[4716]: E1207 16:02:53.657822 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.696231 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.696268 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.696277 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.696292 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.696303 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:53Z","lastTransitionTime":"2025-12-07T16:02:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.798216 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.798280 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.798294 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.798312 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.798324 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:53Z","lastTransitionTime":"2025-12-07T16:02:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.900104 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.900154 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.900167 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.900184 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.900195 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:53Z","lastTransitionTime":"2025-12-07T16:02:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.951048 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qvc7p_c54d618a-dfec-4f2a-a288-f5188c8a128a/ovnkube-controller/1.log" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.954751 4716 scope.go:117] "RemoveContainer" containerID="a8baee0f3d1e686bc2c17b0777f4eacc31c7f1fe44720e68f94612e6c806bc8b" Dec 07 16:02:53 crc kubenswrapper[4716]: E1207 16:02:53.954989 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-qvc7p_openshift-ovn-kubernetes(c54d618a-dfec-4f2a-a288-f5188c8a128a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" podUID="c54d618a-dfec-4f2a-a288-f5188c8a128a" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.955211 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mx85f" event={"ID":"118dd9ba-437c-4e5c-b5ff-4239ae6d4a63","Type":"ContainerStarted","Data":"2ffc069d2b21fa80871b1b97f2c72c37880d9a309cbf8b7daf7d240f2c332a44"} Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.955287 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mx85f" event={"ID":"118dd9ba-437c-4e5c-b5ff-4239ae6d4a63","Type":"ContainerStarted","Data":"bcd6c8deb184d0c713c5e7b7cc82fa38d3cd498517911ba7a0ec3cb88a3437a4"} Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.955311 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mx85f" event={"ID":"118dd9ba-437c-4e5c-b5ff-4239ae6d4a63","Type":"ContainerStarted","Data":"a015a0105713df559050ddefa63a97c55aecd6b40f4c7c7248ae989df36f3fb4"} Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.969751 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef3335ae3875a75ab64685fc7e3590177280d651e8f8292c6f5ea6f472e55789\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:53Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.981171 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:53Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:53 crc kubenswrapper[4716]: I1207 16:02:53.995455 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f95ef61779507288a03bda8eddf745e105f705998a117a30789675bfa452d88f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b955b23163e7c8c458a3bbe2f681633e6a663ce76748febcc0b3a3c37f3ca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:53Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.002886 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.002931 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.002944 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.002969 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.002983 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:54Z","lastTransitionTime":"2025-12-07T16:02:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.014591 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7961cc9a-734e-42c2-b1ff-7075f3f20091\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84a36b76f20618c7885d611ecad9db29bac5101c27572be31d3acbb90c581291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0a0eca3c2087a457c0e448e649345ec4595a9c8a019aed2cddc5f30c77871e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://404426a171b8676eddd4a00996c76afab7e269643fa9608f1e1b6bdba385071d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fd3c2941782921f99a6cdc3de4e892234548bde527fec15a8ac92e513566e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e23abfdca314ed43d594333c9f9e401cc8ef7dad39b8ebac84a836b06ee5c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09b38cd4b389ea477c802a842ee4f2297a64d64684624117163e819539c257c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09b38cd4b389ea477c802a842ee4f2297a64d64684624117163e819539c257c6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c00f7a8e024e0e0dfd7b73e16d1d5437602e79ea95b73c21f24b209edef60e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c00f7a8e024e0e0dfd7b73e16d1d5437602e79ea95b73c21f24b209edef60e98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3f6392d426fa2cd788ec3a5c3bff4ac39483a54dfb2624a1c4370522413a350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3f6392d426fa2cd788ec3a5c3bff4ac39483a54dfb2624a1c4370522413a350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:54Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.028310 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3067de7c-c11c-4be1-bd94-d6572c266795\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d14d018fa044e9e80fe81e7e38df935992d14c773ed2b2193fd5cb873e5d0947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a9e9cf048220ca14c996847e9ced9f2367b9331dc6c2836dd023658855cd79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6d3c911874095b3c7cb062d03070e5090a7bb2456cb20cb4ce078f502515cda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ac6ce7701f1c6aa1ea8fcf7a3035e3b0eaa28ae3cb6afc442372bd9117bdde4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:54Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.040278 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4742ea84b3a2c9461d9b900beee7354a2ded24c39ad23e007cbf270ae12377c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:54Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.052458 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:54Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.062515 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c15b59eb-565d-4556-a4ce-75afdf159dc8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://530a2935826940767ea33fb4f37fa11f7dee4950d867fbcd6af72836d80f7db3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jks8z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ccaf6c1e6a978eb978a1c207b6c4c00b23737f78b743f966579aca781e22e0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jks8z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zcxxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:54Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.077377 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8ps9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98015cea-d753-4e7a-b2b5-5ae7733a81e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d68509b673eecd37725cdcc48e3d0612c3b9c7601433f2b72c21d662140daa59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f84f52a073437543721010dc440339654ffb5e04eeb6cc7afbcc62f00f64c8e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f84f52a073437543721010dc440339654ffb5e04eeb6cc7afbcc62f00f64c8e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f6fe1cb25218cb3ac20488dc47efa4dc6a26a43dbc8306b5ee38cbe1a186c67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f6fe1cb25218cb3ac20488dc47efa4dc6a26a43dbc8306b5ee38cbe1a186c67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14b748a8247f6863e8be61e90057a06626dfd3cd4f01bd3a11629f700f56378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f14b748a8247f6863e8be61e90057a06626dfd3cd4f01bd3a11629f700f56378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ec379c0d573cb9b5ffcfc5ca7c653ba60e2fbdd7e3d37836102e26618ce43e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ec379c0d573cb9b5ffcfc5ca7c653ba60e2fbdd7e3d37836102e26618ce43e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c75bf2a6703cf99865d4ed5fa000e1e3b1330d4946176cf585fec1a2f226df0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c75bf2a6703cf99865d4ed5fa000e1e3b1330d4946176cf585fec1a2f226df0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de45812c9a5a5bfe6b0146828d03fab28362b03d04ef861c150b9ffbf181d345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de45812c9a5a5bfe6b0146828d03fab28362b03d04ef861c150b9ffbf181d345\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8ps9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:54Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.098883 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c54d618a-dfec-4f2a-a288-f5188c8a128a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60324dc7d701bc064e14ae4f474bce10f6ee4ff3c3c942eab749e7166082cf9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ed4f4d3564674bf5cb4e7a76dc0fd95e203ecc8950798e59a1fb6c9ab4a20f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://def9ac840fb8ecf2778bed88ea980eddca7c5b2aba4bd1afb967d62fec3f7bae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b73be664f5412c98b89d971a2cac37c494151c293fdb44a98b390978658a6643\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a847d04e22e98091c0037a89054b9236228b701b8c7ba31940a56df0f37b95c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcae78a6133d97f6a32dd7421ccec4333b970a237404e056a5cc601f8eefdc8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8baee0f3d1e686bc2c17b0777f4eacc31c7f1fe44720e68f94612e6c806bc8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8baee0f3d1e686bc2c17b0777f4eacc31c7f1fe44720e68f94612e6c806bc8b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-07T16:02:52Z\\\",\\\"message\\\":\\\"r occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:51Z is after 2025-08-24T17:21:41Z]\\\\nI1207 16:02:51.943802 6191 services_controller.go:444] Built service openshift-marketplace/redhat-marketplace LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1207 16:02:51.943764 6191 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-route-controller-manager/route-controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"18746a4d-8a63-458a-b7e3-8fb89ff95fc0\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-route-controller-manager/route-controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-route-controller-manager/route-controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-qvc7p_openshift-ovn-kubernetes(c54d618a-dfec-4f2a-a288-f5188c8a128a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcfdffe822635b0ae13b253e6133449fe00461b2792be43d913c4a77219b8255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9bfe2ffce7a7b7c320474117734fc05246cc2b4fd744833e896e8e7c83f5e4d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9bfe2ffce7a7b7c320474117734fc05246cc2b4fd744833e896e8e7c83f5e4d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qvc7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:54Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.105434 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.105494 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.105513 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.105539 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.105557 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:54Z","lastTransitionTime":"2025-12-07T16:02:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.111022 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mx85f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"118dd9ba-437c-4e5c-b5ff-4239ae6d4a63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gg5nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gg5nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mx85f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:54Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.125633 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdee9957-48b9-458f-ba71-e0cd39f80596\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73f85248b4d68a9beecdacb3a01f433e97a0ffea2247ce92a41ef96855509377\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6276df652efe6ca935e745f24ef20f87a16465475139230c63b9999234a4d12a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3b1dd202586ee755ca89fbe12b30bbb12f95a62514f6e4d93cbbdbe55181e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd318647816e4a27849666b0ec4fc2ca4905295661c557dc381e1144f22cf284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d31de0326a05216fa10442bdb47151c6e073c97e9a16c34e793028e5e253756\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"le observer\\\\nW1207 16:02:34.645464 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1207 16:02:34.645587 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 16:02:34.646501 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2082283479/tls.crt::/tmp/serving-cert-2082283479/tls.key\\\\\\\"\\\\nI1207 16:02:34.972041 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 16:02:34.973702 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 16:02:34.973717 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 16:02:34.973738 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 16:02:34.973743 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 16:02:34.977185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1207 16:02:34.977204 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1207 16:02:34.977209 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977217 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 16:02:34.977225 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 16:02:34.977228 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 16:02:34.977232 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1207 16:02:34.979033 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27bd4a54f920b2725987c1c7fd2d9fdf16cd08738c246c333a4e9164a3a582a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:54Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.138796 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:54Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.148375 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2jjld" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dee6f5c-861c-4819-9309-7791574c7b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51c1d4159ab11087839a2c2d192566fb9322447c26df560bf09517c906ececb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmdns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2jjld\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:54Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.154629 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-nx2lh"] Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.155154 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nx2lh" Dec 07 16:02:54 crc kubenswrapper[4716]: E1207 16:02:54.155217 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nx2lh" podUID="04b37f6c-457b-4f9a-8733-e88cb9644483" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.159778 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4wr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0baf8575-26cf-4a72-8a6d-ae94c078eaad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0ca2b3cd5a11033afd219b0125be7ab5acd750935d7eb89a470eff3d93a0838\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p2trd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4wr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:54Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.169182 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-m5jl5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"824ee59f-1474-4994-a70e-462878521514\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a15fe50cc75515704779646cc19aa50004f0db9aa26fe2535110df749bbfb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhwzn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-m5jl5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:54Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.188024 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c54d618a-dfec-4f2a-a288-f5188c8a128a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60324dc7d701bc064e14ae4f474bce10f6ee4ff3c3c942eab749e7166082cf9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ed4f4d3564674bf5cb4e7a76dc0fd95e203ecc8950798e59a1fb6c9ab4a20f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://def9ac840fb8ecf2778bed88ea980eddca7c5b2aba4bd1afb967d62fec3f7bae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b73be664f5412c98b89d971a2cac37c494151c293fdb44a98b390978658a6643\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a847d04e22e98091c0037a89054b9236228b701b8c7ba31940a56df0f37b95c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcae78a6133d97f6a32dd7421ccec4333b970a237404e056a5cc601f8eefdc8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8baee0f3d1e686bc2c17b0777f4eacc31c7f1fe44720e68f94612e6c806bc8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8baee0f3d1e686bc2c17b0777f4eacc31c7f1fe44720e68f94612e6c806bc8b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-07T16:02:52Z\\\",\\\"message\\\":\\\"r occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:51Z is after 2025-08-24T17:21:41Z]\\\\nI1207 16:02:51.943802 6191 services_controller.go:444] Built service openshift-marketplace/redhat-marketplace LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1207 16:02:51.943764 6191 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-route-controller-manager/route-controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"18746a4d-8a63-458a-b7e3-8fb89ff95fc0\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-route-controller-manager/route-controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-route-controller-manager/route-controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-qvc7p_openshift-ovn-kubernetes(c54d618a-dfec-4f2a-a288-f5188c8a128a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcfdffe822635b0ae13b253e6133449fe00461b2792be43d913c4a77219b8255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9bfe2ffce7a7b7c320474117734fc05246cc2b4fd744833e896e8e7c83f5e4d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9bfe2ffce7a7b7c320474117734fc05246cc2b4fd744833e896e8e7c83f5e4d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qvc7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:54Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.197968 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mx85f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"118dd9ba-437c-4e5c-b5ff-4239ae6d4a63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bcd6c8deb184d0c713c5e7b7cc82fa38d3cd498517911ba7a0ec3cb88a3437a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gg5nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ffc069d2b21fa80871b1b97f2c72c37880d9a309cbf8b7daf7d240f2c332a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gg5nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mx85f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:54Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.206922 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.206952 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.206961 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.206975 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.206987 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:54Z","lastTransitionTime":"2025-12-07T16:02:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.209160 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nx2lh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04b37f6c-457b-4f9a-8733-e88cb9644483\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv6td\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv6td\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nx2lh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:54Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.220408 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-m5jl5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"824ee59f-1474-4994-a70e-462878521514\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a15fe50cc75515704779646cc19aa50004f0db9aa26fe2535110df749bbfb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhwzn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-m5jl5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:54Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.238428 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdee9957-48b9-458f-ba71-e0cd39f80596\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73f85248b4d68a9beecdacb3a01f433e97a0ffea2247ce92a41ef96855509377\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6276df652efe6ca935e745f24ef20f87a16465475139230c63b9999234a4d12a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3b1dd202586ee755ca89fbe12b30bbb12f95a62514f6e4d93cbbdbe55181e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd318647816e4a27849666b0ec4fc2ca4905295661c557dc381e1144f22cf284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d31de0326a05216fa10442bdb47151c6e073c97e9a16c34e793028e5e253756\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"le observer\\\\nW1207 16:02:34.645464 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1207 16:02:34.645587 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 16:02:34.646501 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2082283479/tls.crt::/tmp/serving-cert-2082283479/tls.key\\\\\\\"\\\\nI1207 16:02:34.972041 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 16:02:34.973702 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 16:02:34.973717 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 16:02:34.973738 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 16:02:34.973743 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 16:02:34.977185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1207 16:02:34.977204 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1207 16:02:34.977209 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977217 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 16:02:34.977225 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 16:02:34.977228 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 16:02:34.977232 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1207 16:02:34.979033 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27bd4a54f920b2725987c1c7fd2d9fdf16cd08738c246c333a4e9164a3a582a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:54Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.250638 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:54Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.262566 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2jjld" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dee6f5c-861c-4819-9309-7791574c7b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51c1d4159ab11087839a2c2d192566fb9322447c26df560bf09517c906ececb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmdns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2jjld\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:54Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.269040 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bv6td\" (UniqueName: \"kubernetes.io/projected/04b37f6c-457b-4f9a-8733-e88cb9644483-kube-api-access-bv6td\") pod \"network-metrics-daemon-nx2lh\" (UID: \"04b37f6c-457b-4f9a-8733-e88cb9644483\") " pod="openshift-multus/network-metrics-daemon-nx2lh" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.269125 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/04b37f6c-457b-4f9a-8733-e88cb9644483-metrics-certs\") pod \"network-metrics-daemon-nx2lh\" (UID: \"04b37f6c-457b-4f9a-8733-e88cb9644483\") " pod="openshift-multus/network-metrics-daemon-nx2lh" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.281469 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4wr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0baf8575-26cf-4a72-8a6d-ae94c078eaad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0ca2b3cd5a11033afd219b0125be7ab5acd750935d7eb89a470eff3d93a0838\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p2trd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4wr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:54Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.297694 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef3335ae3875a75ab64685fc7e3590177280d651e8f8292c6f5ea6f472e55789\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:54Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.309401 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:54Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.310690 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.310738 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.310748 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.310762 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.310771 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:54Z","lastTransitionTime":"2025-12-07T16:02:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.327173 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f95ef61779507288a03bda8eddf745e105f705998a117a30789675bfa452d88f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b955b23163e7c8c458a3bbe2f681633e6a663ce76748febcc0b3a3c37f3ca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:54Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.341818 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c15b59eb-565d-4556-a4ce-75afdf159dc8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://530a2935826940767ea33fb4f37fa11f7dee4950d867fbcd6af72836d80f7db3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jks8z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ccaf6c1e6a978eb978a1c207b6c4c00b23737f78b743f966579aca781e22e0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jks8z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zcxxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:54Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.364608 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8ps9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98015cea-d753-4e7a-b2b5-5ae7733a81e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d68509b673eecd37725cdcc48e3d0612c3b9c7601433f2b72c21d662140daa59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f84f52a073437543721010dc440339654ffb5e04eeb6cc7afbcc62f00f64c8e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f84f52a073437543721010dc440339654ffb5e04eeb6cc7afbcc62f00f64c8e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f6fe1cb25218cb3ac20488dc47efa4dc6a26a43dbc8306b5ee38cbe1a186c67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f6fe1cb25218cb3ac20488dc47efa4dc6a26a43dbc8306b5ee38cbe1a186c67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14b748a8247f6863e8be61e90057a06626dfd3cd4f01bd3a11629f700f56378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f14b748a8247f6863e8be61e90057a06626dfd3cd4f01bd3a11629f700f56378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ec379c0d573cb9b5ffcfc5ca7c653ba60e2fbdd7e3d37836102e26618ce43e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ec379c0d573cb9b5ffcfc5ca7c653ba60e2fbdd7e3d37836102e26618ce43e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c75bf2a6703cf99865d4ed5fa000e1e3b1330d4946176cf585fec1a2f226df0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c75bf2a6703cf99865d4ed5fa000e1e3b1330d4946176cf585fec1a2f226df0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de45812c9a5a5bfe6b0146828d03fab28362b03d04ef861c150b9ffbf181d345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de45812c9a5a5bfe6b0146828d03fab28362b03d04ef861c150b9ffbf181d345\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8ps9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:54Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.369930 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bv6td\" (UniqueName: \"kubernetes.io/projected/04b37f6c-457b-4f9a-8733-e88cb9644483-kube-api-access-bv6td\") pod \"network-metrics-daemon-nx2lh\" (UID: \"04b37f6c-457b-4f9a-8733-e88cb9644483\") " pod="openshift-multus/network-metrics-daemon-nx2lh" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.370124 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/04b37f6c-457b-4f9a-8733-e88cb9644483-metrics-certs\") pod \"network-metrics-daemon-nx2lh\" (UID: \"04b37f6c-457b-4f9a-8733-e88cb9644483\") " pod="openshift-multus/network-metrics-daemon-nx2lh" Dec 07 16:02:54 crc kubenswrapper[4716]: E1207 16:02:54.370265 4716 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 07 16:02:54 crc kubenswrapper[4716]: E1207 16:02:54.370343 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/04b37f6c-457b-4f9a-8733-e88cb9644483-metrics-certs podName:04b37f6c-457b-4f9a-8733-e88cb9644483 nodeName:}" failed. No retries permitted until 2025-12-07 16:02:54.870320704 +0000 UTC m=+37.560605656 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/04b37f6c-457b-4f9a-8733-e88cb9644483-metrics-certs") pod "network-metrics-daemon-nx2lh" (UID: "04b37f6c-457b-4f9a-8733-e88cb9644483") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.388833 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7961cc9a-734e-42c2-b1ff-7075f3f20091\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84a36b76f20618c7885d611ecad9db29bac5101c27572be31d3acbb90c581291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0a0eca3c2087a457c0e448e649345ec4595a9c8a019aed2cddc5f30c77871e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://404426a171b8676eddd4a00996c76afab7e269643fa9608f1e1b6bdba385071d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fd3c2941782921f99a6cdc3de4e892234548bde527fec15a8ac92e513566e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e23abfdca314ed43d594333c9f9e401cc8ef7dad39b8ebac84a836b06ee5c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09b38cd4b389ea477c802a842ee4f2297a64d64684624117163e819539c257c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09b38cd4b389ea477c802a842ee4f2297a64d64684624117163e819539c257c6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c00f7a8e024e0e0dfd7b73e16d1d5437602e79ea95b73c21f24b209edef60e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c00f7a8e024e0e0dfd7b73e16d1d5437602e79ea95b73c21f24b209edef60e98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3f6392d426fa2cd788ec3a5c3bff4ac39483a54dfb2624a1c4370522413a350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3f6392d426fa2cd788ec3a5c3bff4ac39483a54dfb2624a1c4370522413a350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:54Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.410630 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3067de7c-c11c-4be1-bd94-d6572c266795\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d14d018fa044e9e80fe81e7e38df935992d14c773ed2b2193fd5cb873e5d0947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a9e9cf048220ca14c996847e9ced9f2367b9331dc6c2836dd023658855cd79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6d3c911874095b3c7cb062d03070e5090a7bb2456cb20cb4ce078f502515cda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ac6ce7701f1c6aa1ea8fcf7a3035e3b0eaa28ae3cb6afc442372bd9117bdde4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:54Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.412666 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.412713 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.412731 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.412754 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.412771 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:54Z","lastTransitionTime":"2025-12-07T16:02:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.426775 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4742ea84b3a2c9461d9b900beee7354a2ded24c39ad23e007cbf270ae12377c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:54Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.442487 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:54Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.477176 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bv6td\" (UniqueName: \"kubernetes.io/projected/04b37f6c-457b-4f9a-8733-e88cb9644483-kube-api-access-bv6td\") pod \"network-metrics-daemon-nx2lh\" (UID: \"04b37f6c-457b-4f9a-8733-e88cb9644483\") " pod="openshift-multus/network-metrics-daemon-nx2lh" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.515243 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.515274 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.515285 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.515298 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.515307 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:54Z","lastTransitionTime":"2025-12-07T16:02:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.617687 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.617736 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.617753 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.617775 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.617791 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:54Z","lastTransitionTime":"2025-12-07T16:02:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.657252 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 16:02:54 crc kubenswrapper[4716]: E1207 16:02:54.657433 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.720036 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.720156 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.720182 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.720211 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.720229 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:54Z","lastTransitionTime":"2025-12-07T16:02:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.823177 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.823229 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.823246 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.823268 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.823289 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:54Z","lastTransitionTime":"2025-12-07T16:02:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.875370 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/04b37f6c-457b-4f9a-8733-e88cb9644483-metrics-certs\") pod \"network-metrics-daemon-nx2lh\" (UID: \"04b37f6c-457b-4f9a-8733-e88cb9644483\") " pod="openshift-multus/network-metrics-daemon-nx2lh" Dec 07 16:02:54 crc kubenswrapper[4716]: E1207 16:02:54.875711 4716 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 07 16:02:54 crc kubenswrapper[4716]: E1207 16:02:54.875807 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/04b37f6c-457b-4f9a-8733-e88cb9644483-metrics-certs podName:04b37f6c-457b-4f9a-8733-e88cb9644483 nodeName:}" failed. No retries permitted until 2025-12-07 16:02:55.875773412 +0000 UTC m=+38.566058374 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/04b37f6c-457b-4f9a-8733-e88cb9644483-metrics-certs") pod "network-metrics-daemon-nx2lh" (UID: "04b37f6c-457b-4f9a-8733-e88cb9644483") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.925176 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.925247 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.925260 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.925304 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:54 crc kubenswrapper[4716]: I1207 16:02:54.925316 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:54Z","lastTransitionTime":"2025-12-07T16:02:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:55 crc kubenswrapper[4716]: I1207 16:02:55.028275 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:55 crc kubenswrapper[4716]: I1207 16:02:55.028345 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:55 crc kubenswrapper[4716]: I1207 16:02:55.028364 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:55 crc kubenswrapper[4716]: I1207 16:02:55.028388 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:55 crc kubenswrapper[4716]: I1207 16:02:55.028473 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:55Z","lastTransitionTime":"2025-12-07T16:02:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:55 crc kubenswrapper[4716]: I1207 16:02:55.131410 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:55 crc kubenswrapper[4716]: I1207 16:02:55.131477 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:55 crc kubenswrapper[4716]: I1207 16:02:55.131489 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:55 crc kubenswrapper[4716]: I1207 16:02:55.131508 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:55 crc kubenswrapper[4716]: I1207 16:02:55.131521 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:55Z","lastTransitionTime":"2025-12-07T16:02:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:55 crc kubenswrapper[4716]: I1207 16:02:55.234459 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:55 crc kubenswrapper[4716]: I1207 16:02:55.234500 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:55 crc kubenswrapper[4716]: I1207 16:02:55.234509 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:55 crc kubenswrapper[4716]: I1207 16:02:55.234523 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:55 crc kubenswrapper[4716]: I1207 16:02:55.234533 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:55Z","lastTransitionTime":"2025-12-07T16:02:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:55 crc kubenswrapper[4716]: I1207 16:02:55.337358 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:55 crc kubenswrapper[4716]: I1207 16:02:55.337417 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:55 crc kubenswrapper[4716]: I1207 16:02:55.337434 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:55 crc kubenswrapper[4716]: I1207 16:02:55.337456 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:55 crc kubenswrapper[4716]: I1207 16:02:55.337473 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:55Z","lastTransitionTime":"2025-12-07T16:02:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:55 crc kubenswrapper[4716]: I1207 16:02:55.440971 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:55 crc kubenswrapper[4716]: I1207 16:02:55.441042 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:55 crc kubenswrapper[4716]: I1207 16:02:55.441065 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:55 crc kubenswrapper[4716]: I1207 16:02:55.441131 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:55 crc kubenswrapper[4716]: I1207 16:02:55.441158 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:55Z","lastTransitionTime":"2025-12-07T16:02:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:55 crc kubenswrapper[4716]: I1207 16:02:55.543427 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:55 crc kubenswrapper[4716]: I1207 16:02:55.543460 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:55 crc kubenswrapper[4716]: I1207 16:02:55.543468 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:55 crc kubenswrapper[4716]: I1207 16:02:55.543482 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:55 crc kubenswrapper[4716]: I1207 16:02:55.543492 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:55Z","lastTransitionTime":"2025-12-07T16:02:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:55 crc kubenswrapper[4716]: I1207 16:02:55.645539 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:55 crc kubenswrapper[4716]: I1207 16:02:55.645607 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:55 crc kubenswrapper[4716]: I1207 16:02:55.645625 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:55 crc kubenswrapper[4716]: I1207 16:02:55.645652 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:55 crc kubenswrapper[4716]: I1207 16:02:55.645669 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:55Z","lastTransitionTime":"2025-12-07T16:02:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:55 crc kubenswrapper[4716]: I1207 16:02:55.656937 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 16:02:55 crc kubenswrapper[4716]: I1207 16:02:55.657140 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nx2lh" Dec 07 16:02:55 crc kubenswrapper[4716]: I1207 16:02:55.657171 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:02:55 crc kubenswrapper[4716]: E1207 16:02:55.657344 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 16:02:55 crc kubenswrapper[4716]: E1207 16:02:55.657502 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nx2lh" podUID="04b37f6c-457b-4f9a-8733-e88cb9644483" Dec 07 16:02:55 crc kubenswrapper[4716]: E1207 16:02:55.657711 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 16:02:55 crc kubenswrapper[4716]: I1207 16:02:55.748012 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:55 crc kubenswrapper[4716]: I1207 16:02:55.748050 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:55 crc kubenswrapper[4716]: I1207 16:02:55.748058 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:55 crc kubenswrapper[4716]: I1207 16:02:55.748071 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:55 crc kubenswrapper[4716]: I1207 16:02:55.748099 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:55Z","lastTransitionTime":"2025-12-07T16:02:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:55 crc kubenswrapper[4716]: I1207 16:02:55.851441 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:55 crc kubenswrapper[4716]: I1207 16:02:55.851480 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:55 crc kubenswrapper[4716]: I1207 16:02:55.851490 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:55 crc kubenswrapper[4716]: I1207 16:02:55.851505 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:55 crc kubenswrapper[4716]: I1207 16:02:55.851515 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:55Z","lastTransitionTime":"2025-12-07T16:02:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:55 crc kubenswrapper[4716]: I1207 16:02:55.888570 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/04b37f6c-457b-4f9a-8733-e88cb9644483-metrics-certs\") pod \"network-metrics-daemon-nx2lh\" (UID: \"04b37f6c-457b-4f9a-8733-e88cb9644483\") " pod="openshift-multus/network-metrics-daemon-nx2lh" Dec 07 16:02:55 crc kubenswrapper[4716]: E1207 16:02:55.888770 4716 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 07 16:02:55 crc kubenswrapper[4716]: E1207 16:02:55.888849 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/04b37f6c-457b-4f9a-8733-e88cb9644483-metrics-certs podName:04b37f6c-457b-4f9a-8733-e88cb9644483 nodeName:}" failed. No retries permitted until 2025-12-07 16:02:57.888828847 +0000 UTC m=+40.579113759 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/04b37f6c-457b-4f9a-8733-e88cb9644483-metrics-certs") pod "network-metrics-daemon-nx2lh" (UID: "04b37f6c-457b-4f9a-8733-e88cb9644483") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 07 16:02:55 crc kubenswrapper[4716]: I1207 16:02:55.954012 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:55 crc kubenswrapper[4716]: I1207 16:02:55.954120 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:55 crc kubenswrapper[4716]: I1207 16:02:55.954141 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:55 crc kubenswrapper[4716]: I1207 16:02:55.954165 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:55 crc kubenswrapper[4716]: I1207 16:02:55.954183 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:55Z","lastTransitionTime":"2025-12-07T16:02:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:56 crc kubenswrapper[4716]: I1207 16:02:56.057207 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:56 crc kubenswrapper[4716]: I1207 16:02:56.057262 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:56 crc kubenswrapper[4716]: I1207 16:02:56.057278 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:56 crc kubenswrapper[4716]: I1207 16:02:56.057302 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:56 crc kubenswrapper[4716]: I1207 16:02:56.057367 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:56Z","lastTransitionTime":"2025-12-07T16:02:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:56 crc kubenswrapper[4716]: I1207 16:02:56.159928 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:56 crc kubenswrapper[4716]: I1207 16:02:56.159993 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:56 crc kubenswrapper[4716]: I1207 16:02:56.160011 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:56 crc kubenswrapper[4716]: I1207 16:02:56.160037 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:56 crc kubenswrapper[4716]: I1207 16:02:56.160056 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:56Z","lastTransitionTime":"2025-12-07T16:02:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:56 crc kubenswrapper[4716]: I1207 16:02:56.262540 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:56 crc kubenswrapper[4716]: I1207 16:02:56.262662 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:56 crc kubenswrapper[4716]: I1207 16:02:56.262685 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:56 crc kubenswrapper[4716]: I1207 16:02:56.262708 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:56 crc kubenswrapper[4716]: I1207 16:02:56.262727 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:56Z","lastTransitionTime":"2025-12-07T16:02:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:56 crc kubenswrapper[4716]: I1207 16:02:56.366522 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:56 crc kubenswrapper[4716]: I1207 16:02:56.366568 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:56 crc kubenswrapper[4716]: I1207 16:02:56.366585 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:56 crc kubenswrapper[4716]: I1207 16:02:56.366607 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:56 crc kubenswrapper[4716]: I1207 16:02:56.366624 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:56Z","lastTransitionTime":"2025-12-07T16:02:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:56 crc kubenswrapper[4716]: I1207 16:02:56.389106 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:56 crc kubenswrapper[4716]: I1207 16:02:56.389143 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:56 crc kubenswrapper[4716]: I1207 16:02:56.389151 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:56 crc kubenswrapper[4716]: I1207 16:02:56.389167 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:56 crc kubenswrapper[4716]: I1207 16:02:56.389178 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:56Z","lastTransitionTime":"2025-12-07T16:02:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:56 crc kubenswrapper[4716]: E1207 16:02:56.406637 4716 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:02:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:02:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:02:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:02:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a8d1b7d4-7ec4-40c7-86b0-291db1f38894\\\",\\\"systemUUID\\\":\\\"a57d09b0-0224-46eb-a3ac-a2b60dad2250\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:56Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:56 crc kubenswrapper[4716]: I1207 16:02:56.411005 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:56 crc kubenswrapper[4716]: I1207 16:02:56.411061 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:56 crc kubenswrapper[4716]: I1207 16:02:56.411136 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:56 crc kubenswrapper[4716]: I1207 16:02:56.411166 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:56 crc kubenswrapper[4716]: I1207 16:02:56.411185 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:56Z","lastTransitionTime":"2025-12-07T16:02:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:56 crc kubenswrapper[4716]: E1207 16:02:56.425283 4716 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:02:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:02:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:02:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:02:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a8d1b7d4-7ec4-40c7-86b0-291db1f38894\\\",\\\"systemUUID\\\":\\\"a57d09b0-0224-46eb-a3ac-a2b60dad2250\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:56Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:56 crc kubenswrapper[4716]: I1207 16:02:56.428656 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:56 crc kubenswrapper[4716]: I1207 16:02:56.428683 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:56 crc kubenswrapper[4716]: I1207 16:02:56.428692 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:56 crc kubenswrapper[4716]: I1207 16:02:56.428706 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:56 crc kubenswrapper[4716]: I1207 16:02:56.428717 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:56Z","lastTransitionTime":"2025-12-07T16:02:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:56 crc kubenswrapper[4716]: E1207 16:02:56.439444 4716 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:02:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:02:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:02:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:02:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a8d1b7d4-7ec4-40c7-86b0-291db1f38894\\\",\\\"systemUUID\\\":\\\"a57d09b0-0224-46eb-a3ac-a2b60dad2250\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:56Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:56 crc kubenswrapper[4716]: I1207 16:02:56.443349 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:56 crc kubenswrapper[4716]: I1207 16:02:56.443385 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:56 crc kubenswrapper[4716]: I1207 16:02:56.443394 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:56 crc kubenswrapper[4716]: I1207 16:02:56.443408 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:56 crc kubenswrapper[4716]: I1207 16:02:56.443418 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:56Z","lastTransitionTime":"2025-12-07T16:02:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:56 crc kubenswrapper[4716]: E1207 16:02:56.458953 4716 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:02:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:02:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:02:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:02:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a8d1b7d4-7ec4-40c7-86b0-291db1f38894\\\",\\\"systemUUID\\\":\\\"a57d09b0-0224-46eb-a3ac-a2b60dad2250\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:56Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:56 crc kubenswrapper[4716]: I1207 16:02:56.462307 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:56 crc kubenswrapper[4716]: I1207 16:02:56.462345 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:56 crc kubenswrapper[4716]: I1207 16:02:56.462355 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:56 crc kubenswrapper[4716]: I1207 16:02:56.462369 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:56 crc kubenswrapper[4716]: I1207 16:02:56.462379 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:56Z","lastTransitionTime":"2025-12-07T16:02:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:56 crc kubenswrapper[4716]: E1207 16:02:56.475206 4716 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:02:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:02:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:02:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:02:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a8d1b7d4-7ec4-40c7-86b0-291db1f38894\\\",\\\"systemUUID\\\":\\\"a57d09b0-0224-46eb-a3ac-a2b60dad2250\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:56Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:56 crc kubenswrapper[4716]: E1207 16:02:56.475430 4716 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 07 16:02:56 crc kubenswrapper[4716]: I1207 16:02:56.476935 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:56 crc kubenswrapper[4716]: I1207 16:02:56.476962 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:56 crc kubenswrapper[4716]: I1207 16:02:56.476971 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:56 crc kubenswrapper[4716]: I1207 16:02:56.476981 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:56 crc kubenswrapper[4716]: I1207 16:02:56.476991 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:56Z","lastTransitionTime":"2025-12-07T16:02:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:56 crc kubenswrapper[4716]: I1207 16:02:56.579307 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:56 crc kubenswrapper[4716]: I1207 16:02:56.579361 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:56 crc kubenswrapper[4716]: I1207 16:02:56.579380 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:56 crc kubenswrapper[4716]: I1207 16:02:56.579405 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:56 crc kubenswrapper[4716]: I1207 16:02:56.579423 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:56Z","lastTransitionTime":"2025-12-07T16:02:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:56 crc kubenswrapper[4716]: I1207 16:02:56.656919 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 16:02:56 crc kubenswrapper[4716]: E1207 16:02:56.657106 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 16:02:56 crc kubenswrapper[4716]: I1207 16:02:56.681711 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:56 crc kubenswrapper[4716]: I1207 16:02:56.681759 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:56 crc kubenswrapper[4716]: I1207 16:02:56.681774 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:56 crc kubenswrapper[4716]: I1207 16:02:56.681793 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:56 crc kubenswrapper[4716]: I1207 16:02:56.681809 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:56Z","lastTransitionTime":"2025-12-07T16:02:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:56 crc kubenswrapper[4716]: I1207 16:02:56.784811 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:56 crc kubenswrapper[4716]: I1207 16:02:56.785067 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:56 crc kubenswrapper[4716]: I1207 16:02:56.785098 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:56 crc kubenswrapper[4716]: I1207 16:02:56.785113 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:56 crc kubenswrapper[4716]: I1207 16:02:56.785122 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:56Z","lastTransitionTime":"2025-12-07T16:02:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:56 crc kubenswrapper[4716]: I1207 16:02:56.886896 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:56 crc kubenswrapper[4716]: I1207 16:02:56.886947 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:56 crc kubenswrapper[4716]: I1207 16:02:56.886956 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:56 crc kubenswrapper[4716]: I1207 16:02:56.886969 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:56 crc kubenswrapper[4716]: I1207 16:02:56.886978 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:56Z","lastTransitionTime":"2025-12-07T16:02:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:56 crc kubenswrapper[4716]: I1207 16:02:56.989809 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:56 crc kubenswrapper[4716]: I1207 16:02:56.989876 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:56 crc kubenswrapper[4716]: I1207 16:02:56.989901 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:56 crc kubenswrapper[4716]: I1207 16:02:56.989929 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:56 crc kubenswrapper[4716]: I1207 16:02:56.989951 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:56Z","lastTransitionTime":"2025-12-07T16:02:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:57 crc kubenswrapper[4716]: I1207 16:02:57.092329 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:57 crc kubenswrapper[4716]: I1207 16:02:57.092394 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:57 crc kubenswrapper[4716]: I1207 16:02:57.092412 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:57 crc kubenswrapper[4716]: I1207 16:02:57.092436 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:57 crc kubenswrapper[4716]: I1207 16:02:57.092454 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:57Z","lastTransitionTime":"2025-12-07T16:02:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:57 crc kubenswrapper[4716]: I1207 16:02:57.194658 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:57 crc kubenswrapper[4716]: I1207 16:02:57.194731 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:57 crc kubenswrapper[4716]: I1207 16:02:57.194759 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:57 crc kubenswrapper[4716]: I1207 16:02:57.194790 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:57 crc kubenswrapper[4716]: I1207 16:02:57.194814 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:57Z","lastTransitionTime":"2025-12-07T16:02:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:57 crc kubenswrapper[4716]: I1207 16:02:57.297536 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:57 crc kubenswrapper[4716]: I1207 16:02:57.297573 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:57 crc kubenswrapper[4716]: I1207 16:02:57.297585 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:57 crc kubenswrapper[4716]: I1207 16:02:57.297601 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:57 crc kubenswrapper[4716]: I1207 16:02:57.297612 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:57Z","lastTransitionTime":"2025-12-07T16:02:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:57 crc kubenswrapper[4716]: I1207 16:02:57.400039 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:57 crc kubenswrapper[4716]: I1207 16:02:57.400115 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:57 crc kubenswrapper[4716]: I1207 16:02:57.400126 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:57 crc kubenswrapper[4716]: I1207 16:02:57.400145 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:57 crc kubenswrapper[4716]: I1207 16:02:57.400155 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:57Z","lastTransitionTime":"2025-12-07T16:02:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:57 crc kubenswrapper[4716]: I1207 16:02:57.505909 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:57 crc kubenswrapper[4716]: I1207 16:02:57.506006 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:57 crc kubenswrapper[4716]: I1207 16:02:57.506024 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:57 crc kubenswrapper[4716]: I1207 16:02:57.506054 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:57 crc kubenswrapper[4716]: I1207 16:02:57.506103 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:57Z","lastTransitionTime":"2025-12-07T16:02:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:57 crc kubenswrapper[4716]: I1207 16:02:57.609481 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:57 crc kubenswrapper[4716]: I1207 16:02:57.609543 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:57 crc kubenswrapper[4716]: I1207 16:02:57.609561 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:57 crc kubenswrapper[4716]: I1207 16:02:57.609586 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:57 crc kubenswrapper[4716]: I1207 16:02:57.609605 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:57Z","lastTransitionTime":"2025-12-07T16:02:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:57 crc kubenswrapper[4716]: I1207 16:02:57.656914 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 16:02:57 crc kubenswrapper[4716]: I1207 16:02:57.656974 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nx2lh" Dec 07 16:02:57 crc kubenswrapper[4716]: E1207 16:02:57.657067 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 16:02:57 crc kubenswrapper[4716]: I1207 16:02:57.657161 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:02:57 crc kubenswrapper[4716]: E1207 16:02:57.657239 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nx2lh" podUID="04b37f6c-457b-4f9a-8733-e88cb9644483" Dec 07 16:02:57 crc kubenswrapper[4716]: E1207 16:02:57.657319 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 16:02:57 crc kubenswrapper[4716]: I1207 16:02:57.673743 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdee9957-48b9-458f-ba71-e0cd39f80596\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73f85248b4d68a9beecdacb3a01f433e97a0ffea2247ce92a41ef96855509377\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6276df652efe6ca935e745f24ef20f87a16465475139230c63b9999234a4d12a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3b1dd202586ee755ca89fbe12b30bbb12f95a62514f6e4d93cbbdbe55181e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd318647816e4a27849666b0ec4fc2ca4905295661c557dc381e1144f22cf284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d31de0326a05216fa10442bdb47151c6e073c97e9a16c34e793028e5e253756\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"le observer\\\\nW1207 16:02:34.645464 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1207 16:02:34.645587 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 16:02:34.646501 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2082283479/tls.crt::/tmp/serving-cert-2082283479/tls.key\\\\\\\"\\\\nI1207 16:02:34.972041 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 16:02:34.973702 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 16:02:34.973717 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 16:02:34.973738 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 16:02:34.973743 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 16:02:34.977185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1207 16:02:34.977204 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1207 16:02:34.977209 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977217 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 16:02:34.977225 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 16:02:34.977228 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 16:02:34.977232 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1207 16:02:34.979033 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27bd4a54f920b2725987c1c7fd2d9fdf16cd08738c246c333a4e9164a3a582a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:57Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:57 crc kubenswrapper[4716]: I1207 16:02:57.687048 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:57Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:57 crc kubenswrapper[4716]: I1207 16:02:57.697674 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2jjld" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dee6f5c-861c-4819-9309-7791574c7b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51c1d4159ab11087839a2c2d192566fb9322447c26df560bf09517c906ececb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmdns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2jjld\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:57Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:57 crc kubenswrapper[4716]: I1207 16:02:57.711517 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:57 crc kubenswrapper[4716]: I1207 16:02:57.711560 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:57 crc kubenswrapper[4716]: I1207 16:02:57.711569 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:57 crc kubenswrapper[4716]: I1207 16:02:57.711585 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:57 crc kubenswrapper[4716]: I1207 16:02:57.711594 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:57Z","lastTransitionTime":"2025-12-07T16:02:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:57 crc kubenswrapper[4716]: I1207 16:02:57.733480 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4wr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0baf8575-26cf-4a72-8a6d-ae94c078eaad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0ca2b3cd5a11033afd219b0125be7ab5acd750935d7eb89a470eff3d93a0838\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p2trd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4wr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:57Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:57 crc kubenswrapper[4716]: I1207 16:02:57.766396 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-m5jl5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"824ee59f-1474-4994-a70e-462878521514\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a15fe50cc75515704779646cc19aa50004f0db9aa26fe2535110df749bbfb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhwzn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-m5jl5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:57Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:57 crc kubenswrapper[4716]: I1207 16:02:57.787163 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef3335ae3875a75ab64685fc7e3590177280d651e8f8292c6f5ea6f472e55789\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:57Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:57 crc kubenswrapper[4716]: I1207 16:02:57.801898 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:57Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:57 crc kubenswrapper[4716]: I1207 16:02:57.815943 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:57 crc kubenswrapper[4716]: I1207 16:02:57.815990 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:57 crc kubenswrapper[4716]: I1207 16:02:57.816003 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:57 crc kubenswrapper[4716]: I1207 16:02:57.816024 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:57 crc kubenswrapper[4716]: I1207 16:02:57.816037 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:57Z","lastTransitionTime":"2025-12-07T16:02:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:57 crc kubenswrapper[4716]: I1207 16:02:57.816755 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f95ef61779507288a03bda8eddf745e105f705998a117a30789675bfa452d88f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b955b23163e7c8c458a3bbe2f681633e6a663ce76748febcc0b3a3c37f3ca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:57Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:57 crc kubenswrapper[4716]: I1207 16:02:57.837107 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7961cc9a-734e-42c2-b1ff-7075f3f20091\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84a36b76f20618c7885d611ecad9db29bac5101c27572be31d3acbb90c581291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0a0eca3c2087a457c0e448e649345ec4595a9c8a019aed2cddc5f30c77871e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://404426a171b8676eddd4a00996c76afab7e269643fa9608f1e1b6bdba385071d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fd3c2941782921f99a6cdc3de4e892234548bde527fec15a8ac92e513566e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e23abfdca314ed43d594333c9f9e401cc8ef7dad39b8ebac84a836b06ee5c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09b38cd4b389ea477c802a842ee4f2297a64d64684624117163e819539c257c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09b38cd4b389ea477c802a842ee4f2297a64d64684624117163e819539c257c6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c00f7a8e024e0e0dfd7b73e16d1d5437602e79ea95b73c21f24b209edef60e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c00f7a8e024e0e0dfd7b73e16d1d5437602e79ea95b73c21f24b209edef60e98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3f6392d426fa2cd788ec3a5c3bff4ac39483a54dfb2624a1c4370522413a350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3f6392d426fa2cd788ec3a5c3bff4ac39483a54dfb2624a1c4370522413a350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:57Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:57 crc kubenswrapper[4716]: I1207 16:02:57.850120 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3067de7c-c11c-4be1-bd94-d6572c266795\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d14d018fa044e9e80fe81e7e38df935992d14c773ed2b2193fd5cb873e5d0947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a9e9cf048220ca14c996847e9ced9f2367b9331dc6c2836dd023658855cd79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6d3c911874095b3c7cb062d03070e5090a7bb2456cb20cb4ce078f502515cda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ac6ce7701f1c6aa1ea8fcf7a3035e3b0eaa28ae3cb6afc442372bd9117bdde4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:57Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:57 crc kubenswrapper[4716]: I1207 16:02:57.865488 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4742ea84b3a2c9461d9b900beee7354a2ded24c39ad23e007cbf270ae12377c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:57Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:57 crc kubenswrapper[4716]: I1207 16:02:57.876841 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:57Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:57 crc kubenswrapper[4716]: I1207 16:02:57.888819 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c15b59eb-565d-4556-a4ce-75afdf159dc8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://530a2935826940767ea33fb4f37fa11f7dee4950d867fbcd6af72836d80f7db3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jks8z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ccaf6c1e6a978eb978a1c207b6c4c00b23737f78b743f966579aca781e22e0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jks8z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zcxxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:57Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:57 crc kubenswrapper[4716]: I1207 16:02:57.904985 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8ps9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98015cea-d753-4e7a-b2b5-5ae7733a81e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d68509b673eecd37725cdcc48e3d0612c3b9c7601433f2b72c21d662140daa59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f84f52a073437543721010dc440339654ffb5e04eeb6cc7afbcc62f00f64c8e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f84f52a073437543721010dc440339654ffb5e04eeb6cc7afbcc62f00f64c8e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f6fe1cb25218cb3ac20488dc47efa4dc6a26a43dbc8306b5ee38cbe1a186c67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f6fe1cb25218cb3ac20488dc47efa4dc6a26a43dbc8306b5ee38cbe1a186c67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14b748a8247f6863e8be61e90057a06626dfd3cd4f01bd3a11629f700f56378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f14b748a8247f6863e8be61e90057a06626dfd3cd4f01bd3a11629f700f56378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ec379c0d573cb9b5ffcfc5ca7c653ba60e2fbdd7e3d37836102e26618ce43e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ec379c0d573cb9b5ffcfc5ca7c653ba60e2fbdd7e3d37836102e26618ce43e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c75bf2a6703cf99865d4ed5fa000e1e3b1330d4946176cf585fec1a2f226df0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c75bf2a6703cf99865d4ed5fa000e1e3b1330d4946176cf585fec1a2f226df0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de45812c9a5a5bfe6b0146828d03fab28362b03d04ef861c150b9ffbf181d345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de45812c9a5a5bfe6b0146828d03fab28362b03d04ef861c150b9ffbf181d345\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8ps9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:57Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:57 crc kubenswrapper[4716]: I1207 16:02:57.910309 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/04b37f6c-457b-4f9a-8733-e88cb9644483-metrics-certs\") pod \"network-metrics-daemon-nx2lh\" (UID: \"04b37f6c-457b-4f9a-8733-e88cb9644483\") " pod="openshift-multus/network-metrics-daemon-nx2lh" Dec 07 16:02:57 crc kubenswrapper[4716]: E1207 16:02:57.910407 4716 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 07 16:02:57 crc kubenswrapper[4716]: E1207 16:02:57.910455 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/04b37f6c-457b-4f9a-8733-e88cb9644483-metrics-certs podName:04b37f6c-457b-4f9a-8733-e88cb9644483 nodeName:}" failed. No retries permitted until 2025-12-07 16:03:01.9104422 +0000 UTC m=+44.600727112 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/04b37f6c-457b-4f9a-8733-e88cb9644483-metrics-certs") pod "network-metrics-daemon-nx2lh" (UID: "04b37f6c-457b-4f9a-8733-e88cb9644483") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 07 16:02:57 crc kubenswrapper[4716]: I1207 16:02:57.918642 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:57 crc kubenswrapper[4716]: I1207 16:02:57.918669 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:57 crc kubenswrapper[4716]: I1207 16:02:57.918679 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:57 crc kubenswrapper[4716]: I1207 16:02:57.918693 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:57 crc kubenswrapper[4716]: I1207 16:02:57.918702 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:57Z","lastTransitionTime":"2025-12-07T16:02:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:57 crc kubenswrapper[4716]: I1207 16:02:57.921789 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c54d618a-dfec-4f2a-a288-f5188c8a128a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60324dc7d701bc064e14ae4f474bce10f6ee4ff3c3c942eab749e7166082cf9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ed4f4d3564674bf5cb4e7a76dc0fd95e203ecc8950798e59a1fb6c9ab4a20f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://def9ac840fb8ecf2778bed88ea980eddca7c5b2aba4bd1afb967d62fec3f7bae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b73be664f5412c98b89d971a2cac37c494151c293fdb44a98b390978658a6643\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a847d04e22e98091c0037a89054b9236228b701b8c7ba31940a56df0f37b95c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcae78a6133d97f6a32dd7421ccec4333b970a237404e056a5cc601f8eefdc8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8baee0f3d1e686bc2c17b0777f4eacc31c7f1fe44720e68f94612e6c806bc8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8baee0f3d1e686bc2c17b0777f4eacc31c7f1fe44720e68f94612e6c806bc8b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-07T16:02:52Z\\\",\\\"message\\\":\\\"r occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:51Z is after 2025-08-24T17:21:41Z]\\\\nI1207 16:02:51.943802 6191 services_controller.go:444] Built service openshift-marketplace/redhat-marketplace LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1207 16:02:51.943764 6191 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-route-controller-manager/route-controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"18746a4d-8a63-458a-b7e3-8fb89ff95fc0\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-route-controller-manager/route-controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-route-controller-manager/route-controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-qvc7p_openshift-ovn-kubernetes(c54d618a-dfec-4f2a-a288-f5188c8a128a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcfdffe822635b0ae13b253e6133449fe00461b2792be43d913c4a77219b8255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9bfe2ffce7a7b7c320474117734fc05246cc2b4fd744833e896e8e7c83f5e4d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9bfe2ffce7a7b7c320474117734fc05246cc2b4fd744833e896e8e7c83f5e4d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qvc7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:57Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:57 crc kubenswrapper[4716]: I1207 16:02:57.932740 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mx85f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"118dd9ba-437c-4e5c-b5ff-4239ae6d4a63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bcd6c8deb184d0c713c5e7b7cc82fa38d3cd498517911ba7a0ec3cb88a3437a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gg5nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ffc069d2b21fa80871b1b97f2c72c37880d9a309cbf8b7daf7d240f2c332a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gg5nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mx85f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:57Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:57 crc kubenswrapper[4716]: I1207 16:02:57.942458 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nx2lh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04b37f6c-457b-4f9a-8733-e88cb9644483\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv6td\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv6td\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nx2lh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:57Z is after 2025-08-24T17:21:41Z" Dec 07 16:02:58 crc kubenswrapper[4716]: I1207 16:02:58.021317 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:58 crc kubenswrapper[4716]: I1207 16:02:58.021352 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:58 crc kubenswrapper[4716]: I1207 16:02:58.021362 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:58 crc kubenswrapper[4716]: I1207 16:02:58.021376 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:58 crc kubenswrapper[4716]: I1207 16:02:58.021385 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:58Z","lastTransitionTime":"2025-12-07T16:02:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:58 crc kubenswrapper[4716]: I1207 16:02:58.123543 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:58 crc kubenswrapper[4716]: I1207 16:02:58.123575 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:58 crc kubenswrapper[4716]: I1207 16:02:58.123585 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:58 crc kubenswrapper[4716]: I1207 16:02:58.123600 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:58 crc kubenswrapper[4716]: I1207 16:02:58.123630 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:58Z","lastTransitionTime":"2025-12-07T16:02:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:58 crc kubenswrapper[4716]: I1207 16:02:58.225282 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:58 crc kubenswrapper[4716]: I1207 16:02:58.225314 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:58 crc kubenswrapper[4716]: I1207 16:02:58.225322 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:58 crc kubenswrapper[4716]: I1207 16:02:58.225336 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:58 crc kubenswrapper[4716]: I1207 16:02:58.225345 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:58Z","lastTransitionTime":"2025-12-07T16:02:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:58 crc kubenswrapper[4716]: I1207 16:02:58.327412 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:58 crc kubenswrapper[4716]: I1207 16:02:58.327482 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:58 crc kubenswrapper[4716]: I1207 16:02:58.327506 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:58 crc kubenswrapper[4716]: I1207 16:02:58.327535 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:58 crc kubenswrapper[4716]: I1207 16:02:58.327557 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:58Z","lastTransitionTime":"2025-12-07T16:02:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:58 crc kubenswrapper[4716]: I1207 16:02:58.430595 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:58 crc kubenswrapper[4716]: I1207 16:02:58.430668 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:58 crc kubenswrapper[4716]: I1207 16:02:58.430690 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:58 crc kubenswrapper[4716]: I1207 16:02:58.430752 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:58 crc kubenswrapper[4716]: I1207 16:02:58.430777 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:58Z","lastTransitionTime":"2025-12-07T16:02:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:58 crc kubenswrapper[4716]: I1207 16:02:58.533683 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:58 crc kubenswrapper[4716]: I1207 16:02:58.533742 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:58 crc kubenswrapper[4716]: I1207 16:02:58.533766 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:58 crc kubenswrapper[4716]: I1207 16:02:58.533792 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:58 crc kubenswrapper[4716]: I1207 16:02:58.533811 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:58Z","lastTransitionTime":"2025-12-07T16:02:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:58 crc kubenswrapper[4716]: I1207 16:02:58.636881 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:58 crc kubenswrapper[4716]: I1207 16:02:58.636963 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:58 crc kubenswrapper[4716]: I1207 16:02:58.636981 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:58 crc kubenswrapper[4716]: I1207 16:02:58.637005 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:58 crc kubenswrapper[4716]: I1207 16:02:58.637021 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:58Z","lastTransitionTime":"2025-12-07T16:02:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:58 crc kubenswrapper[4716]: I1207 16:02:58.657183 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 16:02:58 crc kubenswrapper[4716]: E1207 16:02:58.657319 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 16:02:58 crc kubenswrapper[4716]: I1207 16:02:58.739987 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:58 crc kubenswrapper[4716]: I1207 16:02:58.740054 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:58 crc kubenswrapper[4716]: I1207 16:02:58.740120 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:58 crc kubenswrapper[4716]: I1207 16:02:58.740155 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:58 crc kubenswrapper[4716]: I1207 16:02:58.740179 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:58Z","lastTransitionTime":"2025-12-07T16:02:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:58 crc kubenswrapper[4716]: I1207 16:02:58.842808 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:58 crc kubenswrapper[4716]: I1207 16:02:58.843044 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:58 crc kubenswrapper[4716]: I1207 16:02:58.843158 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:58 crc kubenswrapper[4716]: I1207 16:02:58.843229 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:58 crc kubenswrapper[4716]: I1207 16:02:58.843291 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:58Z","lastTransitionTime":"2025-12-07T16:02:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:58 crc kubenswrapper[4716]: I1207 16:02:58.945921 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:58 crc kubenswrapper[4716]: I1207 16:02:58.945958 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:58 crc kubenswrapper[4716]: I1207 16:02:58.945970 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:58 crc kubenswrapper[4716]: I1207 16:02:58.945984 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:58 crc kubenswrapper[4716]: I1207 16:02:58.945997 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:58Z","lastTransitionTime":"2025-12-07T16:02:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:59 crc kubenswrapper[4716]: I1207 16:02:59.048967 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:59 crc kubenswrapper[4716]: I1207 16:02:59.049041 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:59 crc kubenswrapper[4716]: I1207 16:02:59.049061 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:59 crc kubenswrapper[4716]: I1207 16:02:59.049130 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:59 crc kubenswrapper[4716]: I1207 16:02:59.049154 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:59Z","lastTransitionTime":"2025-12-07T16:02:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:59 crc kubenswrapper[4716]: I1207 16:02:59.152604 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:59 crc kubenswrapper[4716]: I1207 16:02:59.152674 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:59 crc kubenswrapper[4716]: I1207 16:02:59.152695 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:59 crc kubenswrapper[4716]: I1207 16:02:59.152722 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:59 crc kubenswrapper[4716]: I1207 16:02:59.152742 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:59Z","lastTransitionTime":"2025-12-07T16:02:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:59 crc kubenswrapper[4716]: I1207 16:02:59.255615 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:59 crc kubenswrapper[4716]: I1207 16:02:59.255650 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:59 crc kubenswrapper[4716]: I1207 16:02:59.255659 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:59 crc kubenswrapper[4716]: I1207 16:02:59.255672 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:59 crc kubenswrapper[4716]: I1207 16:02:59.255680 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:59Z","lastTransitionTime":"2025-12-07T16:02:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:59 crc kubenswrapper[4716]: I1207 16:02:59.357814 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:59 crc kubenswrapper[4716]: I1207 16:02:59.357853 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:59 crc kubenswrapper[4716]: I1207 16:02:59.357863 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:59 crc kubenswrapper[4716]: I1207 16:02:59.357878 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:59 crc kubenswrapper[4716]: I1207 16:02:59.357889 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:59Z","lastTransitionTime":"2025-12-07T16:02:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:59 crc kubenswrapper[4716]: I1207 16:02:59.461172 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:59 crc kubenswrapper[4716]: I1207 16:02:59.461227 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:59 crc kubenswrapper[4716]: I1207 16:02:59.461244 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:59 crc kubenswrapper[4716]: I1207 16:02:59.461267 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:59 crc kubenswrapper[4716]: I1207 16:02:59.461283 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:59Z","lastTransitionTime":"2025-12-07T16:02:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:59 crc kubenswrapper[4716]: I1207 16:02:59.563627 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:59 crc kubenswrapper[4716]: I1207 16:02:59.563662 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:59 crc kubenswrapper[4716]: I1207 16:02:59.563672 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:59 crc kubenswrapper[4716]: I1207 16:02:59.563684 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:59 crc kubenswrapper[4716]: I1207 16:02:59.563693 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:59Z","lastTransitionTime":"2025-12-07T16:02:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:59 crc kubenswrapper[4716]: I1207 16:02:59.657399 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nx2lh" Dec 07 16:02:59 crc kubenswrapper[4716]: I1207 16:02:59.657424 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:02:59 crc kubenswrapper[4716]: I1207 16:02:59.657480 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 16:02:59 crc kubenswrapper[4716]: E1207 16:02:59.657530 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 16:02:59 crc kubenswrapper[4716]: E1207 16:02:59.657650 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 16:02:59 crc kubenswrapper[4716]: E1207 16:02:59.657772 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nx2lh" podUID="04b37f6c-457b-4f9a-8733-e88cb9644483" Dec 07 16:02:59 crc kubenswrapper[4716]: I1207 16:02:59.665588 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:59 crc kubenswrapper[4716]: I1207 16:02:59.665618 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:59 crc kubenswrapper[4716]: I1207 16:02:59.665628 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:59 crc kubenswrapper[4716]: I1207 16:02:59.665641 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:59 crc kubenswrapper[4716]: I1207 16:02:59.665653 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:59Z","lastTransitionTime":"2025-12-07T16:02:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:59 crc kubenswrapper[4716]: I1207 16:02:59.768633 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:59 crc kubenswrapper[4716]: I1207 16:02:59.768696 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:59 crc kubenswrapper[4716]: I1207 16:02:59.768715 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:59 crc kubenswrapper[4716]: I1207 16:02:59.768738 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:59 crc kubenswrapper[4716]: I1207 16:02:59.768755 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:59Z","lastTransitionTime":"2025-12-07T16:02:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:59 crc kubenswrapper[4716]: I1207 16:02:59.871160 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:59 crc kubenswrapper[4716]: I1207 16:02:59.871210 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:59 crc kubenswrapper[4716]: I1207 16:02:59.871219 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:59 crc kubenswrapper[4716]: I1207 16:02:59.871233 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:59 crc kubenswrapper[4716]: I1207 16:02:59.871242 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:59Z","lastTransitionTime":"2025-12-07T16:02:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:02:59 crc kubenswrapper[4716]: I1207 16:02:59.973574 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:02:59 crc kubenswrapper[4716]: I1207 16:02:59.973629 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:02:59 crc kubenswrapper[4716]: I1207 16:02:59.973638 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:02:59 crc kubenswrapper[4716]: I1207 16:02:59.973651 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:02:59 crc kubenswrapper[4716]: I1207 16:02:59.973662 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:02:59Z","lastTransitionTime":"2025-12-07T16:02:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:00 crc kubenswrapper[4716]: I1207 16:03:00.076337 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:00 crc kubenswrapper[4716]: I1207 16:03:00.076405 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:00 crc kubenswrapper[4716]: I1207 16:03:00.076422 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:00 crc kubenswrapper[4716]: I1207 16:03:00.076448 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:00 crc kubenswrapper[4716]: I1207 16:03:00.076466 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:00Z","lastTransitionTime":"2025-12-07T16:03:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:00 crc kubenswrapper[4716]: I1207 16:03:00.180796 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:00 crc kubenswrapper[4716]: I1207 16:03:00.180839 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:00 crc kubenswrapper[4716]: I1207 16:03:00.180848 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:00 crc kubenswrapper[4716]: I1207 16:03:00.180862 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:00 crc kubenswrapper[4716]: I1207 16:03:00.180870 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:00Z","lastTransitionTime":"2025-12-07T16:03:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:00 crc kubenswrapper[4716]: I1207 16:03:00.283784 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:00 crc kubenswrapper[4716]: I1207 16:03:00.283835 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:00 crc kubenswrapper[4716]: I1207 16:03:00.283845 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:00 crc kubenswrapper[4716]: I1207 16:03:00.283859 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:00 crc kubenswrapper[4716]: I1207 16:03:00.283868 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:00Z","lastTransitionTime":"2025-12-07T16:03:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:00 crc kubenswrapper[4716]: I1207 16:03:00.386376 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:00 crc kubenswrapper[4716]: I1207 16:03:00.386410 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:00 crc kubenswrapper[4716]: I1207 16:03:00.386419 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:00 crc kubenswrapper[4716]: I1207 16:03:00.386431 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:00 crc kubenswrapper[4716]: I1207 16:03:00.386440 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:00Z","lastTransitionTime":"2025-12-07T16:03:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:00 crc kubenswrapper[4716]: I1207 16:03:00.489483 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:00 crc kubenswrapper[4716]: I1207 16:03:00.489540 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:00 crc kubenswrapper[4716]: I1207 16:03:00.489557 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:00 crc kubenswrapper[4716]: I1207 16:03:00.489583 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:00 crc kubenswrapper[4716]: I1207 16:03:00.489601 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:00Z","lastTransitionTime":"2025-12-07T16:03:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:00 crc kubenswrapper[4716]: I1207 16:03:00.592642 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:00 crc kubenswrapper[4716]: I1207 16:03:00.592694 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:00 crc kubenswrapper[4716]: I1207 16:03:00.592707 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:00 crc kubenswrapper[4716]: I1207 16:03:00.592726 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:00 crc kubenswrapper[4716]: I1207 16:03:00.592740 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:00Z","lastTransitionTime":"2025-12-07T16:03:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:00 crc kubenswrapper[4716]: I1207 16:03:00.657056 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 16:03:00 crc kubenswrapper[4716]: E1207 16:03:00.657313 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 16:03:00 crc kubenswrapper[4716]: I1207 16:03:00.694560 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:00 crc kubenswrapper[4716]: I1207 16:03:00.694614 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:00 crc kubenswrapper[4716]: I1207 16:03:00.694629 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:00 crc kubenswrapper[4716]: I1207 16:03:00.694652 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:00 crc kubenswrapper[4716]: I1207 16:03:00.694667 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:00Z","lastTransitionTime":"2025-12-07T16:03:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:00 crc kubenswrapper[4716]: I1207 16:03:00.797151 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:00 crc kubenswrapper[4716]: I1207 16:03:00.797224 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:00 crc kubenswrapper[4716]: I1207 16:03:00.797249 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:00 crc kubenswrapper[4716]: I1207 16:03:00.797273 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:00 crc kubenswrapper[4716]: I1207 16:03:00.797290 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:00Z","lastTransitionTime":"2025-12-07T16:03:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:00 crc kubenswrapper[4716]: I1207 16:03:00.899480 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:00 crc kubenswrapper[4716]: I1207 16:03:00.899539 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:00 crc kubenswrapper[4716]: I1207 16:03:00.899556 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:00 crc kubenswrapper[4716]: I1207 16:03:00.899576 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:00 crc kubenswrapper[4716]: I1207 16:03:00.899590 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:00Z","lastTransitionTime":"2025-12-07T16:03:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:01 crc kubenswrapper[4716]: I1207 16:03:01.001529 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:01 crc kubenswrapper[4716]: I1207 16:03:01.001561 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:01 crc kubenswrapper[4716]: I1207 16:03:01.001569 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:01 crc kubenswrapper[4716]: I1207 16:03:01.001581 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:01 crc kubenswrapper[4716]: I1207 16:03:01.001591 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:01Z","lastTransitionTime":"2025-12-07T16:03:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:01 crc kubenswrapper[4716]: I1207 16:03:01.103658 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:01 crc kubenswrapper[4716]: I1207 16:03:01.103697 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:01 crc kubenswrapper[4716]: I1207 16:03:01.103705 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:01 crc kubenswrapper[4716]: I1207 16:03:01.103752 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:01 crc kubenswrapper[4716]: I1207 16:03:01.103763 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:01Z","lastTransitionTime":"2025-12-07T16:03:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:01 crc kubenswrapper[4716]: I1207 16:03:01.206263 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:01 crc kubenswrapper[4716]: I1207 16:03:01.206307 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:01 crc kubenswrapper[4716]: I1207 16:03:01.206323 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:01 crc kubenswrapper[4716]: I1207 16:03:01.206342 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:01 crc kubenswrapper[4716]: I1207 16:03:01.206356 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:01Z","lastTransitionTime":"2025-12-07T16:03:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:01 crc kubenswrapper[4716]: I1207 16:03:01.308512 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:01 crc kubenswrapper[4716]: I1207 16:03:01.308552 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:01 crc kubenswrapper[4716]: I1207 16:03:01.308563 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:01 crc kubenswrapper[4716]: I1207 16:03:01.308582 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:01 crc kubenswrapper[4716]: I1207 16:03:01.308594 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:01Z","lastTransitionTime":"2025-12-07T16:03:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:01 crc kubenswrapper[4716]: I1207 16:03:01.411108 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:01 crc kubenswrapper[4716]: I1207 16:03:01.411146 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:01 crc kubenswrapper[4716]: I1207 16:03:01.411156 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:01 crc kubenswrapper[4716]: I1207 16:03:01.411172 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:01 crc kubenswrapper[4716]: I1207 16:03:01.411184 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:01Z","lastTransitionTime":"2025-12-07T16:03:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:01 crc kubenswrapper[4716]: I1207 16:03:01.513579 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:01 crc kubenswrapper[4716]: I1207 16:03:01.513674 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:01 crc kubenswrapper[4716]: I1207 16:03:01.513705 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:01 crc kubenswrapper[4716]: I1207 16:03:01.513737 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:01 crc kubenswrapper[4716]: I1207 16:03:01.513764 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:01Z","lastTransitionTime":"2025-12-07T16:03:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:01 crc kubenswrapper[4716]: I1207 16:03:01.616472 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:01 crc kubenswrapper[4716]: I1207 16:03:01.616519 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:01 crc kubenswrapper[4716]: I1207 16:03:01.616528 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:01 crc kubenswrapper[4716]: I1207 16:03:01.616542 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:01 crc kubenswrapper[4716]: I1207 16:03:01.616552 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:01Z","lastTransitionTime":"2025-12-07T16:03:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:01 crc kubenswrapper[4716]: I1207 16:03:01.657400 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nx2lh" Dec 07 16:03:01 crc kubenswrapper[4716]: I1207 16:03:01.657421 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:03:01 crc kubenswrapper[4716]: I1207 16:03:01.657438 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 16:03:01 crc kubenswrapper[4716]: E1207 16:03:01.657522 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nx2lh" podUID="04b37f6c-457b-4f9a-8733-e88cb9644483" Dec 07 16:03:01 crc kubenswrapper[4716]: E1207 16:03:01.657626 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 16:03:01 crc kubenswrapper[4716]: E1207 16:03:01.657705 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 16:03:01 crc kubenswrapper[4716]: I1207 16:03:01.719291 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:01 crc kubenswrapper[4716]: I1207 16:03:01.719327 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:01 crc kubenswrapper[4716]: I1207 16:03:01.719336 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:01 crc kubenswrapper[4716]: I1207 16:03:01.719350 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:01 crc kubenswrapper[4716]: I1207 16:03:01.719359 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:01Z","lastTransitionTime":"2025-12-07T16:03:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:01 crc kubenswrapper[4716]: I1207 16:03:01.821121 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:01 crc kubenswrapper[4716]: I1207 16:03:01.821158 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:01 crc kubenswrapper[4716]: I1207 16:03:01.821168 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:01 crc kubenswrapper[4716]: I1207 16:03:01.821195 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:01 crc kubenswrapper[4716]: I1207 16:03:01.821209 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:01Z","lastTransitionTime":"2025-12-07T16:03:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:01 crc kubenswrapper[4716]: I1207 16:03:01.923629 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:01 crc kubenswrapper[4716]: I1207 16:03:01.923663 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:01 crc kubenswrapper[4716]: I1207 16:03:01.923674 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:01 crc kubenswrapper[4716]: I1207 16:03:01.923690 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:01 crc kubenswrapper[4716]: I1207 16:03:01.923702 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:01Z","lastTransitionTime":"2025-12-07T16:03:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:01 crc kubenswrapper[4716]: I1207 16:03:01.951623 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/04b37f6c-457b-4f9a-8733-e88cb9644483-metrics-certs\") pod \"network-metrics-daemon-nx2lh\" (UID: \"04b37f6c-457b-4f9a-8733-e88cb9644483\") " pod="openshift-multus/network-metrics-daemon-nx2lh" Dec 07 16:03:01 crc kubenswrapper[4716]: E1207 16:03:01.951763 4716 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 07 16:03:01 crc kubenswrapper[4716]: E1207 16:03:01.951818 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/04b37f6c-457b-4f9a-8733-e88cb9644483-metrics-certs podName:04b37f6c-457b-4f9a-8733-e88cb9644483 nodeName:}" failed. No retries permitted until 2025-12-07 16:03:09.951801623 +0000 UTC m=+52.642086535 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/04b37f6c-457b-4f9a-8733-e88cb9644483-metrics-certs") pod "network-metrics-daemon-nx2lh" (UID: "04b37f6c-457b-4f9a-8733-e88cb9644483") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 07 16:03:02 crc kubenswrapper[4716]: I1207 16:03:02.026250 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:02 crc kubenswrapper[4716]: I1207 16:03:02.026285 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:02 crc kubenswrapper[4716]: I1207 16:03:02.026293 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:02 crc kubenswrapper[4716]: I1207 16:03:02.026307 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:02 crc kubenswrapper[4716]: I1207 16:03:02.026316 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:02Z","lastTransitionTime":"2025-12-07T16:03:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:02 crc kubenswrapper[4716]: I1207 16:03:02.128500 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:02 crc kubenswrapper[4716]: I1207 16:03:02.128550 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:02 crc kubenswrapper[4716]: I1207 16:03:02.128559 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:02 crc kubenswrapper[4716]: I1207 16:03:02.128575 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:02 crc kubenswrapper[4716]: I1207 16:03:02.128584 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:02Z","lastTransitionTime":"2025-12-07T16:03:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:02 crc kubenswrapper[4716]: I1207 16:03:02.234311 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:02 crc kubenswrapper[4716]: I1207 16:03:02.234369 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:02 crc kubenswrapper[4716]: I1207 16:03:02.234383 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:02 crc kubenswrapper[4716]: I1207 16:03:02.234404 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:02 crc kubenswrapper[4716]: I1207 16:03:02.234419 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:02Z","lastTransitionTime":"2025-12-07T16:03:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:02 crc kubenswrapper[4716]: I1207 16:03:02.337196 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:02 crc kubenswrapper[4716]: I1207 16:03:02.337244 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:02 crc kubenswrapper[4716]: I1207 16:03:02.337256 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:02 crc kubenswrapper[4716]: I1207 16:03:02.337273 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:02 crc kubenswrapper[4716]: I1207 16:03:02.337287 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:02Z","lastTransitionTime":"2025-12-07T16:03:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:02 crc kubenswrapper[4716]: I1207 16:03:02.439277 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:02 crc kubenswrapper[4716]: I1207 16:03:02.439317 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:02 crc kubenswrapper[4716]: I1207 16:03:02.439328 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:02 crc kubenswrapper[4716]: I1207 16:03:02.439343 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:02 crc kubenswrapper[4716]: I1207 16:03:02.439354 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:02Z","lastTransitionTime":"2025-12-07T16:03:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:02 crc kubenswrapper[4716]: I1207 16:03:02.543788 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:02 crc kubenswrapper[4716]: I1207 16:03:02.543832 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:02 crc kubenswrapper[4716]: I1207 16:03:02.543842 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:02 crc kubenswrapper[4716]: I1207 16:03:02.543890 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:02 crc kubenswrapper[4716]: I1207 16:03:02.543916 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:02Z","lastTransitionTime":"2025-12-07T16:03:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:02 crc kubenswrapper[4716]: I1207 16:03:02.646208 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:02 crc kubenswrapper[4716]: I1207 16:03:02.646269 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:02 crc kubenswrapper[4716]: I1207 16:03:02.646285 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:02 crc kubenswrapper[4716]: I1207 16:03:02.646307 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:02 crc kubenswrapper[4716]: I1207 16:03:02.646322 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:02Z","lastTransitionTime":"2025-12-07T16:03:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:02 crc kubenswrapper[4716]: I1207 16:03:02.656706 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 16:03:02 crc kubenswrapper[4716]: E1207 16:03:02.656910 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 16:03:02 crc kubenswrapper[4716]: I1207 16:03:02.748322 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:02 crc kubenswrapper[4716]: I1207 16:03:02.748363 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:02 crc kubenswrapper[4716]: I1207 16:03:02.748374 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:02 crc kubenswrapper[4716]: I1207 16:03:02.748390 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:02 crc kubenswrapper[4716]: I1207 16:03:02.748401 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:02Z","lastTransitionTime":"2025-12-07T16:03:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:02 crc kubenswrapper[4716]: I1207 16:03:02.850191 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:02 crc kubenswrapper[4716]: I1207 16:03:02.850254 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:02 crc kubenswrapper[4716]: I1207 16:03:02.850273 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:02 crc kubenswrapper[4716]: I1207 16:03:02.850297 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:02 crc kubenswrapper[4716]: I1207 16:03:02.850315 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:02Z","lastTransitionTime":"2025-12-07T16:03:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:02 crc kubenswrapper[4716]: I1207 16:03:02.953918 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:02 crc kubenswrapper[4716]: I1207 16:03:02.953970 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:02 crc kubenswrapper[4716]: I1207 16:03:02.953982 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:02 crc kubenswrapper[4716]: I1207 16:03:02.953999 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:02 crc kubenswrapper[4716]: I1207 16:03:02.954010 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:02Z","lastTransitionTime":"2025-12-07T16:03:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:03 crc kubenswrapper[4716]: I1207 16:03:03.056802 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:03 crc kubenswrapper[4716]: I1207 16:03:03.056851 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:03 crc kubenswrapper[4716]: I1207 16:03:03.056863 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:03 crc kubenswrapper[4716]: I1207 16:03:03.056881 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:03 crc kubenswrapper[4716]: I1207 16:03:03.056893 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:03Z","lastTransitionTime":"2025-12-07T16:03:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:03 crc kubenswrapper[4716]: I1207 16:03:03.159767 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:03 crc kubenswrapper[4716]: I1207 16:03:03.159809 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:03 crc kubenswrapper[4716]: I1207 16:03:03.159826 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:03 crc kubenswrapper[4716]: I1207 16:03:03.159848 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:03 crc kubenswrapper[4716]: I1207 16:03:03.159864 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:03Z","lastTransitionTime":"2025-12-07T16:03:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:03 crc kubenswrapper[4716]: I1207 16:03:03.261924 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:03 crc kubenswrapper[4716]: I1207 16:03:03.261987 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:03 crc kubenswrapper[4716]: I1207 16:03:03.262005 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:03 crc kubenswrapper[4716]: I1207 16:03:03.262029 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:03 crc kubenswrapper[4716]: I1207 16:03:03.262047 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:03Z","lastTransitionTime":"2025-12-07T16:03:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:03 crc kubenswrapper[4716]: I1207 16:03:03.364957 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:03 crc kubenswrapper[4716]: I1207 16:03:03.365016 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:03 crc kubenswrapper[4716]: I1207 16:03:03.365034 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:03 crc kubenswrapper[4716]: I1207 16:03:03.365058 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:03 crc kubenswrapper[4716]: I1207 16:03:03.365107 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:03Z","lastTransitionTime":"2025-12-07T16:03:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:03 crc kubenswrapper[4716]: I1207 16:03:03.467659 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:03 crc kubenswrapper[4716]: I1207 16:03:03.467740 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:03 crc kubenswrapper[4716]: I1207 16:03:03.467759 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:03 crc kubenswrapper[4716]: I1207 16:03:03.467783 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:03 crc kubenswrapper[4716]: I1207 16:03:03.467801 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:03Z","lastTransitionTime":"2025-12-07T16:03:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:03 crc kubenswrapper[4716]: I1207 16:03:03.571051 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:03 crc kubenswrapper[4716]: I1207 16:03:03.571248 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:03 crc kubenswrapper[4716]: I1207 16:03:03.571291 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:03 crc kubenswrapper[4716]: I1207 16:03:03.571309 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:03 crc kubenswrapper[4716]: I1207 16:03:03.571319 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:03Z","lastTransitionTime":"2025-12-07T16:03:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:03 crc kubenswrapper[4716]: I1207 16:03:03.657026 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nx2lh" Dec 07 16:03:03 crc kubenswrapper[4716]: I1207 16:03:03.657070 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 16:03:03 crc kubenswrapper[4716]: I1207 16:03:03.657116 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:03:03 crc kubenswrapper[4716]: E1207 16:03:03.657379 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nx2lh" podUID="04b37f6c-457b-4f9a-8733-e88cb9644483" Dec 07 16:03:03 crc kubenswrapper[4716]: E1207 16:03:03.657522 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 16:03:03 crc kubenswrapper[4716]: E1207 16:03:03.657648 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 16:03:03 crc kubenswrapper[4716]: I1207 16:03:03.675256 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:03 crc kubenswrapper[4716]: I1207 16:03:03.675315 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:03 crc kubenswrapper[4716]: I1207 16:03:03.675332 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:03 crc kubenswrapper[4716]: I1207 16:03:03.675358 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:03 crc kubenswrapper[4716]: I1207 16:03:03.675378 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:03Z","lastTransitionTime":"2025-12-07T16:03:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:03 crc kubenswrapper[4716]: I1207 16:03:03.778547 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:03 crc kubenswrapper[4716]: I1207 16:03:03.778637 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:03 crc kubenswrapper[4716]: I1207 16:03:03.778670 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:03 crc kubenswrapper[4716]: I1207 16:03:03.778703 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:03 crc kubenswrapper[4716]: I1207 16:03:03.778729 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:03Z","lastTransitionTime":"2025-12-07T16:03:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:03 crc kubenswrapper[4716]: I1207 16:03:03.881568 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:03 crc kubenswrapper[4716]: I1207 16:03:03.881627 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:03 crc kubenswrapper[4716]: I1207 16:03:03.881679 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:03 crc kubenswrapper[4716]: I1207 16:03:03.881713 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:03 crc kubenswrapper[4716]: I1207 16:03:03.881737 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:03Z","lastTransitionTime":"2025-12-07T16:03:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:03 crc kubenswrapper[4716]: I1207 16:03:03.985003 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:03 crc kubenswrapper[4716]: I1207 16:03:03.985046 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:03 crc kubenswrapper[4716]: I1207 16:03:03.985059 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:03 crc kubenswrapper[4716]: I1207 16:03:03.985106 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:03 crc kubenswrapper[4716]: I1207 16:03:03.985126 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:03Z","lastTransitionTime":"2025-12-07T16:03:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:04 crc kubenswrapper[4716]: I1207 16:03:04.088236 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:04 crc kubenswrapper[4716]: I1207 16:03:04.088297 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:04 crc kubenswrapper[4716]: I1207 16:03:04.088310 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:04 crc kubenswrapper[4716]: I1207 16:03:04.088328 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:04 crc kubenswrapper[4716]: I1207 16:03:04.088340 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:04Z","lastTransitionTime":"2025-12-07T16:03:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:04 crc kubenswrapper[4716]: I1207 16:03:04.191284 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:04 crc kubenswrapper[4716]: I1207 16:03:04.191347 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:04 crc kubenswrapper[4716]: I1207 16:03:04.191371 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:04 crc kubenswrapper[4716]: I1207 16:03:04.191396 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:04 crc kubenswrapper[4716]: I1207 16:03:04.191414 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:04Z","lastTransitionTime":"2025-12-07T16:03:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:04 crc kubenswrapper[4716]: I1207 16:03:04.293836 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:04 crc kubenswrapper[4716]: I1207 16:03:04.293875 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:04 crc kubenswrapper[4716]: I1207 16:03:04.293887 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:04 crc kubenswrapper[4716]: I1207 16:03:04.293905 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:04 crc kubenswrapper[4716]: I1207 16:03:04.293917 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:04Z","lastTransitionTime":"2025-12-07T16:03:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:04 crc kubenswrapper[4716]: I1207 16:03:04.396228 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:04 crc kubenswrapper[4716]: I1207 16:03:04.396328 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:04 crc kubenswrapper[4716]: I1207 16:03:04.396340 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:04 crc kubenswrapper[4716]: I1207 16:03:04.396358 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:04 crc kubenswrapper[4716]: I1207 16:03:04.396369 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:04Z","lastTransitionTime":"2025-12-07T16:03:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:04 crc kubenswrapper[4716]: I1207 16:03:04.498107 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:04 crc kubenswrapper[4716]: I1207 16:03:04.498155 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:04 crc kubenswrapper[4716]: I1207 16:03:04.498163 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:04 crc kubenswrapper[4716]: I1207 16:03:04.498176 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:04 crc kubenswrapper[4716]: I1207 16:03:04.498185 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:04Z","lastTransitionTime":"2025-12-07T16:03:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:04 crc kubenswrapper[4716]: I1207 16:03:04.547149 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 07 16:03:04 crc kubenswrapper[4716]: I1207 16:03:04.553863 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 07 16:03:04 crc kubenswrapper[4716]: I1207 16:03:04.564962 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef3335ae3875a75ab64685fc7e3590177280d651e8f8292c6f5ea6f472e55789\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:04Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:04 crc kubenswrapper[4716]: I1207 16:03:04.580467 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:04Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:04 crc kubenswrapper[4716]: I1207 16:03:04.591575 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f95ef61779507288a03bda8eddf745e105f705998a117a30789675bfa452d88f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b955b23163e7c8c458a3bbe2f681633e6a663ce76748febcc0b3a3c37f3ca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:04Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:04 crc kubenswrapper[4716]: I1207 16:03:04.601811 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:04 crc kubenswrapper[4716]: I1207 16:03:04.601936 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:04 crc kubenswrapper[4716]: I1207 16:03:04.601250 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c15b59eb-565d-4556-a4ce-75afdf159dc8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://530a2935826940767ea33fb4f37fa11f7dee4950d867fbcd6af72836d80f7db3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jks8z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ccaf6c1e6a978eb978a1c207b6c4c00b23737f78b743f966579aca781e22e0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jks8z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zcxxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:04Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:04 crc kubenswrapper[4716]: I1207 16:03:04.601963 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:04 crc kubenswrapper[4716]: I1207 16:03:04.602133 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:04 crc kubenswrapper[4716]: I1207 16:03:04.602152 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:04Z","lastTransitionTime":"2025-12-07T16:03:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:04 crc kubenswrapper[4716]: I1207 16:03:04.614398 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8ps9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98015cea-d753-4e7a-b2b5-5ae7733a81e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d68509b673eecd37725cdcc48e3d0612c3b9c7601433f2b72c21d662140daa59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f84f52a073437543721010dc440339654ffb5e04eeb6cc7afbcc62f00f64c8e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f84f52a073437543721010dc440339654ffb5e04eeb6cc7afbcc62f00f64c8e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f6fe1cb25218cb3ac20488dc47efa4dc6a26a43dbc8306b5ee38cbe1a186c67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f6fe1cb25218cb3ac20488dc47efa4dc6a26a43dbc8306b5ee38cbe1a186c67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14b748a8247f6863e8be61e90057a06626dfd3cd4f01bd3a11629f700f56378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f14b748a8247f6863e8be61e90057a06626dfd3cd4f01bd3a11629f700f56378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ec379c0d573cb9b5ffcfc5ca7c653ba60e2fbdd7e3d37836102e26618ce43e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ec379c0d573cb9b5ffcfc5ca7c653ba60e2fbdd7e3d37836102e26618ce43e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c75bf2a6703cf99865d4ed5fa000e1e3b1330d4946176cf585fec1a2f226df0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c75bf2a6703cf99865d4ed5fa000e1e3b1330d4946176cf585fec1a2f226df0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de45812c9a5a5bfe6b0146828d03fab28362b03d04ef861c150b9ffbf181d345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de45812c9a5a5bfe6b0146828d03fab28362b03d04ef861c150b9ffbf181d345\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8ps9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:04Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:04 crc kubenswrapper[4716]: I1207 16:03:04.635444 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7961cc9a-734e-42c2-b1ff-7075f3f20091\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84a36b76f20618c7885d611ecad9db29bac5101c27572be31d3acbb90c581291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0a0eca3c2087a457c0e448e649345ec4595a9c8a019aed2cddc5f30c77871e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://404426a171b8676eddd4a00996c76afab7e269643fa9608f1e1b6bdba385071d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fd3c2941782921f99a6cdc3de4e892234548bde527fec15a8ac92e513566e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e23abfdca314ed43d594333c9f9e401cc8ef7dad39b8ebac84a836b06ee5c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09b38cd4b389ea477c802a842ee4f2297a64d64684624117163e819539c257c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09b38cd4b389ea477c802a842ee4f2297a64d64684624117163e819539c257c6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c00f7a8e024e0e0dfd7b73e16d1d5437602e79ea95b73c21f24b209edef60e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c00f7a8e024e0e0dfd7b73e16d1d5437602e79ea95b73c21f24b209edef60e98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3f6392d426fa2cd788ec3a5c3bff4ac39483a54dfb2624a1c4370522413a350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3f6392d426fa2cd788ec3a5c3bff4ac39483a54dfb2624a1c4370522413a350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:04Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:04 crc kubenswrapper[4716]: I1207 16:03:04.650688 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3067de7c-c11c-4be1-bd94-d6572c266795\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d14d018fa044e9e80fe81e7e38df935992d14c773ed2b2193fd5cb873e5d0947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a9e9cf048220ca14c996847e9ced9f2367b9331dc6c2836dd023658855cd79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6d3c911874095b3c7cb062d03070e5090a7bb2456cb20cb4ce078f502515cda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ac6ce7701f1c6aa1ea8fcf7a3035e3b0eaa28ae3cb6afc442372bd9117bdde4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:04Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:04 crc kubenswrapper[4716]: I1207 16:03:04.657033 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 16:03:04 crc kubenswrapper[4716]: E1207 16:03:04.657245 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 16:03:04 crc kubenswrapper[4716]: I1207 16:03:04.662830 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4742ea84b3a2c9461d9b900beee7354a2ded24c39ad23e007cbf270ae12377c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:04Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:04 crc kubenswrapper[4716]: I1207 16:03:04.679144 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:04Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:04 crc kubenswrapper[4716]: I1207 16:03:04.706282 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:04 crc kubenswrapper[4716]: I1207 16:03:04.706518 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:04 crc kubenswrapper[4716]: I1207 16:03:04.706726 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:04 crc kubenswrapper[4716]: I1207 16:03:04.706839 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:04 crc kubenswrapper[4716]: I1207 16:03:04.706979 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:04Z","lastTransitionTime":"2025-12-07T16:03:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:04 crc kubenswrapper[4716]: I1207 16:03:04.708263 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c54d618a-dfec-4f2a-a288-f5188c8a128a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60324dc7d701bc064e14ae4f474bce10f6ee4ff3c3c942eab749e7166082cf9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ed4f4d3564674bf5cb4e7a76dc0fd95e203ecc8950798e59a1fb6c9ab4a20f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://def9ac840fb8ecf2778bed88ea980eddca7c5b2aba4bd1afb967d62fec3f7bae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b73be664f5412c98b89d971a2cac37c494151c293fdb44a98b390978658a6643\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a847d04e22e98091c0037a89054b9236228b701b8c7ba31940a56df0f37b95c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcae78a6133d97f6a32dd7421ccec4333b970a237404e056a5cc601f8eefdc8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8baee0f3d1e686bc2c17b0777f4eacc31c7f1fe44720e68f94612e6c806bc8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8baee0f3d1e686bc2c17b0777f4eacc31c7f1fe44720e68f94612e6c806bc8b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-07T16:02:52Z\\\",\\\"message\\\":\\\"r occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:51Z is after 2025-08-24T17:21:41Z]\\\\nI1207 16:02:51.943802 6191 services_controller.go:444] Built service openshift-marketplace/redhat-marketplace LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1207 16:02:51.943764 6191 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-route-controller-manager/route-controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"18746a4d-8a63-458a-b7e3-8fb89ff95fc0\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-route-controller-manager/route-controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-route-controller-manager/route-controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-qvc7p_openshift-ovn-kubernetes(c54d618a-dfec-4f2a-a288-f5188c8a128a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcfdffe822635b0ae13b253e6133449fe00461b2792be43d913c4a77219b8255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9bfe2ffce7a7b7c320474117734fc05246cc2b4fd744833e896e8e7c83f5e4d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9bfe2ffce7a7b7c320474117734fc05246cc2b4fd744833e896e8e7c83f5e4d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qvc7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:04Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:04 crc kubenswrapper[4716]: I1207 16:03:04.724033 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mx85f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"118dd9ba-437c-4e5c-b5ff-4239ae6d4a63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bcd6c8deb184d0c713c5e7b7cc82fa38d3cd498517911ba7a0ec3cb88a3437a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gg5nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ffc069d2b21fa80871b1b97f2c72c37880d9a309cbf8b7daf7d240f2c332a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gg5nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mx85f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:04Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:04 crc kubenswrapper[4716]: I1207 16:03:04.739919 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nx2lh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04b37f6c-457b-4f9a-8733-e88cb9644483\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv6td\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv6td\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nx2lh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:04Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:04 crc kubenswrapper[4716]: I1207 16:03:04.753787 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-m5jl5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"824ee59f-1474-4994-a70e-462878521514\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a15fe50cc75515704779646cc19aa50004f0db9aa26fe2535110df749bbfb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhwzn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-m5jl5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:04Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:04 crc kubenswrapper[4716]: I1207 16:03:04.766195 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdee9957-48b9-458f-ba71-e0cd39f80596\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73f85248b4d68a9beecdacb3a01f433e97a0ffea2247ce92a41ef96855509377\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6276df652efe6ca935e745f24ef20f87a16465475139230c63b9999234a4d12a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3b1dd202586ee755ca89fbe12b30bbb12f95a62514f6e4d93cbbdbe55181e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd318647816e4a27849666b0ec4fc2ca4905295661c557dc381e1144f22cf284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d31de0326a05216fa10442bdb47151c6e073c97e9a16c34e793028e5e253756\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"le observer\\\\nW1207 16:02:34.645464 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1207 16:02:34.645587 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 16:02:34.646501 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2082283479/tls.crt::/tmp/serving-cert-2082283479/tls.key\\\\\\\"\\\\nI1207 16:02:34.972041 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 16:02:34.973702 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 16:02:34.973717 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 16:02:34.973738 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 16:02:34.973743 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 16:02:34.977185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1207 16:02:34.977204 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1207 16:02:34.977209 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977217 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 16:02:34.977225 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 16:02:34.977228 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 16:02:34.977232 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1207 16:02:34.979033 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27bd4a54f920b2725987c1c7fd2d9fdf16cd08738c246c333a4e9164a3a582a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:04Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:04 crc kubenswrapper[4716]: I1207 16:03:04.778561 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:04Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:04 crc kubenswrapper[4716]: I1207 16:03:04.789726 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2jjld" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dee6f5c-861c-4819-9309-7791574c7b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51c1d4159ab11087839a2c2d192566fb9322447c26df560bf09517c906ececb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmdns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2jjld\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:04Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:04 crc kubenswrapper[4716]: I1207 16:03:04.801917 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4wr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0baf8575-26cf-4a72-8a6d-ae94c078eaad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0ca2b3cd5a11033afd219b0125be7ab5acd750935d7eb89a470eff3d93a0838\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p2trd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4wr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:04Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:04 crc kubenswrapper[4716]: I1207 16:03:04.809766 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:04 crc kubenswrapper[4716]: I1207 16:03:04.809830 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:04 crc kubenswrapper[4716]: I1207 16:03:04.809855 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:04 crc kubenswrapper[4716]: I1207 16:03:04.809885 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:04 crc kubenswrapper[4716]: I1207 16:03:04.809907 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:04Z","lastTransitionTime":"2025-12-07T16:03:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:04 crc kubenswrapper[4716]: I1207 16:03:04.913166 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:04 crc kubenswrapper[4716]: I1207 16:03:04.913246 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:04 crc kubenswrapper[4716]: I1207 16:03:04.913260 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:04 crc kubenswrapper[4716]: I1207 16:03:04.913277 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:04 crc kubenswrapper[4716]: I1207 16:03:04.913290 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:04Z","lastTransitionTime":"2025-12-07T16:03:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:05 crc kubenswrapper[4716]: I1207 16:03:05.015557 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:05 crc kubenswrapper[4716]: I1207 16:03:05.015613 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:05 crc kubenswrapper[4716]: I1207 16:03:05.015631 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:05 crc kubenswrapper[4716]: I1207 16:03:05.015656 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:05 crc kubenswrapper[4716]: I1207 16:03:05.015674 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:05Z","lastTransitionTime":"2025-12-07T16:03:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:05 crc kubenswrapper[4716]: I1207 16:03:05.118923 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:05 crc kubenswrapper[4716]: I1207 16:03:05.118978 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:05 crc kubenswrapper[4716]: I1207 16:03:05.118994 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:05 crc kubenswrapper[4716]: I1207 16:03:05.119017 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:05 crc kubenswrapper[4716]: I1207 16:03:05.119038 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:05Z","lastTransitionTime":"2025-12-07T16:03:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:05 crc kubenswrapper[4716]: I1207 16:03:05.222152 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:05 crc kubenswrapper[4716]: I1207 16:03:05.222350 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:05 crc kubenswrapper[4716]: I1207 16:03:05.222408 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:05 crc kubenswrapper[4716]: I1207 16:03:05.222494 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:05 crc kubenswrapper[4716]: I1207 16:03:05.222552 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:05Z","lastTransitionTime":"2025-12-07T16:03:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:05 crc kubenswrapper[4716]: I1207 16:03:05.325816 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:05 crc kubenswrapper[4716]: I1207 16:03:05.325884 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:05 crc kubenswrapper[4716]: I1207 16:03:05.325909 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:05 crc kubenswrapper[4716]: I1207 16:03:05.325941 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:05 crc kubenswrapper[4716]: I1207 16:03:05.325968 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:05Z","lastTransitionTime":"2025-12-07T16:03:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:05 crc kubenswrapper[4716]: I1207 16:03:05.429396 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:05 crc kubenswrapper[4716]: I1207 16:03:05.429439 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:05 crc kubenswrapper[4716]: I1207 16:03:05.429448 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:05 crc kubenswrapper[4716]: I1207 16:03:05.429462 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:05 crc kubenswrapper[4716]: I1207 16:03:05.429472 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:05Z","lastTransitionTime":"2025-12-07T16:03:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:05 crc kubenswrapper[4716]: I1207 16:03:05.531835 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:05 crc kubenswrapper[4716]: I1207 16:03:05.532050 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:05 crc kubenswrapper[4716]: I1207 16:03:05.532130 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:05 crc kubenswrapper[4716]: I1207 16:03:05.532202 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:05 crc kubenswrapper[4716]: I1207 16:03:05.532298 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:05Z","lastTransitionTime":"2025-12-07T16:03:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:05 crc kubenswrapper[4716]: I1207 16:03:05.634540 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:05 crc kubenswrapper[4716]: I1207 16:03:05.634605 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:05 crc kubenswrapper[4716]: I1207 16:03:05.634630 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:05 crc kubenswrapper[4716]: I1207 16:03:05.634652 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:05 crc kubenswrapper[4716]: I1207 16:03:05.634668 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:05Z","lastTransitionTime":"2025-12-07T16:03:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:05 crc kubenswrapper[4716]: I1207 16:03:05.657453 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nx2lh" Dec 07 16:03:05 crc kubenswrapper[4716]: I1207 16:03:05.657541 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:03:05 crc kubenswrapper[4716]: E1207 16:03:05.657601 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nx2lh" podUID="04b37f6c-457b-4f9a-8733-e88cb9644483" Dec 07 16:03:05 crc kubenswrapper[4716]: E1207 16:03:05.657700 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 16:03:05 crc kubenswrapper[4716]: I1207 16:03:05.657808 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 16:03:05 crc kubenswrapper[4716]: E1207 16:03:05.658021 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 16:03:05 crc kubenswrapper[4716]: I1207 16:03:05.737698 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:05 crc kubenswrapper[4716]: I1207 16:03:05.737739 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:05 crc kubenswrapper[4716]: I1207 16:03:05.737749 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:05 crc kubenswrapper[4716]: I1207 16:03:05.737765 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:05 crc kubenswrapper[4716]: I1207 16:03:05.737776 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:05Z","lastTransitionTime":"2025-12-07T16:03:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:05 crc kubenswrapper[4716]: I1207 16:03:05.841213 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:05 crc kubenswrapper[4716]: I1207 16:03:05.841317 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:05 crc kubenswrapper[4716]: I1207 16:03:05.841337 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:05 crc kubenswrapper[4716]: I1207 16:03:05.841364 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:05 crc kubenswrapper[4716]: I1207 16:03:05.841384 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:05Z","lastTransitionTime":"2025-12-07T16:03:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:05 crc kubenswrapper[4716]: I1207 16:03:05.944024 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:05 crc kubenswrapper[4716]: I1207 16:03:05.944067 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:05 crc kubenswrapper[4716]: I1207 16:03:05.944110 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:05 crc kubenswrapper[4716]: I1207 16:03:05.944133 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:05 crc kubenswrapper[4716]: I1207 16:03:05.944150 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:05Z","lastTransitionTime":"2025-12-07T16:03:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:06 crc kubenswrapper[4716]: I1207 16:03:06.046165 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:06 crc kubenswrapper[4716]: I1207 16:03:06.046210 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:06 crc kubenswrapper[4716]: I1207 16:03:06.046224 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:06 crc kubenswrapper[4716]: I1207 16:03:06.046242 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:06 crc kubenswrapper[4716]: I1207 16:03:06.046255 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:06Z","lastTransitionTime":"2025-12-07T16:03:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:06 crc kubenswrapper[4716]: I1207 16:03:06.148063 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:06 crc kubenswrapper[4716]: I1207 16:03:06.148299 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:06 crc kubenswrapper[4716]: I1207 16:03:06.148360 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:06 crc kubenswrapper[4716]: I1207 16:03:06.148422 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:06 crc kubenswrapper[4716]: I1207 16:03:06.148476 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:06Z","lastTransitionTime":"2025-12-07T16:03:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:06 crc kubenswrapper[4716]: I1207 16:03:06.251342 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:06 crc kubenswrapper[4716]: I1207 16:03:06.251401 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:06 crc kubenswrapper[4716]: I1207 16:03:06.251415 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:06 crc kubenswrapper[4716]: I1207 16:03:06.251435 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:06 crc kubenswrapper[4716]: I1207 16:03:06.251448 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:06Z","lastTransitionTime":"2025-12-07T16:03:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:06 crc kubenswrapper[4716]: I1207 16:03:06.354532 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:06 crc kubenswrapper[4716]: I1207 16:03:06.355198 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:06 crc kubenswrapper[4716]: I1207 16:03:06.355224 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:06 crc kubenswrapper[4716]: I1207 16:03:06.355240 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:06 crc kubenswrapper[4716]: I1207 16:03:06.355251 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:06Z","lastTransitionTime":"2025-12-07T16:03:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:06 crc kubenswrapper[4716]: I1207 16:03:06.457806 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:06 crc kubenswrapper[4716]: I1207 16:03:06.457838 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:06 crc kubenswrapper[4716]: I1207 16:03:06.457846 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:06 crc kubenswrapper[4716]: I1207 16:03:06.457858 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:06 crc kubenswrapper[4716]: I1207 16:03:06.457868 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:06Z","lastTransitionTime":"2025-12-07T16:03:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:06 crc kubenswrapper[4716]: I1207 16:03:06.560106 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:06 crc kubenswrapper[4716]: I1207 16:03:06.560143 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:06 crc kubenswrapper[4716]: I1207 16:03:06.560153 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:06 crc kubenswrapper[4716]: I1207 16:03:06.560165 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:06 crc kubenswrapper[4716]: I1207 16:03:06.560173 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:06Z","lastTransitionTime":"2025-12-07T16:03:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:06 crc kubenswrapper[4716]: I1207 16:03:06.657010 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 16:03:06 crc kubenswrapper[4716]: E1207 16:03:06.657155 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 16:03:06 crc kubenswrapper[4716]: I1207 16:03:06.663638 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:06 crc kubenswrapper[4716]: I1207 16:03:06.663679 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:06 crc kubenswrapper[4716]: I1207 16:03:06.663692 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:06 crc kubenswrapper[4716]: I1207 16:03:06.663708 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:06 crc kubenswrapper[4716]: I1207 16:03:06.663722 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:06Z","lastTransitionTime":"2025-12-07T16:03:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:06 crc kubenswrapper[4716]: I1207 16:03:06.756723 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:06 crc kubenswrapper[4716]: I1207 16:03:06.756769 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:06 crc kubenswrapper[4716]: I1207 16:03:06.756780 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:06 crc kubenswrapper[4716]: I1207 16:03:06.756796 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:06 crc kubenswrapper[4716]: I1207 16:03:06.756809 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:06Z","lastTransitionTime":"2025-12-07T16:03:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:06 crc kubenswrapper[4716]: E1207 16:03:06.770239 4716 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a8d1b7d4-7ec4-40c7-86b0-291db1f38894\\\",\\\"systemUUID\\\":\\\"a57d09b0-0224-46eb-a3ac-a2b60dad2250\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:06Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:06 crc kubenswrapper[4716]: I1207 16:03:06.774278 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:06 crc kubenswrapper[4716]: I1207 16:03:06.774374 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:06 crc kubenswrapper[4716]: I1207 16:03:06.774428 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:06 crc kubenswrapper[4716]: I1207 16:03:06.774452 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:06 crc kubenswrapper[4716]: I1207 16:03:06.774468 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:06Z","lastTransitionTime":"2025-12-07T16:03:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:06 crc kubenswrapper[4716]: E1207 16:03:06.793937 4716 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a8d1b7d4-7ec4-40c7-86b0-291db1f38894\\\",\\\"systemUUID\\\":\\\"a57d09b0-0224-46eb-a3ac-a2b60dad2250\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:06Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:06 crc kubenswrapper[4716]: I1207 16:03:06.798190 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:06 crc kubenswrapper[4716]: I1207 16:03:06.798231 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:06 crc kubenswrapper[4716]: I1207 16:03:06.798248 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:06 crc kubenswrapper[4716]: I1207 16:03:06.798270 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:06 crc kubenswrapper[4716]: I1207 16:03:06.798286 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:06Z","lastTransitionTime":"2025-12-07T16:03:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:06 crc kubenswrapper[4716]: E1207 16:03:06.816357 4716 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a8d1b7d4-7ec4-40c7-86b0-291db1f38894\\\",\\\"systemUUID\\\":\\\"a57d09b0-0224-46eb-a3ac-a2b60dad2250\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:06Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:06 crc kubenswrapper[4716]: I1207 16:03:06.821181 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:06 crc kubenswrapper[4716]: I1207 16:03:06.821219 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:06 crc kubenswrapper[4716]: I1207 16:03:06.821229 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:06 crc kubenswrapper[4716]: I1207 16:03:06.821243 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:06 crc kubenswrapper[4716]: I1207 16:03:06.821252 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:06Z","lastTransitionTime":"2025-12-07T16:03:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:06 crc kubenswrapper[4716]: E1207 16:03:06.839111 4716 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a8d1b7d4-7ec4-40c7-86b0-291db1f38894\\\",\\\"systemUUID\\\":\\\"a57d09b0-0224-46eb-a3ac-a2b60dad2250\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:06Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:06 crc kubenswrapper[4716]: I1207 16:03:06.843331 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:06 crc kubenswrapper[4716]: I1207 16:03:06.843357 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:06 crc kubenswrapper[4716]: I1207 16:03:06.843367 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:06 crc kubenswrapper[4716]: I1207 16:03:06.843382 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:06 crc kubenswrapper[4716]: I1207 16:03:06.843392 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:06Z","lastTransitionTime":"2025-12-07T16:03:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:06 crc kubenswrapper[4716]: E1207 16:03:06.859131 4716 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a8d1b7d4-7ec4-40c7-86b0-291db1f38894\\\",\\\"systemUUID\\\":\\\"a57d09b0-0224-46eb-a3ac-a2b60dad2250\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:06Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:06 crc kubenswrapper[4716]: E1207 16:03:06.859292 4716 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 07 16:03:06 crc kubenswrapper[4716]: I1207 16:03:06.860772 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:06 crc kubenswrapper[4716]: I1207 16:03:06.860798 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:06 crc kubenswrapper[4716]: I1207 16:03:06.860807 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:06 crc kubenswrapper[4716]: I1207 16:03:06.860821 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:06 crc kubenswrapper[4716]: I1207 16:03:06.860830 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:06Z","lastTransitionTime":"2025-12-07T16:03:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:06 crc kubenswrapper[4716]: I1207 16:03:06.962717 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:06 crc kubenswrapper[4716]: I1207 16:03:06.962751 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:06 crc kubenswrapper[4716]: I1207 16:03:06.962764 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:06 crc kubenswrapper[4716]: I1207 16:03:06.962780 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:06 crc kubenswrapper[4716]: I1207 16:03:06.962791 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:06Z","lastTransitionTime":"2025-12-07T16:03:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.065496 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.065782 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.065878 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.065947 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.066011 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:07Z","lastTransitionTime":"2025-12-07T16:03:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.168556 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.168870 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.169005 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.169278 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.169415 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:07Z","lastTransitionTime":"2025-12-07T16:03:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.272341 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.272400 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.272420 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.272443 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.272463 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:07Z","lastTransitionTime":"2025-12-07T16:03:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.374561 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.375154 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.375176 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.375191 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.375201 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:07Z","lastTransitionTime":"2025-12-07T16:03:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.414247 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.414346 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 16:03:07 crc kubenswrapper[4716]: E1207 16:03:07.414427 4716 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 07 16:03:07 crc kubenswrapper[4716]: E1207 16:03:07.414445 4716 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 07 16:03:07 crc kubenswrapper[4716]: E1207 16:03:07.414455 4716 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 07 16:03:07 crc kubenswrapper[4716]: E1207 16:03:07.414499 4716 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 07 16:03:07 crc kubenswrapper[4716]: E1207 16:03:07.414524 4716 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 07 16:03:07 crc kubenswrapper[4716]: E1207 16:03:07.414540 4716 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 07 16:03:07 crc kubenswrapper[4716]: E1207 16:03:07.414506 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-07 16:03:39.414492178 +0000 UTC m=+82.104777100 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 07 16:03:07 crc kubenswrapper[4716]: E1207 16:03:07.414792 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-07 16:03:39.414775116 +0000 UTC m=+82.105060068 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.477476 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.477834 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.478063 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.478418 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.478685 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:07Z","lastTransitionTime":"2025-12-07T16:03:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.515340 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 16:03:07 crc kubenswrapper[4716]: E1207 16:03:07.515538 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 16:03:39.515499589 +0000 UTC m=+82.205784531 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.515816 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.515997 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:03:07 crc kubenswrapper[4716]: E1207 16:03:07.516066 4716 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 07 16:03:07 crc kubenswrapper[4716]: E1207 16:03:07.516195 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-07 16:03:39.516167858 +0000 UTC m=+82.206452840 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 07 16:03:07 crc kubenswrapper[4716]: E1207 16:03:07.516210 4716 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 07 16:03:07 crc kubenswrapper[4716]: E1207 16:03:07.516268 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-07 16:03:39.516253351 +0000 UTC m=+82.206538303 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.582244 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.582331 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.582358 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.582387 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.582407 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:07Z","lastTransitionTime":"2025-12-07T16:03:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.657572 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nx2lh" Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.657752 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 16:03:07 crc kubenswrapper[4716]: E1207 16:03:07.658705 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nx2lh" podUID="04b37f6c-457b-4f9a-8733-e88cb9644483" Dec 07 16:03:07 crc kubenswrapper[4716]: E1207 16:03:07.658764 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.659038 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:03:07 crc kubenswrapper[4716]: E1207 16:03:07.660683 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.662277 4716 scope.go:117] "RemoveContainer" containerID="a8baee0f3d1e686bc2c17b0777f4eacc31c7f1fe44720e68f94612e6c806bc8b" Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.678497 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:07Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.686246 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.686322 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.686340 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.686363 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.686414 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:07Z","lastTransitionTime":"2025-12-07T16:03:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.695620 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c15b59eb-565d-4556-a4ce-75afdf159dc8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://530a2935826940767ea33fb4f37fa11f7dee4950d867fbcd6af72836d80f7db3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jks8z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ccaf6c1e6a978eb978a1c207b6c4c00b23737f78b743f966579aca781e22e0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jks8z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zcxxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:07Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.719965 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8ps9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98015cea-d753-4e7a-b2b5-5ae7733a81e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d68509b673eecd37725cdcc48e3d0612c3b9c7601433f2b72c21d662140daa59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f84f52a073437543721010dc440339654ffb5e04eeb6cc7afbcc62f00f64c8e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f84f52a073437543721010dc440339654ffb5e04eeb6cc7afbcc62f00f64c8e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f6fe1cb25218cb3ac20488dc47efa4dc6a26a43dbc8306b5ee38cbe1a186c67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f6fe1cb25218cb3ac20488dc47efa4dc6a26a43dbc8306b5ee38cbe1a186c67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14b748a8247f6863e8be61e90057a06626dfd3cd4f01bd3a11629f700f56378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f14b748a8247f6863e8be61e90057a06626dfd3cd4f01bd3a11629f700f56378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ec379c0d573cb9b5ffcfc5ca7c653ba60e2fbdd7e3d37836102e26618ce43e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ec379c0d573cb9b5ffcfc5ca7c653ba60e2fbdd7e3d37836102e26618ce43e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c75bf2a6703cf99865d4ed5fa000e1e3b1330d4946176cf585fec1a2f226df0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c75bf2a6703cf99865d4ed5fa000e1e3b1330d4946176cf585fec1a2f226df0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de45812c9a5a5bfe6b0146828d03fab28362b03d04ef861c150b9ffbf181d345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de45812c9a5a5bfe6b0146828d03fab28362b03d04ef861c150b9ffbf181d345\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8ps9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:07Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.757771 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7961cc9a-734e-42c2-b1ff-7075f3f20091\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84a36b76f20618c7885d611ecad9db29bac5101c27572be31d3acbb90c581291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0a0eca3c2087a457c0e448e649345ec4595a9c8a019aed2cddc5f30c77871e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://404426a171b8676eddd4a00996c76afab7e269643fa9608f1e1b6bdba385071d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fd3c2941782921f99a6cdc3de4e892234548bde527fec15a8ac92e513566e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e23abfdca314ed43d594333c9f9e401cc8ef7dad39b8ebac84a836b06ee5c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09b38cd4b389ea477c802a842ee4f2297a64d64684624117163e819539c257c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09b38cd4b389ea477c802a842ee4f2297a64d64684624117163e819539c257c6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c00f7a8e024e0e0dfd7b73e16d1d5437602e79ea95b73c21f24b209edef60e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c00f7a8e024e0e0dfd7b73e16d1d5437602e79ea95b73c21f24b209edef60e98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3f6392d426fa2cd788ec3a5c3bff4ac39483a54dfb2624a1c4370522413a350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3f6392d426fa2cd788ec3a5c3bff4ac39483a54dfb2624a1c4370522413a350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:07Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.778512 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3067de7c-c11c-4be1-bd94-d6572c266795\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d14d018fa044e9e80fe81e7e38df935992d14c773ed2b2193fd5cb873e5d0947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a9e9cf048220ca14c996847e9ced9f2367b9331dc6c2836dd023658855cd79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6d3c911874095b3c7cb062d03070e5090a7bb2456cb20cb4ce078f502515cda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ac6ce7701f1c6aa1ea8fcf7a3035e3b0eaa28ae3cb6afc442372bd9117bdde4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:07Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.789570 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.789620 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.789639 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.789665 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.789683 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:07Z","lastTransitionTime":"2025-12-07T16:03:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.800462 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4742ea84b3a2c9461d9b900beee7354a2ded24c39ad23e007cbf270ae12377c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:07Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.813343 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nx2lh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04b37f6c-457b-4f9a-8733-e88cb9644483\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv6td\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv6td\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nx2lh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:07Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.823389 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e903f3c-ba8c-477d-b66a-6d1331a7c64e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ff9c456318bbdb54ec3f1ac3e2c2366968f18ad084158b7100d142acb668948\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1358745250b35426602fc7417e4f85dd411cdd0fb3e8f703c432d72128ca7c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d8c9a434c6cfb9e4fee6a8955f1379291f153c007876d1ffb498ce561c73ba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad8df9b5bd5f5a8dd44ce2899cfa29c19435149b170d339d339183cd3df09e99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad8df9b5bd5f5a8dd44ce2899cfa29c19435149b170d339d339183cd3df09e99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:07Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.845511 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c54d618a-dfec-4f2a-a288-f5188c8a128a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60324dc7d701bc064e14ae4f474bce10f6ee4ff3c3c942eab749e7166082cf9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ed4f4d3564674bf5cb4e7a76dc0fd95e203ecc8950798e59a1fb6c9ab4a20f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://def9ac840fb8ecf2778bed88ea980eddca7c5b2aba4bd1afb967d62fec3f7bae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b73be664f5412c98b89d971a2cac37c494151c293fdb44a98b390978658a6643\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a847d04e22e98091c0037a89054b9236228b701b8c7ba31940a56df0f37b95c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcae78a6133d97f6a32dd7421ccec4333b970a237404e056a5cc601f8eefdc8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8baee0f3d1e686bc2c17b0777f4eacc31c7f1fe44720e68f94612e6c806bc8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8baee0f3d1e686bc2c17b0777f4eacc31c7f1fe44720e68f94612e6c806bc8b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-07T16:02:52Z\\\",\\\"message\\\":\\\"r occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:51Z is after 2025-08-24T17:21:41Z]\\\\nI1207 16:02:51.943802 6191 services_controller.go:444] Built service openshift-marketplace/redhat-marketplace LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1207 16:02:51.943764 6191 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-route-controller-manager/route-controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"18746a4d-8a63-458a-b7e3-8fb89ff95fc0\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-route-controller-manager/route-controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-route-controller-manager/route-controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-qvc7p_openshift-ovn-kubernetes(c54d618a-dfec-4f2a-a288-f5188c8a128a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcfdffe822635b0ae13b253e6133449fe00461b2792be43d913c4a77219b8255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9bfe2ffce7a7b7c320474117734fc05246cc2b4fd744833e896e8e7c83f5e4d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9bfe2ffce7a7b7c320474117734fc05246cc2b4fd744833e896e8e7c83f5e4d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qvc7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:07Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.859251 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mx85f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"118dd9ba-437c-4e5c-b5ff-4239ae6d4a63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bcd6c8deb184d0c713c5e7b7cc82fa38d3cd498517911ba7a0ec3cb88a3437a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gg5nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ffc069d2b21fa80871b1b97f2c72c37880d9a309cbf8b7daf7d240f2c332a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gg5nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mx85f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:07Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.875706 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4wr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0baf8575-26cf-4a72-8a6d-ae94c078eaad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0ca2b3cd5a11033afd219b0125be7ab5acd750935d7eb89a470eff3d93a0838\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p2trd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4wr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:07Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.885977 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-m5jl5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"824ee59f-1474-4994-a70e-462878521514\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a15fe50cc75515704779646cc19aa50004f0db9aa26fe2535110df749bbfb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhwzn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-m5jl5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:07Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.892530 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.892573 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.892589 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.892616 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.892634 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:07Z","lastTransitionTime":"2025-12-07T16:03:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.898470 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdee9957-48b9-458f-ba71-e0cd39f80596\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73f85248b4d68a9beecdacb3a01f433e97a0ffea2247ce92a41ef96855509377\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6276df652efe6ca935e745f24ef20f87a16465475139230c63b9999234a4d12a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3b1dd202586ee755ca89fbe12b30bbb12f95a62514f6e4d93cbbdbe55181e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd318647816e4a27849666b0ec4fc2ca4905295661c557dc381e1144f22cf284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d31de0326a05216fa10442bdb47151c6e073c97e9a16c34e793028e5e253756\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"le observer\\\\nW1207 16:02:34.645464 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1207 16:02:34.645587 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 16:02:34.646501 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2082283479/tls.crt::/tmp/serving-cert-2082283479/tls.key\\\\\\\"\\\\nI1207 16:02:34.972041 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 16:02:34.973702 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 16:02:34.973717 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 16:02:34.973738 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 16:02:34.973743 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 16:02:34.977185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1207 16:02:34.977204 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1207 16:02:34.977209 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977217 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 16:02:34.977225 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 16:02:34.977228 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 16:02:34.977232 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1207 16:02:34.979033 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27bd4a54f920b2725987c1c7fd2d9fdf16cd08738c246c333a4e9164a3a582a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:07Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.912159 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:07Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.922304 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2jjld" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dee6f5c-861c-4819-9309-7791574c7b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51c1d4159ab11087839a2c2d192566fb9322447c26df560bf09517c906ececb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmdns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2jjld\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:07Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.932738 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f95ef61779507288a03bda8eddf745e105f705998a117a30789675bfa452d88f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b955b23163e7c8c458a3bbe2f681633e6a663ce76748febcc0b3a3c37f3ca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:07Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.944333 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef3335ae3875a75ab64685fc7e3590177280d651e8f8292c6f5ea6f472e55789\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:07Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.955888 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:07Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.995314 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.995349 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.995362 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.995378 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:07 crc kubenswrapper[4716]: I1207 16:03:07.995389 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:07Z","lastTransitionTime":"2025-12-07T16:03:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:08 crc kubenswrapper[4716]: I1207 16:03:08.003507 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qvc7p_c54d618a-dfec-4f2a-a288-f5188c8a128a/ovnkube-controller/1.log" Dec 07 16:03:08 crc kubenswrapper[4716]: I1207 16:03:08.005604 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" event={"ID":"c54d618a-dfec-4f2a-a288-f5188c8a128a","Type":"ContainerStarted","Data":"f48167b38ff840f85f11b69d8090ec6c9c634ed26682daef93f5da99fc02ef61"} Dec 07 16:03:08 crc kubenswrapper[4716]: I1207 16:03:08.006649 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" Dec 07 16:03:08 crc kubenswrapper[4716]: I1207 16:03:08.021038 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e903f3c-ba8c-477d-b66a-6d1331a7c64e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ff9c456318bbdb54ec3f1ac3e2c2366968f18ad084158b7100d142acb668948\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1358745250b35426602fc7417e4f85dd411cdd0fb3e8f703c432d72128ca7c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d8c9a434c6cfb9e4fee6a8955f1379291f153c007876d1ffb498ce561c73ba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad8df9b5bd5f5a8dd44ce2899cfa29c19435149b170d339d339183cd3df09e99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad8df9b5bd5f5a8dd44ce2899cfa29c19435149b170d339d339183cd3df09e99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:08Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:08 crc kubenswrapper[4716]: I1207 16:03:08.037479 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c54d618a-dfec-4f2a-a288-f5188c8a128a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60324dc7d701bc064e14ae4f474bce10f6ee4ff3c3c942eab749e7166082cf9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ed4f4d3564674bf5cb4e7a76dc0fd95e203ecc8950798e59a1fb6c9ab4a20f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://def9ac840fb8ecf2778bed88ea980eddca7c5b2aba4bd1afb967d62fec3f7bae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b73be664f5412c98b89d971a2cac37c494151c293fdb44a98b390978658a6643\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a847d04e22e98091c0037a89054b9236228b701b8c7ba31940a56df0f37b95c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcae78a6133d97f6a32dd7421ccec4333b970a237404e056a5cc601f8eefdc8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f48167b38ff840f85f11b69d8090ec6c9c634ed26682daef93f5da99fc02ef61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8baee0f3d1e686bc2c17b0777f4eacc31c7f1fe44720e68f94612e6c806bc8b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-07T16:02:52Z\\\",\\\"message\\\":\\\"r occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:51Z is after 2025-08-24T17:21:41Z]\\\\nI1207 16:02:51.943802 6191 services_controller.go:444] Built service openshift-marketplace/redhat-marketplace LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1207 16:02:51.943764 6191 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-route-controller-manager/route-controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"18746a4d-8a63-458a-b7e3-8fb89ff95fc0\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-route-controller-manager/route-controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-route-controller-manager/route-controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:03:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcfdffe822635b0ae13b253e6133449fe00461b2792be43d913c4a77219b8255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9bfe2ffce7a7b7c320474117734fc05246cc2b4fd744833e896e8e7c83f5e4d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9bfe2ffce7a7b7c320474117734fc05246cc2b4fd744833e896e8e7c83f5e4d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qvc7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:08Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:08 crc kubenswrapper[4716]: I1207 16:03:08.049556 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mx85f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"118dd9ba-437c-4e5c-b5ff-4239ae6d4a63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bcd6c8deb184d0c713c5e7b7cc82fa38d3cd498517911ba7a0ec3cb88a3437a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gg5nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ffc069d2b21fa80871b1b97f2c72c37880d9a309cbf8b7daf7d240f2c332a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gg5nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mx85f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:08Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:08 crc kubenswrapper[4716]: I1207 16:03:08.062044 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nx2lh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04b37f6c-457b-4f9a-8733-e88cb9644483\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv6td\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv6td\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nx2lh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:08Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:08 crc kubenswrapper[4716]: I1207 16:03:08.075071 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdee9957-48b9-458f-ba71-e0cd39f80596\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73f85248b4d68a9beecdacb3a01f433e97a0ffea2247ce92a41ef96855509377\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6276df652efe6ca935e745f24ef20f87a16465475139230c63b9999234a4d12a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3b1dd202586ee755ca89fbe12b30bbb12f95a62514f6e4d93cbbdbe55181e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd318647816e4a27849666b0ec4fc2ca4905295661c557dc381e1144f22cf284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d31de0326a05216fa10442bdb47151c6e073c97e9a16c34e793028e5e253756\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"le observer\\\\nW1207 16:02:34.645464 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1207 16:02:34.645587 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 16:02:34.646501 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2082283479/tls.crt::/tmp/serving-cert-2082283479/tls.key\\\\\\\"\\\\nI1207 16:02:34.972041 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 16:02:34.973702 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 16:02:34.973717 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 16:02:34.973738 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 16:02:34.973743 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 16:02:34.977185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1207 16:02:34.977204 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1207 16:02:34.977209 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977217 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 16:02:34.977225 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 16:02:34.977228 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 16:02:34.977232 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1207 16:02:34.979033 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27bd4a54f920b2725987c1c7fd2d9fdf16cd08738c246c333a4e9164a3a582a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:08Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:08 crc kubenswrapper[4716]: I1207 16:03:08.089385 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:08Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:08 crc kubenswrapper[4716]: I1207 16:03:08.097735 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:08 crc kubenswrapper[4716]: I1207 16:03:08.097774 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:08 crc kubenswrapper[4716]: I1207 16:03:08.097787 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:08 crc kubenswrapper[4716]: I1207 16:03:08.097804 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:08 crc kubenswrapper[4716]: I1207 16:03:08.097816 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:08Z","lastTransitionTime":"2025-12-07T16:03:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:08 crc kubenswrapper[4716]: I1207 16:03:08.101870 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2jjld" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dee6f5c-861c-4819-9309-7791574c7b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51c1d4159ab11087839a2c2d192566fb9322447c26df560bf09517c906ececb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmdns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2jjld\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:08Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:08 crc kubenswrapper[4716]: I1207 16:03:08.116500 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4wr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0baf8575-26cf-4a72-8a6d-ae94c078eaad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0ca2b3cd5a11033afd219b0125be7ab5acd750935d7eb89a470eff3d93a0838\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p2trd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4wr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:08Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:08 crc kubenswrapper[4716]: I1207 16:03:08.125722 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-m5jl5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"824ee59f-1474-4994-a70e-462878521514\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a15fe50cc75515704779646cc19aa50004f0db9aa26fe2535110df749bbfb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhwzn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-m5jl5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:08Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:08 crc kubenswrapper[4716]: I1207 16:03:08.142283 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef3335ae3875a75ab64685fc7e3590177280d651e8f8292c6f5ea6f472e55789\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:08Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:08 crc kubenswrapper[4716]: I1207 16:03:08.153274 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:08Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:08 crc kubenswrapper[4716]: I1207 16:03:08.163957 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f95ef61779507288a03bda8eddf745e105f705998a117a30789675bfa452d88f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b955b23163e7c8c458a3bbe2f681633e6a663ce76748febcc0b3a3c37f3ca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:08Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:08 crc kubenswrapper[4716]: I1207 16:03:08.178212 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8ps9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98015cea-d753-4e7a-b2b5-5ae7733a81e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d68509b673eecd37725cdcc48e3d0612c3b9c7601433f2b72c21d662140daa59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f84f52a073437543721010dc440339654ffb5e04eeb6cc7afbcc62f00f64c8e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f84f52a073437543721010dc440339654ffb5e04eeb6cc7afbcc62f00f64c8e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f6fe1cb25218cb3ac20488dc47efa4dc6a26a43dbc8306b5ee38cbe1a186c67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f6fe1cb25218cb3ac20488dc47efa4dc6a26a43dbc8306b5ee38cbe1a186c67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14b748a8247f6863e8be61e90057a06626dfd3cd4f01bd3a11629f700f56378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f14b748a8247f6863e8be61e90057a06626dfd3cd4f01bd3a11629f700f56378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ec379c0d573cb9b5ffcfc5ca7c653ba60e2fbdd7e3d37836102e26618ce43e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ec379c0d573cb9b5ffcfc5ca7c653ba60e2fbdd7e3d37836102e26618ce43e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c75bf2a6703cf99865d4ed5fa000e1e3b1330d4946176cf585fec1a2f226df0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c75bf2a6703cf99865d4ed5fa000e1e3b1330d4946176cf585fec1a2f226df0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de45812c9a5a5bfe6b0146828d03fab28362b03d04ef861c150b9ffbf181d345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de45812c9a5a5bfe6b0146828d03fab28362b03d04ef861c150b9ffbf181d345\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8ps9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:08Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:08 crc kubenswrapper[4716]: I1207 16:03:08.199102 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7961cc9a-734e-42c2-b1ff-7075f3f20091\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84a36b76f20618c7885d611ecad9db29bac5101c27572be31d3acbb90c581291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0a0eca3c2087a457c0e448e649345ec4595a9c8a019aed2cddc5f30c77871e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://404426a171b8676eddd4a00996c76afab7e269643fa9608f1e1b6bdba385071d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fd3c2941782921f99a6cdc3de4e892234548bde527fec15a8ac92e513566e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e23abfdca314ed43d594333c9f9e401cc8ef7dad39b8ebac84a836b06ee5c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09b38cd4b389ea477c802a842ee4f2297a64d64684624117163e819539c257c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09b38cd4b389ea477c802a842ee4f2297a64d64684624117163e819539c257c6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c00f7a8e024e0e0dfd7b73e16d1d5437602e79ea95b73c21f24b209edef60e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c00f7a8e024e0e0dfd7b73e16d1d5437602e79ea95b73c21f24b209edef60e98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3f6392d426fa2cd788ec3a5c3bff4ac39483a54dfb2624a1c4370522413a350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3f6392d426fa2cd788ec3a5c3bff4ac39483a54dfb2624a1c4370522413a350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:08Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:08 crc kubenswrapper[4716]: I1207 16:03:08.200331 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:08 crc kubenswrapper[4716]: I1207 16:03:08.200358 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:08 crc kubenswrapper[4716]: I1207 16:03:08.200367 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:08 crc kubenswrapper[4716]: I1207 16:03:08.200379 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:08 crc kubenswrapper[4716]: I1207 16:03:08.200388 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:08Z","lastTransitionTime":"2025-12-07T16:03:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:08 crc kubenswrapper[4716]: I1207 16:03:08.213178 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3067de7c-c11c-4be1-bd94-d6572c266795\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d14d018fa044e9e80fe81e7e38df935992d14c773ed2b2193fd5cb873e5d0947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a9e9cf048220ca14c996847e9ced9f2367b9331dc6c2836dd023658855cd79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6d3c911874095b3c7cb062d03070e5090a7bb2456cb20cb4ce078f502515cda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ac6ce7701f1c6aa1ea8fcf7a3035e3b0eaa28ae3cb6afc442372bd9117bdde4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:08Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:08 crc kubenswrapper[4716]: I1207 16:03:08.225536 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4742ea84b3a2c9461d9b900beee7354a2ded24c39ad23e007cbf270ae12377c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:08Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:08 crc kubenswrapper[4716]: I1207 16:03:08.239459 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:08Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:08 crc kubenswrapper[4716]: I1207 16:03:08.249797 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c15b59eb-565d-4556-a4ce-75afdf159dc8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://530a2935826940767ea33fb4f37fa11f7dee4950d867fbcd6af72836d80f7db3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jks8z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ccaf6c1e6a978eb978a1c207b6c4c00b23737f78b743f966579aca781e22e0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jks8z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zcxxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:08Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:08 crc kubenswrapper[4716]: I1207 16:03:08.302448 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:08 crc kubenswrapper[4716]: I1207 16:03:08.302481 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:08 crc kubenswrapper[4716]: I1207 16:03:08.302488 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:08 crc kubenswrapper[4716]: I1207 16:03:08.302501 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:08 crc kubenswrapper[4716]: I1207 16:03:08.302510 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:08Z","lastTransitionTime":"2025-12-07T16:03:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:08 crc kubenswrapper[4716]: I1207 16:03:08.405142 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:08 crc kubenswrapper[4716]: I1207 16:03:08.405178 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:08 crc kubenswrapper[4716]: I1207 16:03:08.405187 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:08 crc kubenswrapper[4716]: I1207 16:03:08.405201 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:08 crc kubenswrapper[4716]: I1207 16:03:08.405212 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:08Z","lastTransitionTime":"2025-12-07T16:03:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:08 crc kubenswrapper[4716]: I1207 16:03:08.507010 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:08 crc kubenswrapper[4716]: I1207 16:03:08.507048 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:08 crc kubenswrapper[4716]: I1207 16:03:08.507058 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:08 crc kubenswrapper[4716]: I1207 16:03:08.507086 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:08 crc kubenswrapper[4716]: I1207 16:03:08.507097 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:08Z","lastTransitionTime":"2025-12-07T16:03:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:08 crc kubenswrapper[4716]: I1207 16:03:08.609880 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:08 crc kubenswrapper[4716]: I1207 16:03:08.609927 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:08 crc kubenswrapper[4716]: I1207 16:03:08.609948 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:08 crc kubenswrapper[4716]: I1207 16:03:08.609975 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:08 crc kubenswrapper[4716]: I1207 16:03:08.609993 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:08Z","lastTransitionTime":"2025-12-07T16:03:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:08 crc kubenswrapper[4716]: I1207 16:03:08.657576 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 16:03:08 crc kubenswrapper[4716]: E1207 16:03:08.657769 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 16:03:08 crc kubenswrapper[4716]: I1207 16:03:08.712604 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:08 crc kubenswrapper[4716]: I1207 16:03:08.712635 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:08 crc kubenswrapper[4716]: I1207 16:03:08.712643 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:08 crc kubenswrapper[4716]: I1207 16:03:08.712656 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:08 crc kubenswrapper[4716]: I1207 16:03:08.712665 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:08Z","lastTransitionTime":"2025-12-07T16:03:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:08 crc kubenswrapper[4716]: I1207 16:03:08.815333 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:08 crc kubenswrapper[4716]: I1207 16:03:08.815395 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:08 crc kubenswrapper[4716]: I1207 16:03:08.815419 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:08 crc kubenswrapper[4716]: I1207 16:03:08.815447 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:08 crc kubenswrapper[4716]: I1207 16:03:08.815467 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:08Z","lastTransitionTime":"2025-12-07T16:03:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:08 crc kubenswrapper[4716]: I1207 16:03:08.918169 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:08 crc kubenswrapper[4716]: I1207 16:03:08.918213 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:08 crc kubenswrapper[4716]: I1207 16:03:08.918228 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:08 crc kubenswrapper[4716]: I1207 16:03:08.918247 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:08 crc kubenswrapper[4716]: I1207 16:03:08.918261 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:08Z","lastTransitionTime":"2025-12-07T16:03:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.011134 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qvc7p_c54d618a-dfec-4f2a-a288-f5188c8a128a/ovnkube-controller/2.log" Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.012235 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qvc7p_c54d618a-dfec-4f2a-a288-f5188c8a128a/ovnkube-controller/1.log" Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.015049 4716 generic.go:334] "Generic (PLEG): container finished" podID="c54d618a-dfec-4f2a-a288-f5188c8a128a" containerID="f48167b38ff840f85f11b69d8090ec6c9c634ed26682daef93f5da99fc02ef61" exitCode=1 Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.015134 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" event={"ID":"c54d618a-dfec-4f2a-a288-f5188c8a128a","Type":"ContainerDied","Data":"f48167b38ff840f85f11b69d8090ec6c9c634ed26682daef93f5da99fc02ef61"} Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.015182 4716 scope.go:117] "RemoveContainer" containerID="a8baee0f3d1e686bc2c17b0777f4eacc31c7f1fe44720e68f94612e6c806bc8b" Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.016289 4716 scope.go:117] "RemoveContainer" containerID="f48167b38ff840f85f11b69d8090ec6c9c634ed26682daef93f5da99fc02ef61" Dec 07 16:03:09 crc kubenswrapper[4716]: E1207 16:03:09.016561 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qvc7p_openshift-ovn-kubernetes(c54d618a-dfec-4f2a-a288-f5188c8a128a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" podUID="c54d618a-dfec-4f2a-a288-f5188c8a128a" Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.020996 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.021039 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.021058 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.021119 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.021139 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:09Z","lastTransitionTime":"2025-12-07T16:03:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.031596 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nx2lh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04b37f6c-457b-4f9a-8733-e88cb9644483\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv6td\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv6td\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nx2lh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:09Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.046507 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e903f3c-ba8c-477d-b66a-6d1331a7c64e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ff9c456318bbdb54ec3f1ac3e2c2366968f18ad084158b7100d142acb668948\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1358745250b35426602fc7417e4f85dd411cdd0fb3e8f703c432d72128ca7c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d8c9a434c6cfb9e4fee6a8955f1379291f153c007876d1ffb498ce561c73ba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad8df9b5bd5f5a8dd44ce2899cfa29c19435149b170d339d339183cd3df09e99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad8df9b5bd5f5a8dd44ce2899cfa29c19435149b170d339d339183cd3df09e99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:09Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.077851 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c54d618a-dfec-4f2a-a288-f5188c8a128a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60324dc7d701bc064e14ae4f474bce10f6ee4ff3c3c942eab749e7166082cf9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ed4f4d3564674bf5cb4e7a76dc0fd95e203ecc8950798e59a1fb6c9ab4a20f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://def9ac840fb8ecf2778bed88ea980eddca7c5b2aba4bd1afb967d62fec3f7bae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b73be664f5412c98b89d971a2cac37c494151c293fdb44a98b390978658a6643\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a847d04e22e98091c0037a89054b9236228b701b8c7ba31940a56df0f37b95c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcae78a6133d97f6a32dd7421ccec4333b970a237404e056a5cc601f8eefdc8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f48167b38ff840f85f11b69d8090ec6c9c634ed26682daef93f5da99fc02ef61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8baee0f3d1e686bc2c17b0777f4eacc31c7f1fe44720e68f94612e6c806bc8b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-07T16:02:52Z\\\",\\\"message\\\":\\\"r occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:02:51Z is after 2025-08-24T17:21:41Z]\\\\nI1207 16:02:51.943802 6191 services_controller.go:444] Built service openshift-marketplace/redhat-marketplace LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1207 16:02:51.943764 6191 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-route-controller-manager/route-controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"18746a4d-8a63-458a-b7e3-8fb89ff95fc0\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-route-controller-manager/route-controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-route-controller-manager/route-controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f48167b38ff840f85f11b69d8090ec6c9c634ed26682daef93f5da99fc02ef61\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-07T16:03:08Z\\\",\\\"message\\\":\\\"er, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:08Z is after 2025-08-24T17:21:41Z]\\\\nI1207 16:03:08.453903 6420 services_controller.go:451] Built service openshift-controller-manager-operator/metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-controller-manager-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-controller-manager-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.58\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.Tem\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T16:03:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcfdffe822635b0ae13b253e6133449fe00461b2792be43d913c4a77219b8255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9bfe2ffce7a7b7c320474117734fc05246cc2b4fd744833e896e8e7c83f5e4d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9bfe2ffce7a7b7c320474117734fc05246cc2b4fd744833e896e8e7c83f5e4d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qvc7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:09Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.090063 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mx85f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"118dd9ba-437c-4e5c-b5ff-4239ae6d4a63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bcd6c8deb184d0c713c5e7b7cc82fa38d3cd498517911ba7a0ec3cb88a3437a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gg5nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ffc069d2b21fa80871b1b97f2c72c37880d9a309cbf8b7daf7d240f2c332a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gg5nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mx85f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:09Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.105578 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4wr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0baf8575-26cf-4a72-8a6d-ae94c078eaad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0ca2b3cd5a11033afd219b0125be7ab5acd750935d7eb89a470eff3d93a0838\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p2trd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4wr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:09Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.116843 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-m5jl5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"824ee59f-1474-4994-a70e-462878521514\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a15fe50cc75515704779646cc19aa50004f0db9aa26fe2535110df749bbfb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhwzn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-m5jl5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:09Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.123691 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.123742 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.123758 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.123780 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.123795 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:09Z","lastTransitionTime":"2025-12-07T16:03:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.132838 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdee9957-48b9-458f-ba71-e0cd39f80596\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73f85248b4d68a9beecdacb3a01f433e97a0ffea2247ce92a41ef96855509377\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6276df652efe6ca935e745f24ef20f87a16465475139230c63b9999234a4d12a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3b1dd202586ee755ca89fbe12b30bbb12f95a62514f6e4d93cbbdbe55181e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd318647816e4a27849666b0ec4fc2ca4905295661c557dc381e1144f22cf284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d31de0326a05216fa10442bdb47151c6e073c97e9a16c34e793028e5e253756\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"le observer\\\\nW1207 16:02:34.645464 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1207 16:02:34.645587 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 16:02:34.646501 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2082283479/tls.crt::/tmp/serving-cert-2082283479/tls.key\\\\\\\"\\\\nI1207 16:02:34.972041 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 16:02:34.973702 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 16:02:34.973717 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 16:02:34.973738 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 16:02:34.973743 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 16:02:34.977185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1207 16:02:34.977204 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1207 16:02:34.977209 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977217 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 16:02:34.977225 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 16:02:34.977228 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 16:02:34.977232 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1207 16:02:34.979033 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27bd4a54f920b2725987c1c7fd2d9fdf16cd08738c246c333a4e9164a3a582a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:09Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.147850 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:09Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.160717 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2jjld" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dee6f5c-861c-4819-9309-7791574c7b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51c1d4159ab11087839a2c2d192566fb9322447c26df560bf09517c906ececb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmdns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2jjld\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:09Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.175295 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f95ef61779507288a03bda8eddf745e105f705998a117a30789675bfa452d88f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b955b23163e7c8c458a3bbe2f681633e6a663ce76748febcc0b3a3c37f3ca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:09Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.193145 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef3335ae3875a75ab64685fc7e3590177280d651e8f8292c6f5ea6f472e55789\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:09Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.211000 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:09Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.225447 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:09Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.226186 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.226242 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.226257 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.226279 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.226297 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:09Z","lastTransitionTime":"2025-12-07T16:03:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.241613 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c15b59eb-565d-4556-a4ce-75afdf159dc8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://530a2935826940767ea33fb4f37fa11f7dee4950d867fbcd6af72836d80f7db3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jks8z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ccaf6c1e6a978eb978a1c207b6c4c00b23737f78b743f966579aca781e22e0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jks8z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zcxxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:09Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.259371 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8ps9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98015cea-d753-4e7a-b2b5-5ae7733a81e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d68509b673eecd37725cdcc48e3d0612c3b9c7601433f2b72c21d662140daa59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f84f52a073437543721010dc440339654ffb5e04eeb6cc7afbcc62f00f64c8e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f84f52a073437543721010dc440339654ffb5e04eeb6cc7afbcc62f00f64c8e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f6fe1cb25218cb3ac20488dc47efa4dc6a26a43dbc8306b5ee38cbe1a186c67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f6fe1cb25218cb3ac20488dc47efa4dc6a26a43dbc8306b5ee38cbe1a186c67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14b748a8247f6863e8be61e90057a06626dfd3cd4f01bd3a11629f700f56378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f14b748a8247f6863e8be61e90057a06626dfd3cd4f01bd3a11629f700f56378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ec379c0d573cb9b5ffcfc5ca7c653ba60e2fbdd7e3d37836102e26618ce43e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ec379c0d573cb9b5ffcfc5ca7c653ba60e2fbdd7e3d37836102e26618ce43e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c75bf2a6703cf99865d4ed5fa000e1e3b1330d4946176cf585fec1a2f226df0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c75bf2a6703cf99865d4ed5fa000e1e3b1330d4946176cf585fec1a2f226df0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de45812c9a5a5bfe6b0146828d03fab28362b03d04ef861c150b9ffbf181d345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de45812c9a5a5bfe6b0146828d03fab28362b03d04ef861c150b9ffbf181d345\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8ps9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:09Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.281675 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7961cc9a-734e-42c2-b1ff-7075f3f20091\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84a36b76f20618c7885d611ecad9db29bac5101c27572be31d3acbb90c581291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0a0eca3c2087a457c0e448e649345ec4595a9c8a019aed2cddc5f30c77871e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://404426a171b8676eddd4a00996c76afab7e269643fa9608f1e1b6bdba385071d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fd3c2941782921f99a6cdc3de4e892234548bde527fec15a8ac92e513566e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e23abfdca314ed43d594333c9f9e401cc8ef7dad39b8ebac84a836b06ee5c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09b38cd4b389ea477c802a842ee4f2297a64d64684624117163e819539c257c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09b38cd4b389ea477c802a842ee4f2297a64d64684624117163e819539c257c6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c00f7a8e024e0e0dfd7b73e16d1d5437602e79ea95b73c21f24b209edef60e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c00f7a8e024e0e0dfd7b73e16d1d5437602e79ea95b73c21f24b209edef60e98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3f6392d426fa2cd788ec3a5c3bff4ac39483a54dfb2624a1c4370522413a350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3f6392d426fa2cd788ec3a5c3bff4ac39483a54dfb2624a1c4370522413a350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:09Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.294800 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3067de7c-c11c-4be1-bd94-d6572c266795\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d14d018fa044e9e80fe81e7e38df935992d14c773ed2b2193fd5cb873e5d0947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a9e9cf048220ca14c996847e9ced9f2367b9331dc6c2836dd023658855cd79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6d3c911874095b3c7cb062d03070e5090a7bb2456cb20cb4ce078f502515cda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ac6ce7701f1c6aa1ea8fcf7a3035e3b0eaa28ae3cb6afc442372bd9117bdde4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:09Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.306775 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4742ea84b3a2c9461d9b900beee7354a2ded24c39ad23e007cbf270ae12377c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:09Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.329238 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.329278 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.329289 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.329309 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.329321 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:09Z","lastTransitionTime":"2025-12-07T16:03:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.431893 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.431938 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.431954 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.431970 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.431979 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:09Z","lastTransitionTime":"2025-12-07T16:03:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.534254 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.534286 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.534294 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.534306 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.534314 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:09Z","lastTransitionTime":"2025-12-07T16:03:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.636460 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.636519 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.636576 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.636604 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.636622 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:09Z","lastTransitionTime":"2025-12-07T16:03:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.657210 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.657225 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nx2lh" Dec 07 16:03:09 crc kubenswrapper[4716]: E1207 16:03:09.657411 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.657226 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:03:09 crc kubenswrapper[4716]: E1207 16:03:09.657475 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nx2lh" podUID="04b37f6c-457b-4f9a-8733-e88cb9644483" Dec 07 16:03:09 crc kubenswrapper[4716]: E1207 16:03:09.657682 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.738973 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.739041 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.739061 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.739119 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.739138 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:09Z","lastTransitionTime":"2025-12-07T16:03:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.842232 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.842304 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.842329 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.842361 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.842385 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:09Z","lastTransitionTime":"2025-12-07T16:03:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.946900 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.946947 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.946962 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.946985 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:09 crc kubenswrapper[4716]: I1207 16:03:09.947001 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:09Z","lastTransitionTime":"2025-12-07T16:03:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:10 crc kubenswrapper[4716]: I1207 16:03:10.020864 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qvc7p_c54d618a-dfec-4f2a-a288-f5188c8a128a/ovnkube-controller/2.log" Dec 07 16:03:10 crc kubenswrapper[4716]: I1207 16:03:10.024887 4716 scope.go:117] "RemoveContainer" containerID="f48167b38ff840f85f11b69d8090ec6c9c634ed26682daef93f5da99fc02ef61" Dec 07 16:03:10 crc kubenswrapper[4716]: E1207 16:03:10.025195 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qvc7p_openshift-ovn-kubernetes(c54d618a-dfec-4f2a-a288-f5188c8a128a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" podUID="c54d618a-dfec-4f2a-a288-f5188c8a128a" Dec 07 16:03:10 crc kubenswrapper[4716]: I1207 16:03:10.045876 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/04b37f6c-457b-4f9a-8733-e88cb9644483-metrics-certs\") pod \"network-metrics-daemon-nx2lh\" (UID: \"04b37f6c-457b-4f9a-8733-e88cb9644483\") " pod="openshift-multus/network-metrics-daemon-nx2lh" Dec 07 16:03:10 crc kubenswrapper[4716]: I1207 16:03:10.045722 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdee9957-48b9-458f-ba71-e0cd39f80596\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73f85248b4d68a9beecdacb3a01f433e97a0ffea2247ce92a41ef96855509377\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6276df652efe6ca935e745f24ef20f87a16465475139230c63b9999234a4d12a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3b1dd202586ee755ca89fbe12b30bbb12f95a62514f6e4d93cbbdbe55181e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd318647816e4a27849666b0ec4fc2ca4905295661c557dc381e1144f22cf284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d31de0326a05216fa10442bdb47151c6e073c97e9a16c34e793028e5e253756\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"le observer\\\\nW1207 16:02:34.645464 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1207 16:02:34.645587 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 16:02:34.646501 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2082283479/tls.crt::/tmp/serving-cert-2082283479/tls.key\\\\\\\"\\\\nI1207 16:02:34.972041 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 16:02:34.973702 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 16:02:34.973717 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 16:02:34.973738 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 16:02:34.973743 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 16:02:34.977185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1207 16:02:34.977204 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1207 16:02:34.977209 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977217 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 16:02:34.977225 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 16:02:34.977228 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 16:02:34.977232 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1207 16:02:34.979033 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27bd4a54f920b2725987c1c7fd2d9fdf16cd08738c246c333a4e9164a3a582a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:10Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:10 crc kubenswrapper[4716]: E1207 16:03:10.046066 4716 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 07 16:03:10 crc kubenswrapper[4716]: E1207 16:03:10.046236 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/04b37f6c-457b-4f9a-8733-e88cb9644483-metrics-certs podName:04b37f6c-457b-4f9a-8733-e88cb9644483 nodeName:}" failed. No retries permitted until 2025-12-07 16:03:26.046173643 +0000 UTC m=+68.736458595 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/04b37f6c-457b-4f9a-8733-e88cb9644483-metrics-certs") pod "network-metrics-daemon-nx2lh" (UID: "04b37f6c-457b-4f9a-8733-e88cb9644483") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 07 16:03:10 crc kubenswrapper[4716]: I1207 16:03:10.049587 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:10 crc kubenswrapper[4716]: I1207 16:03:10.049636 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:10 crc kubenswrapper[4716]: I1207 16:03:10.049647 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:10 crc kubenswrapper[4716]: I1207 16:03:10.049667 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:10 crc kubenswrapper[4716]: I1207 16:03:10.049679 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:10Z","lastTransitionTime":"2025-12-07T16:03:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:10 crc kubenswrapper[4716]: I1207 16:03:10.063524 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:10Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:10 crc kubenswrapper[4716]: I1207 16:03:10.077344 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2jjld" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dee6f5c-861c-4819-9309-7791574c7b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51c1d4159ab11087839a2c2d192566fb9322447c26df560bf09517c906ececb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmdns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2jjld\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:10Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:10 crc kubenswrapper[4716]: I1207 16:03:10.095614 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4wr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0baf8575-26cf-4a72-8a6d-ae94c078eaad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0ca2b3cd5a11033afd219b0125be7ab5acd750935d7eb89a470eff3d93a0838\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p2trd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4wr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:10Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:10 crc kubenswrapper[4716]: I1207 16:03:10.109849 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-m5jl5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"824ee59f-1474-4994-a70e-462878521514\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a15fe50cc75515704779646cc19aa50004f0db9aa26fe2535110df749bbfb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhwzn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-m5jl5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:10Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:10 crc kubenswrapper[4716]: I1207 16:03:10.128001 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef3335ae3875a75ab64685fc7e3590177280d651e8f8292c6f5ea6f472e55789\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:10Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:10 crc kubenswrapper[4716]: I1207 16:03:10.146459 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:10Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:10 crc kubenswrapper[4716]: I1207 16:03:10.152664 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:10 crc kubenswrapper[4716]: I1207 16:03:10.152732 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:10 crc kubenswrapper[4716]: I1207 16:03:10.152759 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:10 crc kubenswrapper[4716]: I1207 16:03:10.152789 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:10 crc kubenswrapper[4716]: I1207 16:03:10.152810 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:10Z","lastTransitionTime":"2025-12-07T16:03:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:10 crc kubenswrapper[4716]: I1207 16:03:10.164819 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f95ef61779507288a03bda8eddf745e105f705998a117a30789675bfa452d88f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b955b23163e7c8c458a3bbe2f681633e6a663ce76748febcc0b3a3c37f3ca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:10Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:10 crc kubenswrapper[4716]: I1207 16:03:10.186519 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8ps9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98015cea-d753-4e7a-b2b5-5ae7733a81e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d68509b673eecd37725cdcc48e3d0612c3b9c7601433f2b72c21d662140daa59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f84f52a073437543721010dc440339654ffb5e04eeb6cc7afbcc62f00f64c8e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f84f52a073437543721010dc440339654ffb5e04eeb6cc7afbcc62f00f64c8e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f6fe1cb25218cb3ac20488dc47efa4dc6a26a43dbc8306b5ee38cbe1a186c67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f6fe1cb25218cb3ac20488dc47efa4dc6a26a43dbc8306b5ee38cbe1a186c67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14b748a8247f6863e8be61e90057a06626dfd3cd4f01bd3a11629f700f56378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f14b748a8247f6863e8be61e90057a06626dfd3cd4f01bd3a11629f700f56378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ec379c0d573cb9b5ffcfc5ca7c653ba60e2fbdd7e3d37836102e26618ce43e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ec379c0d573cb9b5ffcfc5ca7c653ba60e2fbdd7e3d37836102e26618ce43e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c75bf2a6703cf99865d4ed5fa000e1e3b1330d4946176cf585fec1a2f226df0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c75bf2a6703cf99865d4ed5fa000e1e3b1330d4946176cf585fec1a2f226df0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de45812c9a5a5bfe6b0146828d03fab28362b03d04ef861c150b9ffbf181d345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de45812c9a5a5bfe6b0146828d03fab28362b03d04ef861c150b9ffbf181d345\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8ps9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:10Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:10 crc kubenswrapper[4716]: I1207 16:03:10.255740 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:10 crc kubenswrapper[4716]: I1207 16:03:10.255795 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:10 crc kubenswrapper[4716]: I1207 16:03:10.255808 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:10 crc kubenswrapper[4716]: I1207 16:03:10.255825 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:10 crc kubenswrapper[4716]: I1207 16:03:10.255836 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:10Z","lastTransitionTime":"2025-12-07T16:03:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:10 crc kubenswrapper[4716]: I1207 16:03:10.280637 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7961cc9a-734e-42c2-b1ff-7075f3f20091\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84a36b76f20618c7885d611ecad9db29bac5101c27572be31d3acbb90c581291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0a0eca3c2087a457c0e448e649345ec4595a9c8a019aed2cddc5f30c77871e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://404426a171b8676eddd4a00996c76afab7e269643fa9608f1e1b6bdba385071d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fd3c2941782921f99a6cdc3de4e892234548bde527fec15a8ac92e513566e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e23abfdca314ed43d594333c9f9e401cc8ef7dad39b8ebac84a836b06ee5c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09b38cd4b389ea477c802a842ee4f2297a64d64684624117163e819539c257c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09b38cd4b389ea477c802a842ee4f2297a64d64684624117163e819539c257c6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c00f7a8e024e0e0dfd7b73e16d1d5437602e79ea95b73c21f24b209edef60e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c00f7a8e024e0e0dfd7b73e16d1d5437602e79ea95b73c21f24b209edef60e98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3f6392d426fa2cd788ec3a5c3bff4ac39483a54dfb2624a1c4370522413a350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3f6392d426fa2cd788ec3a5c3bff4ac39483a54dfb2624a1c4370522413a350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:10Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:10 crc kubenswrapper[4716]: I1207 16:03:10.296900 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3067de7c-c11c-4be1-bd94-d6572c266795\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d14d018fa044e9e80fe81e7e38df935992d14c773ed2b2193fd5cb873e5d0947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a9e9cf048220ca14c996847e9ced9f2367b9331dc6c2836dd023658855cd79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6d3c911874095b3c7cb062d03070e5090a7bb2456cb20cb4ce078f502515cda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ac6ce7701f1c6aa1ea8fcf7a3035e3b0eaa28ae3cb6afc442372bd9117bdde4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:10Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:10 crc kubenswrapper[4716]: I1207 16:03:10.315258 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4742ea84b3a2c9461d9b900beee7354a2ded24c39ad23e007cbf270ae12377c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:10Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:10 crc kubenswrapper[4716]: I1207 16:03:10.333733 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:10Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:10 crc kubenswrapper[4716]: I1207 16:03:10.346892 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c15b59eb-565d-4556-a4ce-75afdf159dc8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://530a2935826940767ea33fb4f37fa11f7dee4950d867fbcd6af72836d80f7db3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jks8z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ccaf6c1e6a978eb978a1c207b6c4c00b23737f78b743f966579aca781e22e0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jks8z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zcxxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:10Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:10 crc kubenswrapper[4716]: I1207 16:03:10.358271 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:10 crc kubenswrapper[4716]: I1207 16:03:10.358297 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:10 crc kubenswrapper[4716]: I1207 16:03:10.358305 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:10 crc kubenswrapper[4716]: I1207 16:03:10.358318 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:10 crc kubenswrapper[4716]: I1207 16:03:10.358327 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:10Z","lastTransitionTime":"2025-12-07T16:03:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:10 crc kubenswrapper[4716]: I1207 16:03:10.360274 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e903f3c-ba8c-477d-b66a-6d1331a7c64e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ff9c456318bbdb54ec3f1ac3e2c2366968f18ad084158b7100d142acb668948\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1358745250b35426602fc7417e4f85dd411cdd0fb3e8f703c432d72128ca7c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d8c9a434c6cfb9e4fee6a8955f1379291f153c007876d1ffb498ce561c73ba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad8df9b5bd5f5a8dd44ce2899cfa29c19435149b170d339d339183cd3df09e99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad8df9b5bd5f5a8dd44ce2899cfa29c19435149b170d339d339183cd3df09e99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:10Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:10 crc kubenswrapper[4716]: I1207 16:03:10.376531 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c54d618a-dfec-4f2a-a288-f5188c8a128a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60324dc7d701bc064e14ae4f474bce10f6ee4ff3c3c942eab749e7166082cf9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ed4f4d3564674bf5cb4e7a76dc0fd95e203ecc8950798e59a1fb6c9ab4a20f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://def9ac840fb8ecf2778bed88ea980eddca7c5b2aba4bd1afb967d62fec3f7bae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b73be664f5412c98b89d971a2cac37c494151c293fdb44a98b390978658a6643\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a847d04e22e98091c0037a89054b9236228b701b8c7ba31940a56df0f37b95c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcae78a6133d97f6a32dd7421ccec4333b970a237404e056a5cc601f8eefdc8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f48167b38ff840f85f11b69d8090ec6c9c634ed26682daef93f5da99fc02ef61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f48167b38ff840f85f11b69d8090ec6c9c634ed26682daef93f5da99fc02ef61\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-07T16:03:08Z\\\",\\\"message\\\":\\\"er, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:08Z is after 2025-08-24T17:21:41Z]\\\\nI1207 16:03:08.453903 6420 services_controller.go:451] Built service openshift-controller-manager-operator/metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-controller-manager-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-controller-manager-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.58\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.Tem\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T16:03:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qvc7p_openshift-ovn-kubernetes(c54d618a-dfec-4f2a-a288-f5188c8a128a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcfdffe822635b0ae13b253e6133449fe00461b2792be43d913c4a77219b8255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9bfe2ffce7a7b7c320474117734fc05246cc2b4fd744833e896e8e7c83f5e4d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9bfe2ffce7a7b7c320474117734fc05246cc2b4fd744833e896e8e7c83f5e4d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qvc7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:10Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:10 crc kubenswrapper[4716]: I1207 16:03:10.389491 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mx85f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"118dd9ba-437c-4e5c-b5ff-4239ae6d4a63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bcd6c8deb184d0c713c5e7b7cc82fa38d3cd498517911ba7a0ec3cb88a3437a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gg5nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ffc069d2b21fa80871b1b97f2c72c37880d9a309cbf8b7daf7d240f2c332a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gg5nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mx85f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:10Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:10 crc kubenswrapper[4716]: I1207 16:03:10.399532 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nx2lh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04b37f6c-457b-4f9a-8733-e88cb9644483\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv6td\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv6td\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nx2lh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:10Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:10 crc kubenswrapper[4716]: I1207 16:03:10.460948 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:10 crc kubenswrapper[4716]: I1207 16:03:10.460998 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:10 crc kubenswrapper[4716]: I1207 16:03:10.461016 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:10 crc kubenswrapper[4716]: I1207 16:03:10.461040 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:10 crc kubenswrapper[4716]: I1207 16:03:10.461057 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:10Z","lastTransitionTime":"2025-12-07T16:03:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:10 crc kubenswrapper[4716]: I1207 16:03:10.563272 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:10 crc kubenswrapper[4716]: I1207 16:03:10.563311 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:10 crc kubenswrapper[4716]: I1207 16:03:10.563322 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:10 crc kubenswrapper[4716]: I1207 16:03:10.563339 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:10 crc kubenswrapper[4716]: I1207 16:03:10.563350 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:10Z","lastTransitionTime":"2025-12-07T16:03:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:10 crc kubenswrapper[4716]: I1207 16:03:10.657053 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 16:03:10 crc kubenswrapper[4716]: E1207 16:03:10.657242 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 16:03:10 crc kubenswrapper[4716]: I1207 16:03:10.666142 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:10 crc kubenswrapper[4716]: I1207 16:03:10.666186 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:10 crc kubenswrapper[4716]: I1207 16:03:10.666199 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:10 crc kubenswrapper[4716]: I1207 16:03:10.666217 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:10 crc kubenswrapper[4716]: I1207 16:03:10.666228 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:10Z","lastTransitionTime":"2025-12-07T16:03:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:10 crc kubenswrapper[4716]: I1207 16:03:10.769038 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:10 crc kubenswrapper[4716]: I1207 16:03:10.769105 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:10 crc kubenswrapper[4716]: I1207 16:03:10.769121 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:10 crc kubenswrapper[4716]: I1207 16:03:10.769142 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:10 crc kubenswrapper[4716]: I1207 16:03:10.769158 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:10Z","lastTransitionTime":"2025-12-07T16:03:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:10 crc kubenswrapper[4716]: I1207 16:03:10.871061 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:10 crc kubenswrapper[4716]: I1207 16:03:10.871120 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:10 crc kubenswrapper[4716]: I1207 16:03:10.871134 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:10 crc kubenswrapper[4716]: I1207 16:03:10.871152 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:10 crc kubenswrapper[4716]: I1207 16:03:10.871164 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:10Z","lastTransitionTime":"2025-12-07T16:03:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:10 crc kubenswrapper[4716]: I1207 16:03:10.973729 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:10 crc kubenswrapper[4716]: I1207 16:03:10.973786 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:10 crc kubenswrapper[4716]: I1207 16:03:10.973804 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:10 crc kubenswrapper[4716]: I1207 16:03:10.973832 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:10 crc kubenswrapper[4716]: I1207 16:03:10.973850 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:10Z","lastTransitionTime":"2025-12-07T16:03:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:11 crc kubenswrapper[4716]: I1207 16:03:11.076467 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:11 crc kubenswrapper[4716]: I1207 16:03:11.076527 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:11 crc kubenswrapper[4716]: I1207 16:03:11.076546 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:11 crc kubenswrapper[4716]: I1207 16:03:11.076571 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:11 crc kubenswrapper[4716]: I1207 16:03:11.076590 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:11Z","lastTransitionTime":"2025-12-07T16:03:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:11 crc kubenswrapper[4716]: I1207 16:03:11.179192 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:11 crc kubenswrapper[4716]: I1207 16:03:11.179255 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:11 crc kubenswrapper[4716]: I1207 16:03:11.179274 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:11 crc kubenswrapper[4716]: I1207 16:03:11.179300 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:11 crc kubenswrapper[4716]: I1207 16:03:11.179319 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:11Z","lastTransitionTime":"2025-12-07T16:03:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:11 crc kubenswrapper[4716]: I1207 16:03:11.281948 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:11 crc kubenswrapper[4716]: I1207 16:03:11.282010 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:11 crc kubenswrapper[4716]: I1207 16:03:11.282032 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:11 crc kubenswrapper[4716]: I1207 16:03:11.282057 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:11 crc kubenswrapper[4716]: I1207 16:03:11.282074 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:11Z","lastTransitionTime":"2025-12-07T16:03:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:11 crc kubenswrapper[4716]: I1207 16:03:11.384940 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:11 crc kubenswrapper[4716]: I1207 16:03:11.384996 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:11 crc kubenswrapper[4716]: I1207 16:03:11.385014 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:11 crc kubenswrapper[4716]: I1207 16:03:11.385042 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:11 crc kubenswrapper[4716]: I1207 16:03:11.385063 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:11Z","lastTransitionTime":"2025-12-07T16:03:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:11 crc kubenswrapper[4716]: I1207 16:03:11.488011 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:11 crc kubenswrapper[4716]: I1207 16:03:11.488120 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:11 crc kubenswrapper[4716]: I1207 16:03:11.488144 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:11 crc kubenswrapper[4716]: I1207 16:03:11.488170 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:11 crc kubenswrapper[4716]: I1207 16:03:11.488226 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:11Z","lastTransitionTime":"2025-12-07T16:03:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:11 crc kubenswrapper[4716]: I1207 16:03:11.591219 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:11 crc kubenswrapper[4716]: I1207 16:03:11.591284 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:11 crc kubenswrapper[4716]: I1207 16:03:11.591301 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:11 crc kubenswrapper[4716]: I1207 16:03:11.591325 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:11 crc kubenswrapper[4716]: I1207 16:03:11.591344 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:11Z","lastTransitionTime":"2025-12-07T16:03:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:11 crc kubenswrapper[4716]: I1207 16:03:11.657478 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 16:03:11 crc kubenswrapper[4716]: I1207 16:03:11.657557 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nx2lh" Dec 07 16:03:11 crc kubenswrapper[4716]: I1207 16:03:11.657613 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:03:11 crc kubenswrapper[4716]: E1207 16:03:11.657859 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 16:03:11 crc kubenswrapper[4716]: E1207 16:03:11.657993 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nx2lh" podUID="04b37f6c-457b-4f9a-8733-e88cb9644483" Dec 07 16:03:11 crc kubenswrapper[4716]: E1207 16:03:11.658173 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 16:03:11 crc kubenswrapper[4716]: I1207 16:03:11.694278 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:11 crc kubenswrapper[4716]: I1207 16:03:11.694372 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:11 crc kubenswrapper[4716]: I1207 16:03:11.694403 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:11 crc kubenswrapper[4716]: I1207 16:03:11.694435 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:11 crc kubenswrapper[4716]: I1207 16:03:11.694456 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:11Z","lastTransitionTime":"2025-12-07T16:03:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:11 crc kubenswrapper[4716]: I1207 16:03:11.797521 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:11 crc kubenswrapper[4716]: I1207 16:03:11.797592 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:11 crc kubenswrapper[4716]: I1207 16:03:11.797617 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:11 crc kubenswrapper[4716]: I1207 16:03:11.797648 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:11 crc kubenswrapper[4716]: I1207 16:03:11.797674 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:11Z","lastTransitionTime":"2025-12-07T16:03:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:11 crc kubenswrapper[4716]: I1207 16:03:11.900484 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:11 crc kubenswrapper[4716]: I1207 16:03:11.900528 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:11 crc kubenswrapper[4716]: I1207 16:03:11.900540 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:11 crc kubenswrapper[4716]: I1207 16:03:11.900556 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:11 crc kubenswrapper[4716]: I1207 16:03:11.900568 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:11Z","lastTransitionTime":"2025-12-07T16:03:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:12 crc kubenswrapper[4716]: I1207 16:03:12.003565 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:12 crc kubenswrapper[4716]: I1207 16:03:12.003641 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:12 crc kubenswrapper[4716]: I1207 16:03:12.003664 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:12 crc kubenswrapper[4716]: I1207 16:03:12.003696 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:12 crc kubenswrapper[4716]: I1207 16:03:12.003720 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:12Z","lastTransitionTime":"2025-12-07T16:03:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:12 crc kubenswrapper[4716]: I1207 16:03:12.106810 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:12 crc kubenswrapper[4716]: I1207 16:03:12.106853 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:12 crc kubenswrapper[4716]: I1207 16:03:12.106865 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:12 crc kubenswrapper[4716]: I1207 16:03:12.106888 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:12 crc kubenswrapper[4716]: I1207 16:03:12.106905 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:12Z","lastTransitionTime":"2025-12-07T16:03:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:12 crc kubenswrapper[4716]: I1207 16:03:12.210541 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:12 crc kubenswrapper[4716]: I1207 16:03:12.210586 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:12 crc kubenswrapper[4716]: I1207 16:03:12.210597 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:12 crc kubenswrapper[4716]: I1207 16:03:12.210614 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:12 crc kubenswrapper[4716]: I1207 16:03:12.210626 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:12Z","lastTransitionTime":"2025-12-07T16:03:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:12 crc kubenswrapper[4716]: I1207 16:03:12.313673 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:12 crc kubenswrapper[4716]: I1207 16:03:12.313715 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:12 crc kubenswrapper[4716]: I1207 16:03:12.313729 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:12 crc kubenswrapper[4716]: I1207 16:03:12.313750 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:12 crc kubenswrapper[4716]: I1207 16:03:12.313767 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:12Z","lastTransitionTime":"2025-12-07T16:03:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:12 crc kubenswrapper[4716]: I1207 16:03:12.420198 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:12 crc kubenswrapper[4716]: I1207 16:03:12.420266 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:12 crc kubenswrapper[4716]: I1207 16:03:12.420285 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:12 crc kubenswrapper[4716]: I1207 16:03:12.420309 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:12 crc kubenswrapper[4716]: I1207 16:03:12.420326 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:12Z","lastTransitionTime":"2025-12-07T16:03:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:12 crc kubenswrapper[4716]: I1207 16:03:12.523299 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:12 crc kubenswrapper[4716]: I1207 16:03:12.523357 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:12 crc kubenswrapper[4716]: I1207 16:03:12.523372 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:12 crc kubenswrapper[4716]: I1207 16:03:12.523392 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:12 crc kubenswrapper[4716]: I1207 16:03:12.523406 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:12Z","lastTransitionTime":"2025-12-07T16:03:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:12 crc kubenswrapper[4716]: I1207 16:03:12.625918 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:12 crc kubenswrapper[4716]: I1207 16:03:12.625953 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:12 crc kubenswrapper[4716]: I1207 16:03:12.625961 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:12 crc kubenswrapper[4716]: I1207 16:03:12.625975 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:12 crc kubenswrapper[4716]: I1207 16:03:12.625984 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:12Z","lastTransitionTime":"2025-12-07T16:03:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:12 crc kubenswrapper[4716]: I1207 16:03:12.657351 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 16:03:12 crc kubenswrapper[4716]: E1207 16:03:12.657445 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 16:03:12 crc kubenswrapper[4716]: I1207 16:03:12.728009 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:12 crc kubenswrapper[4716]: I1207 16:03:12.728039 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:12 crc kubenswrapper[4716]: I1207 16:03:12.728050 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:12 crc kubenswrapper[4716]: I1207 16:03:12.728064 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:12 crc kubenswrapper[4716]: I1207 16:03:12.728073 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:12Z","lastTransitionTime":"2025-12-07T16:03:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:12 crc kubenswrapper[4716]: I1207 16:03:12.830330 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:12 crc kubenswrapper[4716]: I1207 16:03:12.830376 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:12 crc kubenswrapper[4716]: I1207 16:03:12.830388 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:12 crc kubenswrapper[4716]: I1207 16:03:12.830408 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:12 crc kubenswrapper[4716]: I1207 16:03:12.830421 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:12Z","lastTransitionTime":"2025-12-07T16:03:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:12 crc kubenswrapper[4716]: I1207 16:03:12.932511 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:12 crc kubenswrapper[4716]: I1207 16:03:12.932550 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:12 crc kubenswrapper[4716]: I1207 16:03:12.932559 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:12 crc kubenswrapper[4716]: I1207 16:03:12.932572 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:12 crc kubenswrapper[4716]: I1207 16:03:12.932581 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:12Z","lastTransitionTime":"2025-12-07T16:03:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:13 crc kubenswrapper[4716]: I1207 16:03:13.034603 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:13 crc kubenswrapper[4716]: I1207 16:03:13.034697 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:13 crc kubenswrapper[4716]: I1207 16:03:13.034716 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:13 crc kubenswrapper[4716]: I1207 16:03:13.034776 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:13 crc kubenswrapper[4716]: I1207 16:03:13.034794 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:13Z","lastTransitionTime":"2025-12-07T16:03:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:13 crc kubenswrapper[4716]: I1207 16:03:13.136902 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:13 crc kubenswrapper[4716]: I1207 16:03:13.136956 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:13 crc kubenswrapper[4716]: I1207 16:03:13.136970 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:13 crc kubenswrapper[4716]: I1207 16:03:13.136988 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:13 crc kubenswrapper[4716]: I1207 16:03:13.137003 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:13Z","lastTransitionTime":"2025-12-07T16:03:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:13 crc kubenswrapper[4716]: I1207 16:03:13.240171 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:13 crc kubenswrapper[4716]: I1207 16:03:13.240243 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:13 crc kubenswrapper[4716]: I1207 16:03:13.240267 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:13 crc kubenswrapper[4716]: I1207 16:03:13.240296 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:13 crc kubenswrapper[4716]: I1207 16:03:13.240317 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:13Z","lastTransitionTime":"2025-12-07T16:03:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:13 crc kubenswrapper[4716]: I1207 16:03:13.342457 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:13 crc kubenswrapper[4716]: I1207 16:03:13.342481 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:13 crc kubenswrapper[4716]: I1207 16:03:13.342490 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:13 crc kubenswrapper[4716]: I1207 16:03:13.342502 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:13 crc kubenswrapper[4716]: I1207 16:03:13.342510 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:13Z","lastTransitionTime":"2025-12-07T16:03:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:13 crc kubenswrapper[4716]: I1207 16:03:13.444409 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:13 crc kubenswrapper[4716]: I1207 16:03:13.444432 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:13 crc kubenswrapper[4716]: I1207 16:03:13.444440 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:13 crc kubenswrapper[4716]: I1207 16:03:13.444451 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:13 crc kubenswrapper[4716]: I1207 16:03:13.444459 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:13Z","lastTransitionTime":"2025-12-07T16:03:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:13 crc kubenswrapper[4716]: I1207 16:03:13.546626 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:13 crc kubenswrapper[4716]: I1207 16:03:13.546660 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:13 crc kubenswrapper[4716]: I1207 16:03:13.546690 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:13 crc kubenswrapper[4716]: I1207 16:03:13.546706 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:13 crc kubenswrapper[4716]: I1207 16:03:13.546717 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:13Z","lastTransitionTime":"2025-12-07T16:03:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:13 crc kubenswrapper[4716]: I1207 16:03:13.649564 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:13 crc kubenswrapper[4716]: I1207 16:03:13.649600 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:13 crc kubenswrapper[4716]: I1207 16:03:13.649611 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:13 crc kubenswrapper[4716]: I1207 16:03:13.649626 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:13 crc kubenswrapper[4716]: I1207 16:03:13.649637 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:13Z","lastTransitionTime":"2025-12-07T16:03:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:13 crc kubenswrapper[4716]: I1207 16:03:13.656891 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 16:03:13 crc kubenswrapper[4716]: I1207 16:03:13.656902 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nx2lh" Dec 07 16:03:13 crc kubenswrapper[4716]: I1207 16:03:13.656946 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:03:13 crc kubenswrapper[4716]: E1207 16:03:13.657045 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 16:03:13 crc kubenswrapper[4716]: E1207 16:03:13.657140 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nx2lh" podUID="04b37f6c-457b-4f9a-8733-e88cb9644483" Dec 07 16:03:13 crc kubenswrapper[4716]: E1207 16:03:13.657212 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 16:03:13 crc kubenswrapper[4716]: I1207 16:03:13.751957 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:13 crc kubenswrapper[4716]: I1207 16:03:13.752014 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:13 crc kubenswrapper[4716]: I1207 16:03:13.752032 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:13 crc kubenswrapper[4716]: I1207 16:03:13.752054 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:13 crc kubenswrapper[4716]: I1207 16:03:13.752071 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:13Z","lastTransitionTime":"2025-12-07T16:03:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:13 crc kubenswrapper[4716]: I1207 16:03:13.854399 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:13 crc kubenswrapper[4716]: I1207 16:03:13.854439 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:13 crc kubenswrapper[4716]: I1207 16:03:13.854450 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:13 crc kubenswrapper[4716]: I1207 16:03:13.854466 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:13 crc kubenswrapper[4716]: I1207 16:03:13.854478 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:13Z","lastTransitionTime":"2025-12-07T16:03:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:13 crc kubenswrapper[4716]: I1207 16:03:13.956765 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:13 crc kubenswrapper[4716]: I1207 16:03:13.956803 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:13 crc kubenswrapper[4716]: I1207 16:03:13.956814 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:13 crc kubenswrapper[4716]: I1207 16:03:13.956827 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:13 crc kubenswrapper[4716]: I1207 16:03:13.956856 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:13Z","lastTransitionTime":"2025-12-07T16:03:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:14 crc kubenswrapper[4716]: I1207 16:03:14.059383 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:14 crc kubenswrapper[4716]: I1207 16:03:14.059435 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:14 crc kubenswrapper[4716]: I1207 16:03:14.059443 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:14 crc kubenswrapper[4716]: I1207 16:03:14.059456 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:14 crc kubenswrapper[4716]: I1207 16:03:14.059467 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:14Z","lastTransitionTime":"2025-12-07T16:03:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:14 crc kubenswrapper[4716]: I1207 16:03:14.162070 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:14 crc kubenswrapper[4716]: I1207 16:03:14.162160 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:14 crc kubenswrapper[4716]: I1207 16:03:14.162177 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:14 crc kubenswrapper[4716]: I1207 16:03:14.162201 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:14 crc kubenswrapper[4716]: I1207 16:03:14.162218 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:14Z","lastTransitionTime":"2025-12-07T16:03:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:14 crc kubenswrapper[4716]: I1207 16:03:14.264579 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:14 crc kubenswrapper[4716]: I1207 16:03:14.264637 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:14 crc kubenswrapper[4716]: I1207 16:03:14.264649 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:14 crc kubenswrapper[4716]: I1207 16:03:14.264667 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:14 crc kubenswrapper[4716]: I1207 16:03:14.264679 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:14Z","lastTransitionTime":"2025-12-07T16:03:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:14 crc kubenswrapper[4716]: I1207 16:03:14.366980 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:14 crc kubenswrapper[4716]: I1207 16:03:14.367021 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:14 crc kubenswrapper[4716]: I1207 16:03:14.367029 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:14 crc kubenswrapper[4716]: I1207 16:03:14.367045 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:14 crc kubenswrapper[4716]: I1207 16:03:14.367054 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:14Z","lastTransitionTime":"2025-12-07T16:03:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:14 crc kubenswrapper[4716]: I1207 16:03:14.469824 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:14 crc kubenswrapper[4716]: I1207 16:03:14.469875 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:14 crc kubenswrapper[4716]: I1207 16:03:14.469888 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:14 crc kubenswrapper[4716]: I1207 16:03:14.469905 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:14 crc kubenswrapper[4716]: I1207 16:03:14.469918 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:14Z","lastTransitionTime":"2025-12-07T16:03:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:14 crc kubenswrapper[4716]: I1207 16:03:14.572264 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:14 crc kubenswrapper[4716]: I1207 16:03:14.572313 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:14 crc kubenswrapper[4716]: I1207 16:03:14.572338 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:14 crc kubenswrapper[4716]: I1207 16:03:14.572355 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:14 crc kubenswrapper[4716]: I1207 16:03:14.572367 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:14Z","lastTransitionTime":"2025-12-07T16:03:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:14 crc kubenswrapper[4716]: I1207 16:03:14.657128 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 16:03:14 crc kubenswrapper[4716]: E1207 16:03:14.657285 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 16:03:14 crc kubenswrapper[4716]: I1207 16:03:14.674566 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:14 crc kubenswrapper[4716]: I1207 16:03:14.674628 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:14 crc kubenswrapper[4716]: I1207 16:03:14.674638 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:14 crc kubenswrapper[4716]: I1207 16:03:14.674653 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:14 crc kubenswrapper[4716]: I1207 16:03:14.674663 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:14Z","lastTransitionTime":"2025-12-07T16:03:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:14 crc kubenswrapper[4716]: I1207 16:03:14.777392 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:14 crc kubenswrapper[4716]: I1207 16:03:14.777467 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:14 crc kubenswrapper[4716]: I1207 16:03:14.777489 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:14 crc kubenswrapper[4716]: I1207 16:03:14.777514 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:14 crc kubenswrapper[4716]: I1207 16:03:14.777532 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:14Z","lastTransitionTime":"2025-12-07T16:03:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:14 crc kubenswrapper[4716]: I1207 16:03:14.880069 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:14 crc kubenswrapper[4716]: I1207 16:03:14.880121 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:14 crc kubenswrapper[4716]: I1207 16:03:14.880130 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:14 crc kubenswrapper[4716]: I1207 16:03:14.880146 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:14 crc kubenswrapper[4716]: I1207 16:03:14.880155 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:14Z","lastTransitionTime":"2025-12-07T16:03:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:14 crc kubenswrapper[4716]: I1207 16:03:14.983215 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:14 crc kubenswrapper[4716]: I1207 16:03:14.983272 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:14 crc kubenswrapper[4716]: I1207 16:03:14.983292 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:14 crc kubenswrapper[4716]: I1207 16:03:14.983322 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:14 crc kubenswrapper[4716]: I1207 16:03:14.983344 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:14Z","lastTransitionTime":"2025-12-07T16:03:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:15 crc kubenswrapper[4716]: I1207 16:03:15.085856 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:15 crc kubenswrapper[4716]: I1207 16:03:15.085890 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:15 crc kubenswrapper[4716]: I1207 16:03:15.085898 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:15 crc kubenswrapper[4716]: I1207 16:03:15.085913 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:15 crc kubenswrapper[4716]: I1207 16:03:15.085922 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:15Z","lastTransitionTime":"2025-12-07T16:03:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:15 crc kubenswrapper[4716]: I1207 16:03:15.188846 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:15 crc kubenswrapper[4716]: I1207 16:03:15.188898 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:15 crc kubenswrapper[4716]: I1207 16:03:15.188914 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:15 crc kubenswrapper[4716]: I1207 16:03:15.188936 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:15 crc kubenswrapper[4716]: I1207 16:03:15.188952 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:15Z","lastTransitionTime":"2025-12-07T16:03:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:15 crc kubenswrapper[4716]: I1207 16:03:15.291382 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:15 crc kubenswrapper[4716]: I1207 16:03:15.291439 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:15 crc kubenswrapper[4716]: I1207 16:03:15.291490 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:15 crc kubenswrapper[4716]: I1207 16:03:15.291512 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:15 crc kubenswrapper[4716]: I1207 16:03:15.291530 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:15Z","lastTransitionTime":"2025-12-07T16:03:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:15 crc kubenswrapper[4716]: I1207 16:03:15.394228 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:15 crc kubenswrapper[4716]: I1207 16:03:15.394274 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:15 crc kubenswrapper[4716]: I1207 16:03:15.394286 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:15 crc kubenswrapper[4716]: I1207 16:03:15.394304 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:15 crc kubenswrapper[4716]: I1207 16:03:15.394318 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:15Z","lastTransitionTime":"2025-12-07T16:03:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:15 crc kubenswrapper[4716]: I1207 16:03:15.496908 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:15 crc kubenswrapper[4716]: I1207 16:03:15.496948 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:15 crc kubenswrapper[4716]: I1207 16:03:15.496957 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:15 crc kubenswrapper[4716]: I1207 16:03:15.496974 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:15 crc kubenswrapper[4716]: I1207 16:03:15.496987 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:15Z","lastTransitionTime":"2025-12-07T16:03:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:15 crc kubenswrapper[4716]: I1207 16:03:15.600274 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:15 crc kubenswrapper[4716]: I1207 16:03:15.600311 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:15 crc kubenswrapper[4716]: I1207 16:03:15.600320 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:15 crc kubenswrapper[4716]: I1207 16:03:15.600337 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:15 crc kubenswrapper[4716]: I1207 16:03:15.600348 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:15Z","lastTransitionTime":"2025-12-07T16:03:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:15 crc kubenswrapper[4716]: I1207 16:03:15.657287 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 16:03:15 crc kubenswrapper[4716]: I1207 16:03:15.657363 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:03:15 crc kubenswrapper[4716]: I1207 16:03:15.657287 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nx2lh" Dec 07 16:03:15 crc kubenswrapper[4716]: E1207 16:03:15.657472 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 16:03:15 crc kubenswrapper[4716]: E1207 16:03:15.657570 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 16:03:15 crc kubenswrapper[4716]: E1207 16:03:15.657824 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nx2lh" podUID="04b37f6c-457b-4f9a-8733-e88cb9644483" Dec 07 16:03:15 crc kubenswrapper[4716]: I1207 16:03:15.702673 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:15 crc kubenswrapper[4716]: I1207 16:03:15.702710 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:15 crc kubenswrapper[4716]: I1207 16:03:15.702718 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:15 crc kubenswrapper[4716]: I1207 16:03:15.702734 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:15 crc kubenswrapper[4716]: I1207 16:03:15.702744 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:15Z","lastTransitionTime":"2025-12-07T16:03:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:15 crc kubenswrapper[4716]: I1207 16:03:15.806410 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:15 crc kubenswrapper[4716]: I1207 16:03:15.806450 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:15 crc kubenswrapper[4716]: I1207 16:03:15.806459 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:15 crc kubenswrapper[4716]: I1207 16:03:15.806475 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:15 crc kubenswrapper[4716]: I1207 16:03:15.806485 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:15Z","lastTransitionTime":"2025-12-07T16:03:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:15 crc kubenswrapper[4716]: I1207 16:03:15.909245 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:15 crc kubenswrapper[4716]: I1207 16:03:15.909292 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:15 crc kubenswrapper[4716]: I1207 16:03:15.909306 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:15 crc kubenswrapper[4716]: I1207 16:03:15.909328 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:15 crc kubenswrapper[4716]: I1207 16:03:15.909344 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:15Z","lastTransitionTime":"2025-12-07T16:03:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:16 crc kubenswrapper[4716]: I1207 16:03:16.011448 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:16 crc kubenswrapper[4716]: I1207 16:03:16.011493 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:16 crc kubenswrapper[4716]: I1207 16:03:16.011503 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:16 crc kubenswrapper[4716]: I1207 16:03:16.011520 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:16 crc kubenswrapper[4716]: I1207 16:03:16.011532 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:16Z","lastTransitionTime":"2025-12-07T16:03:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:16 crc kubenswrapper[4716]: I1207 16:03:16.113796 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:16 crc kubenswrapper[4716]: I1207 16:03:16.113858 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:16 crc kubenswrapper[4716]: I1207 16:03:16.113872 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:16 crc kubenswrapper[4716]: I1207 16:03:16.113888 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:16 crc kubenswrapper[4716]: I1207 16:03:16.113900 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:16Z","lastTransitionTime":"2025-12-07T16:03:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:16 crc kubenswrapper[4716]: I1207 16:03:16.216445 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:16 crc kubenswrapper[4716]: I1207 16:03:16.216481 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:16 crc kubenswrapper[4716]: I1207 16:03:16.216490 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:16 crc kubenswrapper[4716]: I1207 16:03:16.216505 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:16 crc kubenswrapper[4716]: I1207 16:03:16.216516 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:16Z","lastTransitionTime":"2025-12-07T16:03:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:16 crc kubenswrapper[4716]: I1207 16:03:16.320745 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:16 crc kubenswrapper[4716]: I1207 16:03:16.320838 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:16 crc kubenswrapper[4716]: I1207 16:03:16.320867 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:16 crc kubenswrapper[4716]: I1207 16:03:16.320915 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:16 crc kubenswrapper[4716]: I1207 16:03:16.320942 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:16Z","lastTransitionTime":"2025-12-07T16:03:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:16 crc kubenswrapper[4716]: I1207 16:03:16.423143 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:16 crc kubenswrapper[4716]: I1207 16:03:16.423212 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:16 crc kubenswrapper[4716]: I1207 16:03:16.423224 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:16 crc kubenswrapper[4716]: I1207 16:03:16.423296 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:16 crc kubenswrapper[4716]: I1207 16:03:16.423310 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:16Z","lastTransitionTime":"2025-12-07T16:03:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:16 crc kubenswrapper[4716]: I1207 16:03:16.526880 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:16 crc kubenswrapper[4716]: I1207 16:03:16.526923 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:16 crc kubenswrapper[4716]: I1207 16:03:16.526935 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:16 crc kubenswrapper[4716]: I1207 16:03:16.526953 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:16 crc kubenswrapper[4716]: I1207 16:03:16.526968 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:16Z","lastTransitionTime":"2025-12-07T16:03:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:16 crc kubenswrapper[4716]: I1207 16:03:16.629596 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:16 crc kubenswrapper[4716]: I1207 16:03:16.629700 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:16 crc kubenswrapper[4716]: I1207 16:03:16.629719 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:16 crc kubenswrapper[4716]: I1207 16:03:16.629744 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:16 crc kubenswrapper[4716]: I1207 16:03:16.629760 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:16Z","lastTransitionTime":"2025-12-07T16:03:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:16 crc kubenswrapper[4716]: I1207 16:03:16.657125 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 16:03:16 crc kubenswrapper[4716]: E1207 16:03:16.657274 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 16:03:16 crc kubenswrapper[4716]: I1207 16:03:16.731930 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:16 crc kubenswrapper[4716]: I1207 16:03:16.731993 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:16 crc kubenswrapper[4716]: I1207 16:03:16.732010 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:16 crc kubenswrapper[4716]: I1207 16:03:16.732036 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:16 crc kubenswrapper[4716]: I1207 16:03:16.732059 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:16Z","lastTransitionTime":"2025-12-07T16:03:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:16 crc kubenswrapper[4716]: I1207 16:03:16.834283 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:16 crc kubenswrapper[4716]: I1207 16:03:16.834317 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:16 crc kubenswrapper[4716]: I1207 16:03:16.834326 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:16 crc kubenswrapper[4716]: I1207 16:03:16.834340 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:16 crc kubenswrapper[4716]: I1207 16:03:16.834348 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:16Z","lastTransitionTime":"2025-12-07T16:03:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:16 crc kubenswrapper[4716]: I1207 16:03:16.914542 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:16 crc kubenswrapper[4716]: I1207 16:03:16.914578 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:16 crc kubenswrapper[4716]: I1207 16:03:16.914587 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:16 crc kubenswrapper[4716]: I1207 16:03:16.914604 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:16 crc kubenswrapper[4716]: I1207 16:03:16.914614 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:16Z","lastTransitionTime":"2025-12-07T16:03:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:16 crc kubenswrapper[4716]: E1207 16:03:16.933218 4716 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a8d1b7d4-7ec4-40c7-86b0-291db1f38894\\\",\\\"systemUUID\\\":\\\"a57d09b0-0224-46eb-a3ac-a2b60dad2250\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:16Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:16 crc kubenswrapper[4716]: I1207 16:03:16.936876 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:16 crc kubenswrapper[4716]: I1207 16:03:16.936936 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:16 crc kubenswrapper[4716]: I1207 16:03:16.936958 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:16 crc kubenswrapper[4716]: I1207 16:03:16.936990 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:16 crc kubenswrapper[4716]: I1207 16:03:16.937013 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:16Z","lastTransitionTime":"2025-12-07T16:03:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:16 crc kubenswrapper[4716]: E1207 16:03:16.953908 4716 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a8d1b7d4-7ec4-40c7-86b0-291db1f38894\\\",\\\"systemUUID\\\":\\\"a57d09b0-0224-46eb-a3ac-a2b60dad2250\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:16Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:16 crc kubenswrapper[4716]: I1207 16:03:16.958648 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:16 crc kubenswrapper[4716]: I1207 16:03:16.958697 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:16 crc kubenswrapper[4716]: I1207 16:03:16.958706 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:16 crc kubenswrapper[4716]: I1207 16:03:16.958720 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:16 crc kubenswrapper[4716]: I1207 16:03:16.958731 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:16Z","lastTransitionTime":"2025-12-07T16:03:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:16 crc kubenswrapper[4716]: E1207 16:03:16.972307 4716 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a8d1b7d4-7ec4-40c7-86b0-291db1f38894\\\",\\\"systemUUID\\\":\\\"a57d09b0-0224-46eb-a3ac-a2b60dad2250\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:16Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:16 crc kubenswrapper[4716]: I1207 16:03:16.979368 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:16 crc kubenswrapper[4716]: I1207 16:03:16.979399 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:16 crc kubenswrapper[4716]: I1207 16:03:16.979421 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:16 crc kubenswrapper[4716]: I1207 16:03:16.979435 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:16 crc kubenswrapper[4716]: I1207 16:03:16.979444 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:16Z","lastTransitionTime":"2025-12-07T16:03:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:16 crc kubenswrapper[4716]: E1207 16:03:16.992729 4716 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a8d1b7d4-7ec4-40c7-86b0-291db1f38894\\\",\\\"systemUUID\\\":\\\"a57d09b0-0224-46eb-a3ac-a2b60dad2250\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:16Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:16 crc kubenswrapper[4716]: I1207 16:03:16.996849 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:16 crc kubenswrapper[4716]: I1207 16:03:16.996885 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:16 crc kubenswrapper[4716]: I1207 16:03:16.996897 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:16 crc kubenswrapper[4716]: I1207 16:03:16.996916 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:16 crc kubenswrapper[4716]: I1207 16:03:16.996928 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:16Z","lastTransitionTime":"2025-12-07T16:03:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:17 crc kubenswrapper[4716]: E1207 16:03:17.012983 4716 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a8d1b7d4-7ec4-40c7-86b0-291db1f38894\\\",\\\"systemUUID\\\":\\\"a57d09b0-0224-46eb-a3ac-a2b60dad2250\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:17Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:17 crc kubenswrapper[4716]: E1207 16:03:17.013318 4716 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 07 16:03:17 crc kubenswrapper[4716]: I1207 16:03:17.014756 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:17 crc kubenswrapper[4716]: I1207 16:03:17.014786 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:17 crc kubenswrapper[4716]: I1207 16:03:17.014794 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:17 crc kubenswrapper[4716]: I1207 16:03:17.014807 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:17 crc kubenswrapper[4716]: I1207 16:03:17.014817 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:17Z","lastTransitionTime":"2025-12-07T16:03:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:17 crc kubenswrapper[4716]: I1207 16:03:17.117760 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:17 crc kubenswrapper[4716]: I1207 16:03:17.117812 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:17 crc kubenswrapper[4716]: I1207 16:03:17.117824 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:17 crc kubenswrapper[4716]: I1207 16:03:17.117846 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:17 crc kubenswrapper[4716]: I1207 16:03:17.117860 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:17Z","lastTransitionTime":"2025-12-07T16:03:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:17 crc kubenswrapper[4716]: I1207 16:03:17.221141 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:17 crc kubenswrapper[4716]: I1207 16:03:17.221185 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:17 crc kubenswrapper[4716]: I1207 16:03:17.221196 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:17 crc kubenswrapper[4716]: I1207 16:03:17.221215 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:17 crc kubenswrapper[4716]: I1207 16:03:17.221228 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:17Z","lastTransitionTime":"2025-12-07T16:03:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:17 crc kubenswrapper[4716]: I1207 16:03:17.323906 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:17 crc kubenswrapper[4716]: I1207 16:03:17.323968 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:17 crc kubenswrapper[4716]: I1207 16:03:17.323978 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:17 crc kubenswrapper[4716]: I1207 16:03:17.323993 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:17 crc kubenswrapper[4716]: I1207 16:03:17.324003 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:17Z","lastTransitionTime":"2025-12-07T16:03:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:17 crc kubenswrapper[4716]: I1207 16:03:17.426552 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:17 crc kubenswrapper[4716]: I1207 16:03:17.426588 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:17 crc kubenswrapper[4716]: I1207 16:03:17.426598 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:17 crc kubenswrapper[4716]: I1207 16:03:17.426611 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:17 crc kubenswrapper[4716]: I1207 16:03:17.426621 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:17Z","lastTransitionTime":"2025-12-07T16:03:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:17 crc kubenswrapper[4716]: I1207 16:03:17.528373 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:17 crc kubenswrapper[4716]: I1207 16:03:17.528407 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:17 crc kubenswrapper[4716]: I1207 16:03:17.528417 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:17 crc kubenswrapper[4716]: I1207 16:03:17.528471 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:17 crc kubenswrapper[4716]: I1207 16:03:17.528481 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:17Z","lastTransitionTime":"2025-12-07T16:03:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:17 crc kubenswrapper[4716]: I1207 16:03:17.631210 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:17 crc kubenswrapper[4716]: I1207 16:03:17.631254 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:17 crc kubenswrapper[4716]: I1207 16:03:17.631266 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:17 crc kubenswrapper[4716]: I1207 16:03:17.631283 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:17 crc kubenswrapper[4716]: I1207 16:03:17.631294 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:17Z","lastTransitionTime":"2025-12-07T16:03:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:17 crc kubenswrapper[4716]: I1207 16:03:17.656780 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nx2lh" Dec 07 16:03:17 crc kubenswrapper[4716]: I1207 16:03:17.656980 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 16:03:17 crc kubenswrapper[4716]: E1207 16:03:17.657111 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nx2lh" podUID="04b37f6c-457b-4f9a-8733-e88cb9644483" Dec 07 16:03:17 crc kubenswrapper[4716]: I1207 16:03:17.657572 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:03:17 crc kubenswrapper[4716]: E1207 16:03:17.657660 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 16:03:17 crc kubenswrapper[4716]: E1207 16:03:17.658668 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 16:03:17 crc kubenswrapper[4716]: I1207 16:03:17.671557 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:17Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:17 crc kubenswrapper[4716]: I1207 16:03:17.681442 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c15b59eb-565d-4556-a4ce-75afdf159dc8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://530a2935826940767ea33fb4f37fa11f7dee4950d867fbcd6af72836d80f7db3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jks8z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ccaf6c1e6a978eb978a1c207b6c4c00b23737f78b743f966579aca781e22e0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jks8z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zcxxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:17Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:17 crc kubenswrapper[4716]: I1207 16:03:17.693041 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8ps9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98015cea-d753-4e7a-b2b5-5ae7733a81e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d68509b673eecd37725cdcc48e3d0612c3b9c7601433f2b72c21d662140daa59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f84f52a073437543721010dc440339654ffb5e04eeb6cc7afbcc62f00f64c8e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f84f52a073437543721010dc440339654ffb5e04eeb6cc7afbcc62f00f64c8e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f6fe1cb25218cb3ac20488dc47efa4dc6a26a43dbc8306b5ee38cbe1a186c67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f6fe1cb25218cb3ac20488dc47efa4dc6a26a43dbc8306b5ee38cbe1a186c67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14b748a8247f6863e8be61e90057a06626dfd3cd4f01bd3a11629f700f56378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f14b748a8247f6863e8be61e90057a06626dfd3cd4f01bd3a11629f700f56378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ec379c0d573cb9b5ffcfc5ca7c653ba60e2fbdd7e3d37836102e26618ce43e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ec379c0d573cb9b5ffcfc5ca7c653ba60e2fbdd7e3d37836102e26618ce43e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c75bf2a6703cf99865d4ed5fa000e1e3b1330d4946176cf585fec1a2f226df0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c75bf2a6703cf99865d4ed5fa000e1e3b1330d4946176cf585fec1a2f226df0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de45812c9a5a5bfe6b0146828d03fab28362b03d04ef861c150b9ffbf181d345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de45812c9a5a5bfe6b0146828d03fab28362b03d04ef861c150b9ffbf181d345\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8ps9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:17Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:17 crc kubenswrapper[4716]: I1207 16:03:17.709288 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7961cc9a-734e-42c2-b1ff-7075f3f20091\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84a36b76f20618c7885d611ecad9db29bac5101c27572be31d3acbb90c581291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0a0eca3c2087a457c0e448e649345ec4595a9c8a019aed2cddc5f30c77871e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://404426a171b8676eddd4a00996c76afab7e269643fa9608f1e1b6bdba385071d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fd3c2941782921f99a6cdc3de4e892234548bde527fec15a8ac92e513566e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e23abfdca314ed43d594333c9f9e401cc8ef7dad39b8ebac84a836b06ee5c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09b38cd4b389ea477c802a842ee4f2297a64d64684624117163e819539c257c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09b38cd4b389ea477c802a842ee4f2297a64d64684624117163e819539c257c6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c00f7a8e024e0e0dfd7b73e16d1d5437602e79ea95b73c21f24b209edef60e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c00f7a8e024e0e0dfd7b73e16d1d5437602e79ea95b73c21f24b209edef60e98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3f6392d426fa2cd788ec3a5c3bff4ac39483a54dfb2624a1c4370522413a350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3f6392d426fa2cd788ec3a5c3bff4ac39483a54dfb2624a1c4370522413a350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:17Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:17 crc kubenswrapper[4716]: I1207 16:03:17.720735 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3067de7c-c11c-4be1-bd94-d6572c266795\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d14d018fa044e9e80fe81e7e38df935992d14c773ed2b2193fd5cb873e5d0947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a9e9cf048220ca14c996847e9ced9f2367b9331dc6c2836dd023658855cd79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6d3c911874095b3c7cb062d03070e5090a7bb2456cb20cb4ce078f502515cda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ac6ce7701f1c6aa1ea8fcf7a3035e3b0eaa28ae3cb6afc442372bd9117bdde4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:17Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:17 crc kubenswrapper[4716]: I1207 16:03:17.733400 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:17 crc kubenswrapper[4716]: I1207 16:03:17.733441 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:17 crc kubenswrapper[4716]: I1207 16:03:17.733450 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:17 crc kubenswrapper[4716]: I1207 16:03:17.733462 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:17 crc kubenswrapper[4716]: I1207 16:03:17.733470 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:17Z","lastTransitionTime":"2025-12-07T16:03:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:17 crc kubenswrapper[4716]: I1207 16:03:17.734619 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4742ea84b3a2c9461d9b900beee7354a2ded24c39ad23e007cbf270ae12377c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:17Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:17 crc kubenswrapper[4716]: I1207 16:03:17.745884 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nx2lh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04b37f6c-457b-4f9a-8733-e88cb9644483\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv6td\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv6td\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nx2lh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:17Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:17 crc kubenswrapper[4716]: I1207 16:03:17.755504 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e903f3c-ba8c-477d-b66a-6d1331a7c64e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ff9c456318bbdb54ec3f1ac3e2c2366968f18ad084158b7100d142acb668948\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1358745250b35426602fc7417e4f85dd411cdd0fb3e8f703c432d72128ca7c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d8c9a434c6cfb9e4fee6a8955f1379291f153c007876d1ffb498ce561c73ba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad8df9b5bd5f5a8dd44ce2899cfa29c19435149b170d339d339183cd3df09e99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad8df9b5bd5f5a8dd44ce2899cfa29c19435149b170d339d339183cd3df09e99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:17Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:17 crc kubenswrapper[4716]: I1207 16:03:17.770282 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c54d618a-dfec-4f2a-a288-f5188c8a128a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60324dc7d701bc064e14ae4f474bce10f6ee4ff3c3c942eab749e7166082cf9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ed4f4d3564674bf5cb4e7a76dc0fd95e203ecc8950798e59a1fb6c9ab4a20f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://def9ac840fb8ecf2778bed88ea980eddca7c5b2aba4bd1afb967d62fec3f7bae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b73be664f5412c98b89d971a2cac37c494151c293fdb44a98b390978658a6643\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a847d04e22e98091c0037a89054b9236228b701b8c7ba31940a56df0f37b95c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcae78a6133d97f6a32dd7421ccec4333b970a237404e056a5cc601f8eefdc8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f48167b38ff840f85f11b69d8090ec6c9c634ed26682daef93f5da99fc02ef61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f48167b38ff840f85f11b69d8090ec6c9c634ed26682daef93f5da99fc02ef61\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-07T16:03:08Z\\\",\\\"message\\\":\\\"er, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:08Z is after 2025-08-24T17:21:41Z]\\\\nI1207 16:03:08.453903 6420 services_controller.go:451] Built service openshift-controller-manager-operator/metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-controller-manager-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-controller-manager-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.58\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.Tem\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T16:03:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qvc7p_openshift-ovn-kubernetes(c54d618a-dfec-4f2a-a288-f5188c8a128a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcfdffe822635b0ae13b253e6133449fe00461b2792be43d913c4a77219b8255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9bfe2ffce7a7b7c320474117734fc05246cc2b4fd744833e896e8e7c83f5e4d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9bfe2ffce7a7b7c320474117734fc05246cc2b4fd744833e896e8e7c83f5e4d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qvc7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:17Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:17 crc kubenswrapper[4716]: I1207 16:03:17.780109 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mx85f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"118dd9ba-437c-4e5c-b5ff-4239ae6d4a63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bcd6c8deb184d0c713c5e7b7cc82fa38d3cd498517911ba7a0ec3cb88a3437a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gg5nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ffc069d2b21fa80871b1b97f2c72c37880d9a309cbf8b7daf7d240f2c332a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gg5nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mx85f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:17Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:17 crc kubenswrapper[4716]: I1207 16:03:17.792199 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4wr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0baf8575-26cf-4a72-8a6d-ae94c078eaad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0ca2b3cd5a11033afd219b0125be7ab5acd750935d7eb89a470eff3d93a0838\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p2trd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4wr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:17Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:17 crc kubenswrapper[4716]: I1207 16:03:17.801073 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-m5jl5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"824ee59f-1474-4994-a70e-462878521514\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a15fe50cc75515704779646cc19aa50004f0db9aa26fe2535110df749bbfb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhwzn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-m5jl5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:17Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:17 crc kubenswrapper[4716]: I1207 16:03:17.812292 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdee9957-48b9-458f-ba71-e0cd39f80596\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73f85248b4d68a9beecdacb3a01f433e97a0ffea2247ce92a41ef96855509377\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6276df652efe6ca935e745f24ef20f87a16465475139230c63b9999234a4d12a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3b1dd202586ee755ca89fbe12b30bbb12f95a62514f6e4d93cbbdbe55181e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd318647816e4a27849666b0ec4fc2ca4905295661c557dc381e1144f22cf284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d31de0326a05216fa10442bdb47151c6e073c97e9a16c34e793028e5e253756\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"le observer\\\\nW1207 16:02:34.645464 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1207 16:02:34.645587 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 16:02:34.646501 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2082283479/tls.crt::/tmp/serving-cert-2082283479/tls.key\\\\\\\"\\\\nI1207 16:02:34.972041 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 16:02:34.973702 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 16:02:34.973717 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 16:02:34.973738 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 16:02:34.973743 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 16:02:34.977185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1207 16:02:34.977204 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1207 16:02:34.977209 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977217 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 16:02:34.977225 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 16:02:34.977228 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 16:02:34.977232 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1207 16:02:34.979033 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27bd4a54f920b2725987c1c7fd2d9fdf16cd08738c246c333a4e9164a3a582a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:17Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:17 crc kubenswrapper[4716]: I1207 16:03:17.837567 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:17Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:17 crc kubenswrapper[4716]: I1207 16:03:17.849383 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:17 crc kubenswrapper[4716]: I1207 16:03:17.849435 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:17 crc kubenswrapper[4716]: I1207 16:03:17.849453 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:17 crc kubenswrapper[4716]: I1207 16:03:17.849476 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:17 crc kubenswrapper[4716]: I1207 16:03:17.849494 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:17Z","lastTransitionTime":"2025-12-07T16:03:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:17 crc kubenswrapper[4716]: I1207 16:03:17.852140 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2jjld" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dee6f5c-861c-4819-9309-7791574c7b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51c1d4159ab11087839a2c2d192566fb9322447c26df560bf09517c906ececb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmdns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2jjld\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:17Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:17 crc kubenswrapper[4716]: I1207 16:03:17.872163 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f95ef61779507288a03bda8eddf745e105f705998a117a30789675bfa452d88f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b955b23163e7c8c458a3bbe2f681633e6a663ce76748febcc0b3a3c37f3ca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:17Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:17 crc kubenswrapper[4716]: I1207 16:03:17.884789 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef3335ae3875a75ab64685fc7e3590177280d651e8f8292c6f5ea6f472e55789\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:17Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:17 crc kubenswrapper[4716]: I1207 16:03:17.897839 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:17Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:17 crc kubenswrapper[4716]: I1207 16:03:17.952387 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:17 crc kubenswrapper[4716]: I1207 16:03:17.952464 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:17 crc kubenswrapper[4716]: I1207 16:03:17.952520 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:17 crc kubenswrapper[4716]: I1207 16:03:17.952556 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:17 crc kubenswrapper[4716]: I1207 16:03:17.952581 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:17Z","lastTransitionTime":"2025-12-07T16:03:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:18 crc kubenswrapper[4716]: I1207 16:03:18.054954 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:18 crc kubenswrapper[4716]: I1207 16:03:18.054984 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:18 crc kubenswrapper[4716]: I1207 16:03:18.054992 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:18 crc kubenswrapper[4716]: I1207 16:03:18.055004 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:18 crc kubenswrapper[4716]: I1207 16:03:18.055013 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:18Z","lastTransitionTime":"2025-12-07T16:03:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:18 crc kubenswrapper[4716]: I1207 16:03:18.157456 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:18 crc kubenswrapper[4716]: I1207 16:03:18.157506 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:18 crc kubenswrapper[4716]: I1207 16:03:18.157517 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:18 crc kubenswrapper[4716]: I1207 16:03:18.157532 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:18 crc kubenswrapper[4716]: I1207 16:03:18.157541 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:18Z","lastTransitionTime":"2025-12-07T16:03:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:18 crc kubenswrapper[4716]: I1207 16:03:18.263142 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:18 crc kubenswrapper[4716]: I1207 16:03:18.263728 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:18 crc kubenswrapper[4716]: I1207 16:03:18.263827 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:18 crc kubenswrapper[4716]: I1207 16:03:18.263910 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:18 crc kubenswrapper[4716]: I1207 16:03:18.263990 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:18Z","lastTransitionTime":"2025-12-07T16:03:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:18 crc kubenswrapper[4716]: I1207 16:03:18.366033 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:18 crc kubenswrapper[4716]: I1207 16:03:18.366287 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:18 crc kubenswrapper[4716]: I1207 16:03:18.366447 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:18 crc kubenswrapper[4716]: I1207 16:03:18.366583 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:18 crc kubenswrapper[4716]: I1207 16:03:18.366722 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:18Z","lastTransitionTime":"2025-12-07T16:03:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:18 crc kubenswrapper[4716]: I1207 16:03:18.469334 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:18 crc kubenswrapper[4716]: I1207 16:03:18.469719 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:18 crc kubenswrapper[4716]: I1207 16:03:18.469838 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:18 crc kubenswrapper[4716]: I1207 16:03:18.469975 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:18 crc kubenswrapper[4716]: I1207 16:03:18.470109 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:18Z","lastTransitionTime":"2025-12-07T16:03:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:18 crc kubenswrapper[4716]: I1207 16:03:18.572882 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:18 crc kubenswrapper[4716]: I1207 16:03:18.573169 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:18 crc kubenswrapper[4716]: I1207 16:03:18.573308 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:18 crc kubenswrapper[4716]: I1207 16:03:18.573413 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:18 crc kubenswrapper[4716]: I1207 16:03:18.573501 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:18Z","lastTransitionTime":"2025-12-07T16:03:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:18 crc kubenswrapper[4716]: I1207 16:03:18.657141 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 16:03:18 crc kubenswrapper[4716]: E1207 16:03:18.657857 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 16:03:18 crc kubenswrapper[4716]: I1207 16:03:18.675613 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:18 crc kubenswrapper[4716]: I1207 16:03:18.675652 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:18 crc kubenswrapper[4716]: I1207 16:03:18.675663 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:18 crc kubenswrapper[4716]: I1207 16:03:18.675678 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:18 crc kubenswrapper[4716]: I1207 16:03:18.675688 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:18Z","lastTransitionTime":"2025-12-07T16:03:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:18 crc kubenswrapper[4716]: I1207 16:03:18.778405 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:18 crc kubenswrapper[4716]: I1207 16:03:18.778460 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:18 crc kubenswrapper[4716]: I1207 16:03:18.778475 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:18 crc kubenswrapper[4716]: I1207 16:03:18.778496 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:18 crc kubenswrapper[4716]: I1207 16:03:18.778511 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:18Z","lastTransitionTime":"2025-12-07T16:03:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:18 crc kubenswrapper[4716]: I1207 16:03:18.880453 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:18 crc kubenswrapper[4716]: I1207 16:03:18.880480 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:18 crc kubenswrapper[4716]: I1207 16:03:18.880490 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:18 crc kubenswrapper[4716]: I1207 16:03:18.880506 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:18 crc kubenswrapper[4716]: I1207 16:03:18.880516 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:18Z","lastTransitionTime":"2025-12-07T16:03:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:18 crc kubenswrapper[4716]: I1207 16:03:18.982579 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:18 crc kubenswrapper[4716]: I1207 16:03:18.982621 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:18 crc kubenswrapper[4716]: I1207 16:03:18.982633 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:18 crc kubenswrapper[4716]: I1207 16:03:18.982648 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:18 crc kubenswrapper[4716]: I1207 16:03:18.982658 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:18Z","lastTransitionTime":"2025-12-07T16:03:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:19 crc kubenswrapper[4716]: I1207 16:03:19.085814 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:19 crc kubenswrapper[4716]: I1207 16:03:19.085914 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:19 crc kubenswrapper[4716]: I1207 16:03:19.085932 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:19 crc kubenswrapper[4716]: I1207 16:03:19.085960 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:19 crc kubenswrapper[4716]: I1207 16:03:19.085977 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:19Z","lastTransitionTime":"2025-12-07T16:03:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:19 crc kubenswrapper[4716]: I1207 16:03:19.188402 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:19 crc kubenswrapper[4716]: I1207 16:03:19.188485 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:19 crc kubenswrapper[4716]: I1207 16:03:19.188496 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:19 crc kubenswrapper[4716]: I1207 16:03:19.188509 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:19 crc kubenswrapper[4716]: I1207 16:03:19.188518 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:19Z","lastTransitionTime":"2025-12-07T16:03:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:19 crc kubenswrapper[4716]: I1207 16:03:19.290806 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:19 crc kubenswrapper[4716]: I1207 16:03:19.290862 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:19 crc kubenswrapper[4716]: I1207 16:03:19.290879 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:19 crc kubenswrapper[4716]: I1207 16:03:19.290900 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:19 crc kubenswrapper[4716]: I1207 16:03:19.290917 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:19Z","lastTransitionTime":"2025-12-07T16:03:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:19 crc kubenswrapper[4716]: I1207 16:03:19.393289 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:19 crc kubenswrapper[4716]: I1207 16:03:19.393363 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:19 crc kubenswrapper[4716]: I1207 16:03:19.393389 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:19 crc kubenswrapper[4716]: I1207 16:03:19.393418 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:19 crc kubenswrapper[4716]: I1207 16:03:19.393439 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:19Z","lastTransitionTime":"2025-12-07T16:03:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:19 crc kubenswrapper[4716]: I1207 16:03:19.496240 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:19 crc kubenswrapper[4716]: I1207 16:03:19.496305 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:19 crc kubenswrapper[4716]: I1207 16:03:19.496323 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:19 crc kubenswrapper[4716]: I1207 16:03:19.496348 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:19 crc kubenswrapper[4716]: I1207 16:03:19.496368 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:19Z","lastTransitionTime":"2025-12-07T16:03:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:19 crc kubenswrapper[4716]: I1207 16:03:19.600963 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:19 crc kubenswrapper[4716]: I1207 16:03:19.601036 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:19 crc kubenswrapper[4716]: I1207 16:03:19.601056 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:19 crc kubenswrapper[4716]: I1207 16:03:19.601106 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:19 crc kubenswrapper[4716]: I1207 16:03:19.601124 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:19Z","lastTransitionTime":"2025-12-07T16:03:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:19 crc kubenswrapper[4716]: I1207 16:03:19.657090 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nx2lh" Dec 07 16:03:19 crc kubenswrapper[4716]: I1207 16:03:19.657140 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:03:19 crc kubenswrapper[4716]: E1207 16:03:19.657219 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nx2lh" podUID="04b37f6c-457b-4f9a-8733-e88cb9644483" Dec 07 16:03:19 crc kubenswrapper[4716]: I1207 16:03:19.657140 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 16:03:19 crc kubenswrapper[4716]: E1207 16:03:19.657287 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 16:03:19 crc kubenswrapper[4716]: E1207 16:03:19.657344 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 16:03:19 crc kubenswrapper[4716]: I1207 16:03:19.704236 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:19 crc kubenswrapper[4716]: I1207 16:03:19.704333 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:19 crc kubenswrapper[4716]: I1207 16:03:19.704350 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:19 crc kubenswrapper[4716]: I1207 16:03:19.704374 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:19 crc kubenswrapper[4716]: I1207 16:03:19.704422 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:19Z","lastTransitionTime":"2025-12-07T16:03:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:19 crc kubenswrapper[4716]: I1207 16:03:19.807521 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:19 crc kubenswrapper[4716]: I1207 16:03:19.807637 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:19 crc kubenswrapper[4716]: I1207 16:03:19.807709 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:19 crc kubenswrapper[4716]: I1207 16:03:19.807741 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:19 crc kubenswrapper[4716]: I1207 16:03:19.807819 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:19Z","lastTransitionTime":"2025-12-07T16:03:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:19 crc kubenswrapper[4716]: I1207 16:03:19.910707 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:19 crc kubenswrapper[4716]: I1207 16:03:19.910784 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:19 crc kubenswrapper[4716]: I1207 16:03:19.910809 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:19 crc kubenswrapper[4716]: I1207 16:03:19.910839 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:19 crc kubenswrapper[4716]: I1207 16:03:19.910861 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:19Z","lastTransitionTime":"2025-12-07T16:03:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:20 crc kubenswrapper[4716]: I1207 16:03:20.015480 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:20 crc kubenswrapper[4716]: I1207 16:03:20.015554 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:20 crc kubenswrapper[4716]: I1207 16:03:20.015582 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:20 crc kubenswrapper[4716]: I1207 16:03:20.015615 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:20 crc kubenswrapper[4716]: I1207 16:03:20.015639 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:20Z","lastTransitionTime":"2025-12-07T16:03:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:20 crc kubenswrapper[4716]: I1207 16:03:20.119046 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:20 crc kubenswrapper[4716]: I1207 16:03:20.119151 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:20 crc kubenswrapper[4716]: I1207 16:03:20.119176 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:20 crc kubenswrapper[4716]: I1207 16:03:20.119205 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:20 crc kubenswrapper[4716]: I1207 16:03:20.119228 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:20Z","lastTransitionTime":"2025-12-07T16:03:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:20 crc kubenswrapper[4716]: I1207 16:03:20.222158 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:20 crc kubenswrapper[4716]: I1207 16:03:20.222238 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:20 crc kubenswrapper[4716]: I1207 16:03:20.222259 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:20 crc kubenswrapper[4716]: I1207 16:03:20.222288 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:20 crc kubenswrapper[4716]: I1207 16:03:20.222305 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:20Z","lastTransitionTime":"2025-12-07T16:03:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:20 crc kubenswrapper[4716]: I1207 16:03:20.325295 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:20 crc kubenswrapper[4716]: I1207 16:03:20.325326 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:20 crc kubenswrapper[4716]: I1207 16:03:20.325334 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:20 crc kubenswrapper[4716]: I1207 16:03:20.325349 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:20 crc kubenswrapper[4716]: I1207 16:03:20.325358 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:20Z","lastTransitionTime":"2025-12-07T16:03:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:20 crc kubenswrapper[4716]: I1207 16:03:20.428392 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:20 crc kubenswrapper[4716]: I1207 16:03:20.428597 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:20 crc kubenswrapper[4716]: I1207 16:03:20.428619 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:20 crc kubenswrapper[4716]: I1207 16:03:20.428652 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:20 crc kubenswrapper[4716]: I1207 16:03:20.428673 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:20Z","lastTransitionTime":"2025-12-07T16:03:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:20 crc kubenswrapper[4716]: I1207 16:03:20.531513 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:20 crc kubenswrapper[4716]: I1207 16:03:20.531553 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:20 crc kubenswrapper[4716]: I1207 16:03:20.531570 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:20 crc kubenswrapper[4716]: I1207 16:03:20.531585 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:20 crc kubenswrapper[4716]: I1207 16:03:20.531595 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:20Z","lastTransitionTime":"2025-12-07T16:03:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:20 crc kubenswrapper[4716]: I1207 16:03:20.634533 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:20 crc kubenswrapper[4716]: I1207 16:03:20.634568 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:20 crc kubenswrapper[4716]: I1207 16:03:20.634576 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:20 crc kubenswrapper[4716]: I1207 16:03:20.634590 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:20 crc kubenswrapper[4716]: I1207 16:03:20.634599 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:20Z","lastTransitionTime":"2025-12-07T16:03:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:20 crc kubenswrapper[4716]: I1207 16:03:20.657186 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 16:03:20 crc kubenswrapper[4716]: E1207 16:03:20.657294 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 16:03:20 crc kubenswrapper[4716]: I1207 16:03:20.737017 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:20 crc kubenswrapper[4716]: I1207 16:03:20.737069 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:20 crc kubenswrapper[4716]: I1207 16:03:20.737095 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:20 crc kubenswrapper[4716]: I1207 16:03:20.737111 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:20 crc kubenswrapper[4716]: I1207 16:03:20.737124 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:20Z","lastTransitionTime":"2025-12-07T16:03:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:20 crc kubenswrapper[4716]: I1207 16:03:20.842745 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:20 crc kubenswrapper[4716]: I1207 16:03:20.842839 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:20 crc kubenswrapper[4716]: I1207 16:03:20.842849 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:20 crc kubenswrapper[4716]: I1207 16:03:20.842863 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:20 crc kubenswrapper[4716]: I1207 16:03:20.842876 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:20Z","lastTransitionTime":"2025-12-07T16:03:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:20 crc kubenswrapper[4716]: I1207 16:03:20.945431 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:20 crc kubenswrapper[4716]: I1207 16:03:20.945463 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:20 crc kubenswrapper[4716]: I1207 16:03:20.945471 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:20 crc kubenswrapper[4716]: I1207 16:03:20.945484 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:20 crc kubenswrapper[4716]: I1207 16:03:20.945492 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:20Z","lastTransitionTime":"2025-12-07T16:03:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:21 crc kubenswrapper[4716]: I1207 16:03:21.047512 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:21 crc kubenswrapper[4716]: I1207 16:03:21.047547 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:21 crc kubenswrapper[4716]: I1207 16:03:21.047556 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:21 crc kubenswrapper[4716]: I1207 16:03:21.047571 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:21 crc kubenswrapper[4716]: I1207 16:03:21.047581 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:21Z","lastTransitionTime":"2025-12-07T16:03:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:21 crc kubenswrapper[4716]: I1207 16:03:21.150629 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:21 crc kubenswrapper[4716]: I1207 16:03:21.150666 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:21 crc kubenswrapper[4716]: I1207 16:03:21.150677 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:21 crc kubenswrapper[4716]: I1207 16:03:21.150691 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:21 crc kubenswrapper[4716]: I1207 16:03:21.150702 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:21Z","lastTransitionTime":"2025-12-07T16:03:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:21 crc kubenswrapper[4716]: I1207 16:03:21.252709 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:21 crc kubenswrapper[4716]: I1207 16:03:21.252745 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:21 crc kubenswrapper[4716]: I1207 16:03:21.252755 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:21 crc kubenswrapper[4716]: I1207 16:03:21.252771 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:21 crc kubenswrapper[4716]: I1207 16:03:21.252781 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:21Z","lastTransitionTime":"2025-12-07T16:03:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:21 crc kubenswrapper[4716]: I1207 16:03:21.355590 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:21 crc kubenswrapper[4716]: I1207 16:03:21.355635 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:21 crc kubenswrapper[4716]: I1207 16:03:21.355650 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:21 crc kubenswrapper[4716]: I1207 16:03:21.355667 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:21 crc kubenswrapper[4716]: I1207 16:03:21.355678 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:21Z","lastTransitionTime":"2025-12-07T16:03:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:21 crc kubenswrapper[4716]: I1207 16:03:21.458211 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:21 crc kubenswrapper[4716]: I1207 16:03:21.458256 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:21 crc kubenswrapper[4716]: I1207 16:03:21.458265 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:21 crc kubenswrapper[4716]: I1207 16:03:21.458304 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:21 crc kubenswrapper[4716]: I1207 16:03:21.458313 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:21Z","lastTransitionTime":"2025-12-07T16:03:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:21 crc kubenswrapper[4716]: I1207 16:03:21.561264 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:21 crc kubenswrapper[4716]: I1207 16:03:21.561325 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:21 crc kubenswrapper[4716]: I1207 16:03:21.561342 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:21 crc kubenswrapper[4716]: I1207 16:03:21.561366 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:21 crc kubenswrapper[4716]: I1207 16:03:21.561384 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:21Z","lastTransitionTime":"2025-12-07T16:03:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:21 crc kubenswrapper[4716]: I1207 16:03:21.657605 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 16:03:21 crc kubenswrapper[4716]: I1207 16:03:21.657645 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:03:21 crc kubenswrapper[4716]: I1207 16:03:21.657616 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nx2lh" Dec 07 16:03:21 crc kubenswrapper[4716]: E1207 16:03:21.657734 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 16:03:21 crc kubenswrapper[4716]: E1207 16:03:21.657815 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nx2lh" podUID="04b37f6c-457b-4f9a-8733-e88cb9644483" Dec 07 16:03:21 crc kubenswrapper[4716]: E1207 16:03:21.657875 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 16:03:21 crc kubenswrapper[4716]: I1207 16:03:21.664068 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:21 crc kubenswrapper[4716]: I1207 16:03:21.664130 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:21 crc kubenswrapper[4716]: I1207 16:03:21.664142 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:21 crc kubenswrapper[4716]: I1207 16:03:21.664160 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:21 crc kubenswrapper[4716]: I1207 16:03:21.664172 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:21Z","lastTransitionTime":"2025-12-07T16:03:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:21 crc kubenswrapper[4716]: I1207 16:03:21.766765 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:21 crc kubenswrapper[4716]: I1207 16:03:21.766816 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:21 crc kubenswrapper[4716]: I1207 16:03:21.766826 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:21 crc kubenswrapper[4716]: I1207 16:03:21.766847 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:21 crc kubenswrapper[4716]: I1207 16:03:21.766858 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:21Z","lastTransitionTime":"2025-12-07T16:03:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:21 crc kubenswrapper[4716]: I1207 16:03:21.869341 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:21 crc kubenswrapper[4716]: I1207 16:03:21.869385 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:21 crc kubenswrapper[4716]: I1207 16:03:21.869400 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:21 crc kubenswrapper[4716]: I1207 16:03:21.869418 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:21 crc kubenswrapper[4716]: I1207 16:03:21.869429 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:21Z","lastTransitionTime":"2025-12-07T16:03:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:21 crc kubenswrapper[4716]: I1207 16:03:21.972126 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:21 crc kubenswrapper[4716]: I1207 16:03:21.972159 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:21 crc kubenswrapper[4716]: I1207 16:03:21.972170 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:21 crc kubenswrapper[4716]: I1207 16:03:21.972184 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:21 crc kubenswrapper[4716]: I1207 16:03:21.972194 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:21Z","lastTransitionTime":"2025-12-07T16:03:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:22 crc kubenswrapper[4716]: I1207 16:03:22.074515 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:22 crc kubenswrapper[4716]: I1207 16:03:22.074551 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:22 crc kubenswrapper[4716]: I1207 16:03:22.074560 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:22 crc kubenswrapper[4716]: I1207 16:03:22.074574 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:22 crc kubenswrapper[4716]: I1207 16:03:22.074584 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:22Z","lastTransitionTime":"2025-12-07T16:03:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:22 crc kubenswrapper[4716]: I1207 16:03:22.177401 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:22 crc kubenswrapper[4716]: I1207 16:03:22.177440 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:22 crc kubenswrapper[4716]: I1207 16:03:22.177449 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:22 crc kubenswrapper[4716]: I1207 16:03:22.177465 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:22 crc kubenswrapper[4716]: I1207 16:03:22.177475 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:22Z","lastTransitionTime":"2025-12-07T16:03:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:22 crc kubenswrapper[4716]: I1207 16:03:22.279503 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:22 crc kubenswrapper[4716]: I1207 16:03:22.279544 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:22 crc kubenswrapper[4716]: I1207 16:03:22.279552 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:22 crc kubenswrapper[4716]: I1207 16:03:22.279567 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:22 crc kubenswrapper[4716]: I1207 16:03:22.279577 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:22Z","lastTransitionTime":"2025-12-07T16:03:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:22 crc kubenswrapper[4716]: I1207 16:03:22.381525 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:22 crc kubenswrapper[4716]: I1207 16:03:22.381551 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:22 crc kubenswrapper[4716]: I1207 16:03:22.381558 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:22 crc kubenswrapper[4716]: I1207 16:03:22.381572 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:22 crc kubenswrapper[4716]: I1207 16:03:22.381581 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:22Z","lastTransitionTime":"2025-12-07T16:03:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:22 crc kubenswrapper[4716]: I1207 16:03:22.483654 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:22 crc kubenswrapper[4716]: I1207 16:03:22.483693 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:22 crc kubenswrapper[4716]: I1207 16:03:22.483704 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:22 crc kubenswrapper[4716]: I1207 16:03:22.483722 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:22 crc kubenswrapper[4716]: I1207 16:03:22.483733 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:22Z","lastTransitionTime":"2025-12-07T16:03:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:22 crc kubenswrapper[4716]: I1207 16:03:22.586761 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:22 crc kubenswrapper[4716]: I1207 16:03:22.586798 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:22 crc kubenswrapper[4716]: I1207 16:03:22.586806 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:22 crc kubenswrapper[4716]: I1207 16:03:22.586823 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:22 crc kubenswrapper[4716]: I1207 16:03:22.586832 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:22Z","lastTransitionTime":"2025-12-07T16:03:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:22 crc kubenswrapper[4716]: I1207 16:03:22.657300 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 16:03:22 crc kubenswrapper[4716]: E1207 16:03:22.657479 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 16:03:22 crc kubenswrapper[4716]: I1207 16:03:22.689429 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:22 crc kubenswrapper[4716]: I1207 16:03:22.689491 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:22 crc kubenswrapper[4716]: I1207 16:03:22.689510 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:22 crc kubenswrapper[4716]: I1207 16:03:22.689534 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:22 crc kubenswrapper[4716]: I1207 16:03:22.689550 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:22Z","lastTransitionTime":"2025-12-07T16:03:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:22 crc kubenswrapper[4716]: I1207 16:03:22.791517 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:22 crc kubenswrapper[4716]: I1207 16:03:22.791551 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:22 crc kubenswrapper[4716]: I1207 16:03:22.791559 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:22 crc kubenswrapper[4716]: I1207 16:03:22.791573 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:22 crc kubenswrapper[4716]: I1207 16:03:22.791583 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:22Z","lastTransitionTime":"2025-12-07T16:03:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:22 crc kubenswrapper[4716]: I1207 16:03:22.894422 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:22 crc kubenswrapper[4716]: I1207 16:03:22.894451 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:22 crc kubenswrapper[4716]: I1207 16:03:22.894459 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:22 crc kubenswrapper[4716]: I1207 16:03:22.894470 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:22 crc kubenswrapper[4716]: I1207 16:03:22.894479 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:22Z","lastTransitionTime":"2025-12-07T16:03:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:22 crc kubenswrapper[4716]: I1207 16:03:22.997130 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:22 crc kubenswrapper[4716]: I1207 16:03:22.997180 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:22 crc kubenswrapper[4716]: I1207 16:03:22.997199 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:22 crc kubenswrapper[4716]: I1207 16:03:22.997221 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:22 crc kubenswrapper[4716]: I1207 16:03:22.997237 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:22Z","lastTransitionTime":"2025-12-07T16:03:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:23 crc kubenswrapper[4716]: I1207 16:03:23.099869 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:23 crc kubenswrapper[4716]: I1207 16:03:23.099918 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:23 crc kubenswrapper[4716]: I1207 16:03:23.099933 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:23 crc kubenswrapper[4716]: I1207 16:03:23.099953 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:23 crc kubenswrapper[4716]: I1207 16:03:23.099969 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:23Z","lastTransitionTime":"2025-12-07T16:03:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:23 crc kubenswrapper[4716]: I1207 16:03:23.202483 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:23 crc kubenswrapper[4716]: I1207 16:03:23.202522 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:23 crc kubenswrapper[4716]: I1207 16:03:23.202534 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:23 crc kubenswrapper[4716]: I1207 16:03:23.202551 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:23 crc kubenswrapper[4716]: I1207 16:03:23.202561 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:23Z","lastTransitionTime":"2025-12-07T16:03:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:23 crc kubenswrapper[4716]: I1207 16:03:23.305040 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:23 crc kubenswrapper[4716]: I1207 16:03:23.305318 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:23 crc kubenswrapper[4716]: I1207 16:03:23.305332 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:23 crc kubenswrapper[4716]: I1207 16:03:23.305362 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:23 crc kubenswrapper[4716]: I1207 16:03:23.305377 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:23Z","lastTransitionTime":"2025-12-07T16:03:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:23 crc kubenswrapper[4716]: I1207 16:03:23.408777 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:23 crc kubenswrapper[4716]: I1207 16:03:23.408846 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:23 crc kubenswrapper[4716]: I1207 16:03:23.408859 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:23 crc kubenswrapper[4716]: I1207 16:03:23.408879 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:23 crc kubenswrapper[4716]: I1207 16:03:23.408892 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:23Z","lastTransitionTime":"2025-12-07T16:03:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:23 crc kubenswrapper[4716]: I1207 16:03:23.511884 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:23 crc kubenswrapper[4716]: I1207 16:03:23.511931 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:23 crc kubenswrapper[4716]: I1207 16:03:23.511943 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:23 crc kubenswrapper[4716]: I1207 16:03:23.511964 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:23 crc kubenswrapper[4716]: I1207 16:03:23.511977 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:23Z","lastTransitionTime":"2025-12-07T16:03:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:23 crc kubenswrapper[4716]: I1207 16:03:23.614633 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:23 crc kubenswrapper[4716]: I1207 16:03:23.614919 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:23 crc kubenswrapper[4716]: I1207 16:03:23.614998 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:23 crc kubenswrapper[4716]: I1207 16:03:23.615107 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:23 crc kubenswrapper[4716]: I1207 16:03:23.615184 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:23Z","lastTransitionTime":"2025-12-07T16:03:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:23 crc kubenswrapper[4716]: I1207 16:03:23.657134 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 16:03:23 crc kubenswrapper[4716]: E1207 16:03:23.657283 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 16:03:23 crc kubenswrapper[4716]: I1207 16:03:23.657363 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:03:23 crc kubenswrapper[4716]: I1207 16:03:23.657542 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nx2lh" Dec 07 16:03:23 crc kubenswrapper[4716]: E1207 16:03:23.657570 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 16:03:23 crc kubenswrapper[4716]: E1207 16:03:23.657820 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nx2lh" podUID="04b37f6c-457b-4f9a-8733-e88cb9644483" Dec 07 16:03:23 crc kubenswrapper[4716]: I1207 16:03:23.717797 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:23 crc kubenswrapper[4716]: I1207 16:03:23.717918 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:23 crc kubenswrapper[4716]: I1207 16:03:23.717936 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:23 crc kubenswrapper[4716]: I1207 16:03:23.717966 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:23 crc kubenswrapper[4716]: I1207 16:03:23.717983 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:23Z","lastTransitionTime":"2025-12-07T16:03:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:23 crc kubenswrapper[4716]: I1207 16:03:23.820945 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:23 crc kubenswrapper[4716]: I1207 16:03:23.820977 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:23 crc kubenswrapper[4716]: I1207 16:03:23.820987 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:23 crc kubenswrapper[4716]: I1207 16:03:23.821000 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:23 crc kubenswrapper[4716]: I1207 16:03:23.821011 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:23Z","lastTransitionTime":"2025-12-07T16:03:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:23 crc kubenswrapper[4716]: I1207 16:03:23.923483 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:23 crc kubenswrapper[4716]: I1207 16:03:23.923528 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:23 crc kubenswrapper[4716]: I1207 16:03:23.923540 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:23 crc kubenswrapper[4716]: I1207 16:03:23.923556 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:23 crc kubenswrapper[4716]: I1207 16:03:23.923568 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:23Z","lastTransitionTime":"2025-12-07T16:03:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:24 crc kubenswrapper[4716]: I1207 16:03:24.025655 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:24 crc kubenswrapper[4716]: I1207 16:03:24.025712 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:24 crc kubenswrapper[4716]: I1207 16:03:24.025731 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:24 crc kubenswrapper[4716]: I1207 16:03:24.025754 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:24 crc kubenswrapper[4716]: I1207 16:03:24.025770 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:24Z","lastTransitionTime":"2025-12-07T16:03:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:24 crc kubenswrapper[4716]: I1207 16:03:24.128269 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:24 crc kubenswrapper[4716]: I1207 16:03:24.128343 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:24 crc kubenswrapper[4716]: I1207 16:03:24.128365 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:24 crc kubenswrapper[4716]: I1207 16:03:24.128398 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:24 crc kubenswrapper[4716]: I1207 16:03:24.128418 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:24Z","lastTransitionTime":"2025-12-07T16:03:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:24 crc kubenswrapper[4716]: I1207 16:03:24.230387 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:24 crc kubenswrapper[4716]: I1207 16:03:24.230421 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:24 crc kubenswrapper[4716]: I1207 16:03:24.230429 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:24 crc kubenswrapper[4716]: I1207 16:03:24.230443 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:24 crc kubenswrapper[4716]: I1207 16:03:24.230451 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:24Z","lastTransitionTime":"2025-12-07T16:03:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:24 crc kubenswrapper[4716]: I1207 16:03:24.331994 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:24 crc kubenswrapper[4716]: I1207 16:03:24.332035 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:24 crc kubenswrapper[4716]: I1207 16:03:24.332044 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:24 crc kubenswrapper[4716]: I1207 16:03:24.332056 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:24 crc kubenswrapper[4716]: I1207 16:03:24.332066 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:24Z","lastTransitionTime":"2025-12-07T16:03:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:24 crc kubenswrapper[4716]: I1207 16:03:24.433488 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:24 crc kubenswrapper[4716]: I1207 16:03:24.433523 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:24 crc kubenswrapper[4716]: I1207 16:03:24.433535 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:24 crc kubenswrapper[4716]: I1207 16:03:24.433549 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:24 crc kubenswrapper[4716]: I1207 16:03:24.433561 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:24Z","lastTransitionTime":"2025-12-07T16:03:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:24 crc kubenswrapper[4716]: I1207 16:03:24.535986 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:24 crc kubenswrapper[4716]: I1207 16:03:24.536025 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:24 crc kubenswrapper[4716]: I1207 16:03:24.536033 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:24 crc kubenswrapper[4716]: I1207 16:03:24.536048 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:24 crc kubenswrapper[4716]: I1207 16:03:24.536058 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:24Z","lastTransitionTime":"2025-12-07T16:03:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:24 crc kubenswrapper[4716]: I1207 16:03:24.637657 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:24 crc kubenswrapper[4716]: I1207 16:03:24.637726 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:24 crc kubenswrapper[4716]: I1207 16:03:24.637738 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:24 crc kubenswrapper[4716]: I1207 16:03:24.637754 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:24 crc kubenswrapper[4716]: I1207 16:03:24.637767 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:24Z","lastTransitionTime":"2025-12-07T16:03:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:24 crc kubenswrapper[4716]: I1207 16:03:24.656938 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 16:03:24 crc kubenswrapper[4716]: E1207 16:03:24.657063 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 16:03:24 crc kubenswrapper[4716]: I1207 16:03:24.657847 4716 scope.go:117] "RemoveContainer" containerID="f48167b38ff840f85f11b69d8090ec6c9c634ed26682daef93f5da99fc02ef61" Dec 07 16:03:24 crc kubenswrapper[4716]: E1207 16:03:24.658100 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qvc7p_openshift-ovn-kubernetes(c54d618a-dfec-4f2a-a288-f5188c8a128a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" podUID="c54d618a-dfec-4f2a-a288-f5188c8a128a" Dec 07 16:03:24 crc kubenswrapper[4716]: I1207 16:03:24.739344 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:24 crc kubenswrapper[4716]: I1207 16:03:24.739386 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:24 crc kubenswrapper[4716]: I1207 16:03:24.739394 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:24 crc kubenswrapper[4716]: I1207 16:03:24.739406 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:24 crc kubenswrapper[4716]: I1207 16:03:24.739417 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:24Z","lastTransitionTime":"2025-12-07T16:03:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:24 crc kubenswrapper[4716]: I1207 16:03:24.841690 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:24 crc kubenswrapper[4716]: I1207 16:03:24.841729 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:24 crc kubenswrapper[4716]: I1207 16:03:24.841737 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:24 crc kubenswrapper[4716]: I1207 16:03:24.841751 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:24 crc kubenswrapper[4716]: I1207 16:03:24.841763 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:24Z","lastTransitionTime":"2025-12-07T16:03:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:24 crc kubenswrapper[4716]: I1207 16:03:24.944173 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:24 crc kubenswrapper[4716]: I1207 16:03:24.944209 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:24 crc kubenswrapper[4716]: I1207 16:03:24.944217 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:24 crc kubenswrapper[4716]: I1207 16:03:24.944231 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:24 crc kubenswrapper[4716]: I1207 16:03:24.944241 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:24Z","lastTransitionTime":"2025-12-07T16:03:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:25 crc kubenswrapper[4716]: I1207 16:03:25.046816 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:25 crc kubenswrapper[4716]: I1207 16:03:25.046852 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:25 crc kubenswrapper[4716]: I1207 16:03:25.046863 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:25 crc kubenswrapper[4716]: I1207 16:03:25.046878 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:25 crc kubenswrapper[4716]: I1207 16:03:25.046891 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:25Z","lastTransitionTime":"2025-12-07T16:03:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:25 crc kubenswrapper[4716]: I1207 16:03:25.149011 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:25 crc kubenswrapper[4716]: I1207 16:03:25.149044 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:25 crc kubenswrapper[4716]: I1207 16:03:25.149054 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:25 crc kubenswrapper[4716]: I1207 16:03:25.149066 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:25 crc kubenswrapper[4716]: I1207 16:03:25.149088 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:25Z","lastTransitionTime":"2025-12-07T16:03:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:25 crc kubenswrapper[4716]: I1207 16:03:25.250806 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:25 crc kubenswrapper[4716]: I1207 16:03:25.250842 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:25 crc kubenswrapper[4716]: I1207 16:03:25.250852 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:25 crc kubenswrapper[4716]: I1207 16:03:25.250865 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:25 crc kubenswrapper[4716]: I1207 16:03:25.250875 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:25Z","lastTransitionTime":"2025-12-07T16:03:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:25 crc kubenswrapper[4716]: I1207 16:03:25.353145 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:25 crc kubenswrapper[4716]: I1207 16:03:25.353197 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:25 crc kubenswrapper[4716]: I1207 16:03:25.353206 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:25 crc kubenswrapper[4716]: I1207 16:03:25.353222 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:25 crc kubenswrapper[4716]: I1207 16:03:25.353230 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:25Z","lastTransitionTime":"2025-12-07T16:03:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:25 crc kubenswrapper[4716]: I1207 16:03:25.455503 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:25 crc kubenswrapper[4716]: I1207 16:03:25.455549 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:25 crc kubenswrapper[4716]: I1207 16:03:25.455562 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:25 crc kubenswrapper[4716]: I1207 16:03:25.455580 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:25 crc kubenswrapper[4716]: I1207 16:03:25.455591 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:25Z","lastTransitionTime":"2025-12-07T16:03:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:25 crc kubenswrapper[4716]: I1207 16:03:25.558146 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:25 crc kubenswrapper[4716]: I1207 16:03:25.558185 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:25 crc kubenswrapper[4716]: I1207 16:03:25.558194 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:25 crc kubenswrapper[4716]: I1207 16:03:25.558208 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:25 crc kubenswrapper[4716]: I1207 16:03:25.558218 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:25Z","lastTransitionTime":"2025-12-07T16:03:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:25 crc kubenswrapper[4716]: I1207 16:03:25.656951 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 16:03:25 crc kubenswrapper[4716]: I1207 16:03:25.657028 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nx2lh" Dec 07 16:03:25 crc kubenswrapper[4716]: E1207 16:03:25.657131 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 16:03:25 crc kubenswrapper[4716]: I1207 16:03:25.657173 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:03:25 crc kubenswrapper[4716]: E1207 16:03:25.657210 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nx2lh" podUID="04b37f6c-457b-4f9a-8733-e88cb9644483" Dec 07 16:03:25 crc kubenswrapper[4716]: E1207 16:03:25.657333 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 16:03:25 crc kubenswrapper[4716]: I1207 16:03:25.660247 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:25 crc kubenswrapper[4716]: I1207 16:03:25.660275 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:25 crc kubenswrapper[4716]: I1207 16:03:25.660285 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:25 crc kubenswrapper[4716]: I1207 16:03:25.660845 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:25 crc kubenswrapper[4716]: I1207 16:03:25.660902 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:25Z","lastTransitionTime":"2025-12-07T16:03:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:25 crc kubenswrapper[4716]: I1207 16:03:25.763508 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:25 crc kubenswrapper[4716]: I1207 16:03:25.763549 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:25 crc kubenswrapper[4716]: I1207 16:03:25.763562 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:25 crc kubenswrapper[4716]: I1207 16:03:25.763579 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:25 crc kubenswrapper[4716]: I1207 16:03:25.763589 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:25Z","lastTransitionTime":"2025-12-07T16:03:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:25 crc kubenswrapper[4716]: I1207 16:03:25.866713 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:25 crc kubenswrapper[4716]: I1207 16:03:25.866757 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:25 crc kubenswrapper[4716]: I1207 16:03:25.866768 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:25 crc kubenswrapper[4716]: I1207 16:03:25.866782 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:25 crc kubenswrapper[4716]: I1207 16:03:25.866791 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:25Z","lastTransitionTime":"2025-12-07T16:03:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:25 crc kubenswrapper[4716]: I1207 16:03:25.969120 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:25 crc kubenswrapper[4716]: I1207 16:03:25.969165 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:25 crc kubenswrapper[4716]: I1207 16:03:25.969177 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:25 crc kubenswrapper[4716]: I1207 16:03:25.969193 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:25 crc kubenswrapper[4716]: I1207 16:03:25.969204 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:25Z","lastTransitionTime":"2025-12-07T16:03:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:26 crc kubenswrapper[4716]: I1207 16:03:26.071649 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:26 crc kubenswrapper[4716]: I1207 16:03:26.071693 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:26 crc kubenswrapper[4716]: I1207 16:03:26.071705 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:26 crc kubenswrapper[4716]: I1207 16:03:26.071721 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:26 crc kubenswrapper[4716]: I1207 16:03:26.071730 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:26Z","lastTransitionTime":"2025-12-07T16:03:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:26 crc kubenswrapper[4716]: I1207 16:03:26.121579 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/04b37f6c-457b-4f9a-8733-e88cb9644483-metrics-certs\") pod \"network-metrics-daemon-nx2lh\" (UID: \"04b37f6c-457b-4f9a-8733-e88cb9644483\") " pod="openshift-multus/network-metrics-daemon-nx2lh" Dec 07 16:03:26 crc kubenswrapper[4716]: E1207 16:03:26.121718 4716 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 07 16:03:26 crc kubenswrapper[4716]: E1207 16:03:26.121782 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/04b37f6c-457b-4f9a-8733-e88cb9644483-metrics-certs podName:04b37f6c-457b-4f9a-8733-e88cb9644483 nodeName:}" failed. No retries permitted until 2025-12-07 16:03:58.121754279 +0000 UTC m=+100.812039191 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/04b37f6c-457b-4f9a-8733-e88cb9644483-metrics-certs") pod "network-metrics-daemon-nx2lh" (UID: "04b37f6c-457b-4f9a-8733-e88cb9644483") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 07 16:03:26 crc kubenswrapper[4716]: I1207 16:03:26.174165 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:26 crc kubenswrapper[4716]: I1207 16:03:26.174203 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:26 crc kubenswrapper[4716]: I1207 16:03:26.174213 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:26 crc kubenswrapper[4716]: I1207 16:03:26.174227 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:26 crc kubenswrapper[4716]: I1207 16:03:26.174237 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:26Z","lastTransitionTime":"2025-12-07T16:03:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:26 crc kubenswrapper[4716]: I1207 16:03:26.276599 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:26 crc kubenswrapper[4716]: I1207 16:03:26.276638 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:26 crc kubenswrapper[4716]: I1207 16:03:26.276646 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:26 crc kubenswrapper[4716]: I1207 16:03:26.276664 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:26 crc kubenswrapper[4716]: I1207 16:03:26.276676 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:26Z","lastTransitionTime":"2025-12-07T16:03:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:26 crc kubenswrapper[4716]: I1207 16:03:26.378886 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:26 crc kubenswrapper[4716]: I1207 16:03:26.378933 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:26 crc kubenswrapper[4716]: I1207 16:03:26.378945 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:26 crc kubenswrapper[4716]: I1207 16:03:26.378963 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:26 crc kubenswrapper[4716]: I1207 16:03:26.378976 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:26Z","lastTransitionTime":"2025-12-07T16:03:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:26 crc kubenswrapper[4716]: I1207 16:03:26.481486 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:26 crc kubenswrapper[4716]: I1207 16:03:26.481537 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:26 crc kubenswrapper[4716]: I1207 16:03:26.481548 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:26 crc kubenswrapper[4716]: I1207 16:03:26.481566 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:26 crc kubenswrapper[4716]: I1207 16:03:26.481576 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:26Z","lastTransitionTime":"2025-12-07T16:03:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:26 crc kubenswrapper[4716]: I1207 16:03:26.583636 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:26 crc kubenswrapper[4716]: I1207 16:03:26.583674 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:26 crc kubenswrapper[4716]: I1207 16:03:26.583686 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:26 crc kubenswrapper[4716]: I1207 16:03:26.583701 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:26 crc kubenswrapper[4716]: I1207 16:03:26.583713 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:26Z","lastTransitionTime":"2025-12-07T16:03:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:26 crc kubenswrapper[4716]: I1207 16:03:26.656911 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 16:03:26 crc kubenswrapper[4716]: E1207 16:03:26.657051 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 16:03:26 crc kubenswrapper[4716]: I1207 16:03:26.685449 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:26 crc kubenswrapper[4716]: I1207 16:03:26.685483 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:26 crc kubenswrapper[4716]: I1207 16:03:26.685495 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:26 crc kubenswrapper[4716]: I1207 16:03:26.685510 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:26 crc kubenswrapper[4716]: I1207 16:03:26.685524 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:26Z","lastTransitionTime":"2025-12-07T16:03:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:26 crc kubenswrapper[4716]: I1207 16:03:26.787525 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:26 crc kubenswrapper[4716]: I1207 16:03:26.787557 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:26 crc kubenswrapper[4716]: I1207 16:03:26.787568 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:26 crc kubenswrapper[4716]: I1207 16:03:26.787584 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:26 crc kubenswrapper[4716]: I1207 16:03:26.787594 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:26Z","lastTransitionTime":"2025-12-07T16:03:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:26 crc kubenswrapper[4716]: I1207 16:03:26.890704 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:26 crc kubenswrapper[4716]: I1207 16:03:26.890767 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:26 crc kubenswrapper[4716]: I1207 16:03:26.890779 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:26 crc kubenswrapper[4716]: I1207 16:03:26.890797 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:26 crc kubenswrapper[4716]: I1207 16:03:26.890808 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:26Z","lastTransitionTime":"2025-12-07T16:03:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:26 crc kubenswrapper[4716]: I1207 16:03:26.992593 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:26 crc kubenswrapper[4716]: I1207 16:03:26.992662 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:26 crc kubenswrapper[4716]: I1207 16:03:26.992677 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:26 crc kubenswrapper[4716]: I1207 16:03:26.992693 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:26 crc kubenswrapper[4716]: I1207 16:03:26.992706 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:26Z","lastTransitionTime":"2025-12-07T16:03:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.079373 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-k4wr2_0baf8575-26cf-4a72-8a6d-ae94c078eaad/kube-multus/0.log" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.079415 4716 generic.go:334] "Generic (PLEG): container finished" podID="0baf8575-26cf-4a72-8a6d-ae94c078eaad" containerID="f0ca2b3cd5a11033afd219b0125be7ab5acd750935d7eb89a470eff3d93a0838" exitCode=1 Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.079441 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-k4wr2" event={"ID":"0baf8575-26cf-4a72-8a6d-ae94c078eaad","Type":"ContainerDied","Data":"f0ca2b3cd5a11033afd219b0125be7ab5acd750935d7eb89a470eff3d93a0838"} Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.079756 4716 scope.go:117] "RemoveContainer" containerID="f0ca2b3cd5a11033afd219b0125be7ab5acd750935d7eb89a470eff3d93a0838" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.092966 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e903f3c-ba8c-477d-b66a-6d1331a7c64e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ff9c456318bbdb54ec3f1ac3e2c2366968f18ad084158b7100d142acb668948\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1358745250b35426602fc7417e4f85dd411cdd0fb3e8f703c432d72128ca7c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d8c9a434c6cfb9e4fee6a8955f1379291f153c007876d1ffb498ce561c73ba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad8df9b5bd5f5a8dd44ce2899cfa29c19435149b170d339d339183cd3df09e99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad8df9b5bd5f5a8dd44ce2899cfa29c19435149b170d339d339183cd3df09e99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:27Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.100930 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.101145 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.101154 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.101167 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.101178 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:27Z","lastTransitionTime":"2025-12-07T16:03:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.115835 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c54d618a-dfec-4f2a-a288-f5188c8a128a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60324dc7d701bc064e14ae4f474bce10f6ee4ff3c3c942eab749e7166082cf9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ed4f4d3564674bf5cb4e7a76dc0fd95e203ecc8950798e59a1fb6c9ab4a20f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://def9ac840fb8ecf2778bed88ea980eddca7c5b2aba4bd1afb967d62fec3f7bae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b73be664f5412c98b89d971a2cac37c494151c293fdb44a98b390978658a6643\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a847d04e22e98091c0037a89054b9236228b701b8c7ba31940a56df0f37b95c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcae78a6133d97f6a32dd7421ccec4333b970a237404e056a5cc601f8eefdc8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f48167b38ff840f85f11b69d8090ec6c9c634ed26682daef93f5da99fc02ef61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f48167b38ff840f85f11b69d8090ec6c9c634ed26682daef93f5da99fc02ef61\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-07T16:03:08Z\\\",\\\"message\\\":\\\"er, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:08Z is after 2025-08-24T17:21:41Z]\\\\nI1207 16:03:08.453903 6420 services_controller.go:451] Built service openshift-controller-manager-operator/metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-controller-manager-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-controller-manager-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.58\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.Tem\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T16:03:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qvc7p_openshift-ovn-kubernetes(c54d618a-dfec-4f2a-a288-f5188c8a128a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcfdffe822635b0ae13b253e6133449fe00461b2792be43d913c4a77219b8255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9bfe2ffce7a7b7c320474117734fc05246cc2b4fd744833e896e8e7c83f5e4d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9bfe2ffce7a7b7c320474117734fc05246cc2b4fd744833e896e8e7c83f5e4d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qvc7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:27Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.130620 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mx85f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"118dd9ba-437c-4e5c-b5ff-4239ae6d4a63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bcd6c8deb184d0c713c5e7b7cc82fa38d3cd498517911ba7a0ec3cb88a3437a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gg5nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ffc069d2b21fa80871b1b97f2c72c37880d9a309cbf8b7daf7d240f2c332a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gg5nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mx85f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:27Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.142481 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nx2lh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04b37f6c-457b-4f9a-8733-e88cb9644483\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv6td\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv6td\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nx2lh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:27Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.158230 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdee9957-48b9-458f-ba71-e0cd39f80596\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73f85248b4d68a9beecdacb3a01f433e97a0ffea2247ce92a41ef96855509377\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6276df652efe6ca935e745f24ef20f87a16465475139230c63b9999234a4d12a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3b1dd202586ee755ca89fbe12b30bbb12f95a62514f6e4d93cbbdbe55181e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd318647816e4a27849666b0ec4fc2ca4905295661c557dc381e1144f22cf284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d31de0326a05216fa10442bdb47151c6e073c97e9a16c34e793028e5e253756\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"le observer\\\\nW1207 16:02:34.645464 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1207 16:02:34.645587 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 16:02:34.646501 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2082283479/tls.crt::/tmp/serving-cert-2082283479/tls.key\\\\\\\"\\\\nI1207 16:02:34.972041 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 16:02:34.973702 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 16:02:34.973717 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 16:02:34.973738 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 16:02:34.973743 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 16:02:34.977185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1207 16:02:34.977204 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1207 16:02:34.977209 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977217 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 16:02:34.977225 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 16:02:34.977228 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 16:02:34.977232 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1207 16:02:34.979033 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27bd4a54f920b2725987c1c7fd2d9fdf16cd08738c246c333a4e9164a3a582a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:27Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.181290 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:27Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.194502 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2jjld" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dee6f5c-861c-4819-9309-7791574c7b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51c1d4159ab11087839a2c2d192566fb9322447c26df560bf09517c906ececb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmdns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2jjld\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:27Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.203459 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.203499 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.203510 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.203525 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.203535 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:27Z","lastTransitionTime":"2025-12-07T16:03:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.217369 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4wr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0baf8575-26cf-4a72-8a6d-ae94c078eaad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0ca2b3cd5a11033afd219b0125be7ab5acd750935d7eb89a470eff3d93a0838\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0ca2b3cd5a11033afd219b0125be7ab5acd750935d7eb89a470eff3d93a0838\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-07T16:03:26Z\\\",\\\"message\\\":\\\"2025-12-07T16:02:41+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_09397239-db79-454b-b301-47a1e4cdfc48\\\\n2025-12-07T16:02:41+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_09397239-db79-454b-b301-47a1e4cdfc48 to /host/opt/cni/bin/\\\\n2025-12-07T16:02:41Z [verbose] multus-daemon started\\\\n2025-12-07T16:02:41Z [verbose] Readiness Indicator file check\\\\n2025-12-07T16:03:26Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p2trd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4wr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:27Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.231004 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-m5jl5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"824ee59f-1474-4994-a70e-462878521514\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a15fe50cc75515704779646cc19aa50004f0db9aa26fe2535110df749bbfb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhwzn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-m5jl5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:27Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.251815 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef3335ae3875a75ab64685fc7e3590177280d651e8f8292c6f5ea6f472e55789\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:27Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.270807 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:27Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.289119 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f95ef61779507288a03bda8eddf745e105f705998a117a30789675bfa452d88f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b955b23163e7c8c458a3bbe2f681633e6a663ce76748febcc0b3a3c37f3ca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:27Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.306135 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.306562 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.306633 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.306973 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.307156 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:27Z","lastTransitionTime":"2025-12-07T16:03:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.319669 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7961cc9a-734e-42c2-b1ff-7075f3f20091\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84a36b76f20618c7885d611ecad9db29bac5101c27572be31d3acbb90c581291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0a0eca3c2087a457c0e448e649345ec4595a9c8a019aed2cddc5f30c77871e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://404426a171b8676eddd4a00996c76afab7e269643fa9608f1e1b6bdba385071d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fd3c2941782921f99a6cdc3de4e892234548bde527fec15a8ac92e513566e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e23abfdca314ed43d594333c9f9e401cc8ef7dad39b8ebac84a836b06ee5c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09b38cd4b389ea477c802a842ee4f2297a64d64684624117163e819539c257c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09b38cd4b389ea477c802a842ee4f2297a64d64684624117163e819539c257c6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c00f7a8e024e0e0dfd7b73e16d1d5437602e79ea95b73c21f24b209edef60e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c00f7a8e024e0e0dfd7b73e16d1d5437602e79ea95b73c21f24b209edef60e98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3f6392d426fa2cd788ec3a5c3bff4ac39483a54dfb2624a1c4370522413a350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3f6392d426fa2cd788ec3a5c3bff4ac39483a54dfb2624a1c4370522413a350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:27Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.335680 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3067de7c-c11c-4be1-bd94-d6572c266795\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d14d018fa044e9e80fe81e7e38df935992d14c773ed2b2193fd5cb873e5d0947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a9e9cf048220ca14c996847e9ced9f2367b9331dc6c2836dd023658855cd79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6d3c911874095b3c7cb062d03070e5090a7bb2456cb20cb4ce078f502515cda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ac6ce7701f1c6aa1ea8fcf7a3035e3b0eaa28ae3cb6afc442372bd9117bdde4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:27Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.355864 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.356219 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.356378 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.356528 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.356666 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:27Z","lastTransitionTime":"2025-12-07T16:03:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.365534 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4742ea84b3a2c9461d9b900beee7354a2ded24c39ad23e007cbf270ae12377c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:27Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:27 crc kubenswrapper[4716]: E1207 16:03:27.373826 4716 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a8d1b7d4-7ec4-40c7-86b0-291db1f38894\\\",\\\"systemUUID\\\":\\\"a57d09b0-0224-46eb-a3ac-a2b60dad2250\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:27Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.378242 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.378295 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.378308 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.378330 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.378343 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:27Z","lastTransitionTime":"2025-12-07T16:03:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.384070 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:27Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.397897 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c15b59eb-565d-4556-a4ce-75afdf159dc8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://530a2935826940767ea33fb4f37fa11f7dee4950d867fbcd6af72836d80f7db3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jks8z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ccaf6c1e6a978eb978a1c207b6c4c00b23737f78b743f966579aca781e22e0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jks8z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zcxxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:27Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:27 crc kubenswrapper[4716]: E1207 16:03:27.407143 4716 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a8d1b7d4-7ec4-40c7-86b0-291db1f38894\\\",\\\"systemUUID\\\":\\\"a57d09b0-0224-46eb-a3ac-a2b60dad2250\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:27Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.410956 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.410977 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.410985 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.410999 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.411008 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:27Z","lastTransitionTime":"2025-12-07T16:03:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.412485 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8ps9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98015cea-d753-4e7a-b2b5-5ae7733a81e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d68509b673eecd37725cdcc48e3d0612c3b9c7601433f2b72c21d662140daa59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f84f52a073437543721010dc440339654ffb5e04eeb6cc7afbcc62f00f64c8e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f84f52a073437543721010dc440339654ffb5e04eeb6cc7afbcc62f00f64c8e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f6fe1cb25218cb3ac20488dc47efa4dc6a26a43dbc8306b5ee38cbe1a186c67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f6fe1cb25218cb3ac20488dc47efa4dc6a26a43dbc8306b5ee38cbe1a186c67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14b748a8247f6863e8be61e90057a06626dfd3cd4f01bd3a11629f700f56378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f14b748a8247f6863e8be61e90057a06626dfd3cd4f01bd3a11629f700f56378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ec379c0d573cb9b5ffcfc5ca7c653ba60e2fbdd7e3d37836102e26618ce43e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ec379c0d573cb9b5ffcfc5ca7c653ba60e2fbdd7e3d37836102e26618ce43e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c75bf2a6703cf99865d4ed5fa000e1e3b1330d4946176cf585fec1a2f226df0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c75bf2a6703cf99865d4ed5fa000e1e3b1330d4946176cf585fec1a2f226df0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de45812c9a5a5bfe6b0146828d03fab28362b03d04ef861c150b9ffbf181d345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de45812c9a5a5bfe6b0146828d03fab28362b03d04ef861c150b9ffbf181d345\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8ps9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:27Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:27 crc kubenswrapper[4716]: E1207 16:03:27.429299 4716 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a8d1b7d4-7ec4-40c7-86b0-291db1f38894\\\",\\\"systemUUID\\\":\\\"a57d09b0-0224-46eb-a3ac-a2b60dad2250\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:27Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.432530 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.432588 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.432606 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.432617 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.432626 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:27Z","lastTransitionTime":"2025-12-07T16:03:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:27 crc kubenswrapper[4716]: E1207 16:03:27.452495 4716 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a8d1b7d4-7ec4-40c7-86b0-291db1f38894\\\",\\\"systemUUID\\\":\\\"a57d09b0-0224-46eb-a3ac-a2b60dad2250\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:27Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.456336 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.456393 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.456412 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.456438 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.456457 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:27Z","lastTransitionTime":"2025-12-07T16:03:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:27 crc kubenswrapper[4716]: E1207 16:03:27.475581 4716 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a8d1b7d4-7ec4-40c7-86b0-291db1f38894\\\",\\\"systemUUID\\\":\\\"a57d09b0-0224-46eb-a3ac-a2b60dad2250\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:27Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:27 crc kubenswrapper[4716]: E1207 16:03:27.475702 4716 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.477475 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.477616 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.477731 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.477848 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.477965 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:27Z","lastTransitionTime":"2025-12-07T16:03:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.579887 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.579926 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.579937 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.579952 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.579962 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:27Z","lastTransitionTime":"2025-12-07T16:03:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.657569 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.657604 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nx2lh" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.657659 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:03:27 crc kubenswrapper[4716]: E1207 16:03:27.657711 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 16:03:27 crc kubenswrapper[4716]: E1207 16:03:27.657828 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 16:03:27 crc kubenswrapper[4716]: E1207 16:03:27.657907 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nx2lh" podUID="04b37f6c-457b-4f9a-8733-e88cb9644483" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.675961 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e903f3c-ba8c-477d-b66a-6d1331a7c64e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ff9c456318bbdb54ec3f1ac3e2c2366968f18ad084158b7100d142acb668948\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1358745250b35426602fc7417e4f85dd411cdd0fb3e8f703c432d72128ca7c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d8c9a434c6cfb9e4fee6a8955f1379291f153c007876d1ffb498ce561c73ba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad8df9b5bd5f5a8dd44ce2899cfa29c19435149b170d339d339183cd3df09e99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad8df9b5bd5f5a8dd44ce2899cfa29c19435149b170d339d339183cd3df09e99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:27Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.682292 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.682328 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.682337 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.682350 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.682360 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:27Z","lastTransitionTime":"2025-12-07T16:03:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.698813 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c54d618a-dfec-4f2a-a288-f5188c8a128a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60324dc7d701bc064e14ae4f474bce10f6ee4ff3c3c942eab749e7166082cf9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ed4f4d3564674bf5cb4e7a76dc0fd95e203ecc8950798e59a1fb6c9ab4a20f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://def9ac840fb8ecf2778bed88ea980eddca7c5b2aba4bd1afb967d62fec3f7bae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b73be664f5412c98b89d971a2cac37c494151c293fdb44a98b390978658a6643\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a847d04e22e98091c0037a89054b9236228b701b8c7ba31940a56df0f37b95c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcae78a6133d97f6a32dd7421ccec4333b970a237404e056a5cc601f8eefdc8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f48167b38ff840f85f11b69d8090ec6c9c634ed26682daef93f5da99fc02ef61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f48167b38ff840f85f11b69d8090ec6c9c634ed26682daef93f5da99fc02ef61\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-07T16:03:08Z\\\",\\\"message\\\":\\\"er, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:08Z is after 2025-08-24T17:21:41Z]\\\\nI1207 16:03:08.453903 6420 services_controller.go:451] Built service openshift-controller-manager-operator/metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-controller-manager-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-controller-manager-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.58\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.Tem\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T16:03:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qvc7p_openshift-ovn-kubernetes(c54d618a-dfec-4f2a-a288-f5188c8a128a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcfdffe822635b0ae13b253e6133449fe00461b2792be43d913c4a77219b8255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9bfe2ffce7a7b7c320474117734fc05246cc2b4fd744833e896e8e7c83f5e4d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9bfe2ffce7a7b7c320474117734fc05246cc2b4fd744833e896e8e7c83f5e4d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qvc7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:27Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.710807 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mx85f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"118dd9ba-437c-4e5c-b5ff-4239ae6d4a63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bcd6c8deb184d0c713c5e7b7cc82fa38d3cd498517911ba7a0ec3cb88a3437a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gg5nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ffc069d2b21fa80871b1b97f2c72c37880d9a309cbf8b7daf7d240f2c332a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gg5nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mx85f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:27Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.727980 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nx2lh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04b37f6c-457b-4f9a-8733-e88cb9644483\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv6td\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv6td\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nx2lh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:27Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.743870 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdee9957-48b9-458f-ba71-e0cd39f80596\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73f85248b4d68a9beecdacb3a01f433e97a0ffea2247ce92a41ef96855509377\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6276df652efe6ca935e745f24ef20f87a16465475139230c63b9999234a4d12a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3b1dd202586ee755ca89fbe12b30bbb12f95a62514f6e4d93cbbdbe55181e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd318647816e4a27849666b0ec4fc2ca4905295661c557dc381e1144f22cf284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d31de0326a05216fa10442bdb47151c6e073c97e9a16c34e793028e5e253756\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"le observer\\\\nW1207 16:02:34.645464 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1207 16:02:34.645587 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 16:02:34.646501 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2082283479/tls.crt::/tmp/serving-cert-2082283479/tls.key\\\\\\\"\\\\nI1207 16:02:34.972041 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 16:02:34.973702 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 16:02:34.973717 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 16:02:34.973738 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 16:02:34.973743 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 16:02:34.977185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1207 16:02:34.977204 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1207 16:02:34.977209 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977217 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 16:02:34.977225 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 16:02:34.977228 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 16:02:34.977232 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1207 16:02:34.979033 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27bd4a54f920b2725987c1c7fd2d9fdf16cd08738c246c333a4e9164a3a582a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:27Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.764672 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:27Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.776756 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2jjld" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dee6f5c-861c-4819-9309-7791574c7b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51c1d4159ab11087839a2c2d192566fb9322447c26df560bf09517c906ececb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmdns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2jjld\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:27Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.784837 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.784930 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.784948 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.784968 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.784983 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:27Z","lastTransitionTime":"2025-12-07T16:03:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.794280 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4wr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0baf8575-26cf-4a72-8a6d-ae94c078eaad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0ca2b3cd5a11033afd219b0125be7ab5acd750935d7eb89a470eff3d93a0838\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0ca2b3cd5a11033afd219b0125be7ab5acd750935d7eb89a470eff3d93a0838\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-07T16:03:26Z\\\",\\\"message\\\":\\\"2025-12-07T16:02:41+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_09397239-db79-454b-b301-47a1e4cdfc48\\\\n2025-12-07T16:02:41+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_09397239-db79-454b-b301-47a1e4cdfc48 to /host/opt/cni/bin/\\\\n2025-12-07T16:02:41Z [verbose] multus-daemon started\\\\n2025-12-07T16:02:41Z [verbose] Readiness Indicator file check\\\\n2025-12-07T16:03:26Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p2trd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4wr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:27Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.804551 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-m5jl5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"824ee59f-1474-4994-a70e-462878521514\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a15fe50cc75515704779646cc19aa50004f0db9aa26fe2535110df749bbfb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhwzn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-m5jl5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:27Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.817328 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef3335ae3875a75ab64685fc7e3590177280d651e8f8292c6f5ea6f472e55789\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:27Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.834283 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:27Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.848383 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f95ef61779507288a03bda8eddf745e105f705998a117a30789675bfa452d88f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b955b23163e7c8c458a3bbe2f681633e6a663ce76748febcc0b3a3c37f3ca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:27Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.870563 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7961cc9a-734e-42c2-b1ff-7075f3f20091\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84a36b76f20618c7885d611ecad9db29bac5101c27572be31d3acbb90c581291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0a0eca3c2087a457c0e448e649345ec4595a9c8a019aed2cddc5f30c77871e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://404426a171b8676eddd4a00996c76afab7e269643fa9608f1e1b6bdba385071d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fd3c2941782921f99a6cdc3de4e892234548bde527fec15a8ac92e513566e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e23abfdca314ed43d594333c9f9e401cc8ef7dad39b8ebac84a836b06ee5c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09b38cd4b389ea477c802a842ee4f2297a64d64684624117163e819539c257c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09b38cd4b389ea477c802a842ee4f2297a64d64684624117163e819539c257c6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c00f7a8e024e0e0dfd7b73e16d1d5437602e79ea95b73c21f24b209edef60e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c00f7a8e024e0e0dfd7b73e16d1d5437602e79ea95b73c21f24b209edef60e98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3f6392d426fa2cd788ec3a5c3bff4ac39483a54dfb2624a1c4370522413a350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3f6392d426fa2cd788ec3a5c3bff4ac39483a54dfb2624a1c4370522413a350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:27Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.886226 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3067de7c-c11c-4be1-bd94-d6572c266795\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d14d018fa044e9e80fe81e7e38df935992d14c773ed2b2193fd5cb873e5d0947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a9e9cf048220ca14c996847e9ced9f2367b9331dc6c2836dd023658855cd79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6d3c911874095b3c7cb062d03070e5090a7bb2456cb20cb4ce078f502515cda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ac6ce7701f1c6aa1ea8fcf7a3035e3b0eaa28ae3cb6afc442372bd9117bdde4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:27Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.887579 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.887608 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.887617 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.887631 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.887639 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:27Z","lastTransitionTime":"2025-12-07T16:03:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.900726 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4742ea84b3a2c9461d9b900beee7354a2ded24c39ad23e007cbf270ae12377c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:27Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.913755 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:27Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.925993 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c15b59eb-565d-4556-a4ce-75afdf159dc8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://530a2935826940767ea33fb4f37fa11f7dee4950d867fbcd6af72836d80f7db3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jks8z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ccaf6c1e6a978eb978a1c207b6c4c00b23737f78b743f966579aca781e22e0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jks8z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zcxxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:27Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.944240 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8ps9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98015cea-d753-4e7a-b2b5-5ae7733a81e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d68509b673eecd37725cdcc48e3d0612c3b9c7601433f2b72c21d662140daa59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f84f52a073437543721010dc440339654ffb5e04eeb6cc7afbcc62f00f64c8e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f84f52a073437543721010dc440339654ffb5e04eeb6cc7afbcc62f00f64c8e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f6fe1cb25218cb3ac20488dc47efa4dc6a26a43dbc8306b5ee38cbe1a186c67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f6fe1cb25218cb3ac20488dc47efa4dc6a26a43dbc8306b5ee38cbe1a186c67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14b748a8247f6863e8be61e90057a06626dfd3cd4f01bd3a11629f700f56378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f14b748a8247f6863e8be61e90057a06626dfd3cd4f01bd3a11629f700f56378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ec379c0d573cb9b5ffcfc5ca7c653ba60e2fbdd7e3d37836102e26618ce43e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ec379c0d573cb9b5ffcfc5ca7c653ba60e2fbdd7e3d37836102e26618ce43e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c75bf2a6703cf99865d4ed5fa000e1e3b1330d4946176cf585fec1a2f226df0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c75bf2a6703cf99865d4ed5fa000e1e3b1330d4946176cf585fec1a2f226df0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de45812c9a5a5bfe6b0146828d03fab28362b03d04ef861c150b9ffbf181d345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de45812c9a5a5bfe6b0146828d03fab28362b03d04ef861c150b9ffbf181d345\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8ps9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:27Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.991441 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.991648 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.991773 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.991873 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:27 crc kubenswrapper[4716]: I1207 16:03:27.991966 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:27Z","lastTransitionTime":"2025-12-07T16:03:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:28 crc kubenswrapper[4716]: I1207 16:03:28.083591 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-k4wr2_0baf8575-26cf-4a72-8a6d-ae94c078eaad/kube-multus/0.log" Dec 07 16:03:28 crc kubenswrapper[4716]: I1207 16:03:28.083642 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-k4wr2" event={"ID":"0baf8575-26cf-4a72-8a6d-ae94c078eaad","Type":"ContainerStarted","Data":"6fca3b2ea503451cfaf7afb00445fe0adb0acc87d436d1eebba5068452b8e7cb"} Dec 07 16:03:28 crc kubenswrapper[4716]: I1207 16:03:28.094294 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:28 crc kubenswrapper[4716]: I1207 16:03:28.094324 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:28 crc kubenswrapper[4716]: I1207 16:03:28.094332 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:28 crc kubenswrapper[4716]: I1207 16:03:28.094344 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:28 crc kubenswrapper[4716]: I1207 16:03:28.094353 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:28Z","lastTransitionTime":"2025-12-07T16:03:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:28 crc kubenswrapper[4716]: I1207 16:03:28.101559 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7961cc9a-734e-42c2-b1ff-7075f3f20091\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84a36b76f20618c7885d611ecad9db29bac5101c27572be31d3acbb90c581291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0a0eca3c2087a457c0e448e649345ec4595a9c8a019aed2cddc5f30c77871e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://404426a171b8676eddd4a00996c76afab7e269643fa9608f1e1b6bdba385071d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fd3c2941782921f99a6cdc3de4e892234548bde527fec15a8ac92e513566e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e23abfdca314ed43d594333c9f9e401cc8ef7dad39b8ebac84a836b06ee5c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09b38cd4b389ea477c802a842ee4f2297a64d64684624117163e819539c257c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09b38cd4b389ea477c802a842ee4f2297a64d64684624117163e819539c257c6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c00f7a8e024e0e0dfd7b73e16d1d5437602e79ea95b73c21f24b209edef60e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c00f7a8e024e0e0dfd7b73e16d1d5437602e79ea95b73c21f24b209edef60e98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3f6392d426fa2cd788ec3a5c3bff4ac39483a54dfb2624a1c4370522413a350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3f6392d426fa2cd788ec3a5c3bff4ac39483a54dfb2624a1c4370522413a350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:28Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:28 crc kubenswrapper[4716]: I1207 16:03:28.112189 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3067de7c-c11c-4be1-bd94-d6572c266795\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d14d018fa044e9e80fe81e7e38df935992d14c773ed2b2193fd5cb873e5d0947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a9e9cf048220ca14c996847e9ced9f2367b9331dc6c2836dd023658855cd79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6d3c911874095b3c7cb062d03070e5090a7bb2456cb20cb4ce078f502515cda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ac6ce7701f1c6aa1ea8fcf7a3035e3b0eaa28ae3cb6afc442372bd9117bdde4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:28Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:28 crc kubenswrapper[4716]: I1207 16:03:28.121943 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4742ea84b3a2c9461d9b900beee7354a2ded24c39ad23e007cbf270ae12377c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:28Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:28 crc kubenswrapper[4716]: I1207 16:03:28.132858 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:28Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:28 crc kubenswrapper[4716]: I1207 16:03:28.141960 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c15b59eb-565d-4556-a4ce-75afdf159dc8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://530a2935826940767ea33fb4f37fa11f7dee4950d867fbcd6af72836d80f7db3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jks8z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ccaf6c1e6a978eb978a1c207b6c4c00b23737f78b743f966579aca781e22e0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jks8z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zcxxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:28Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:28 crc kubenswrapper[4716]: I1207 16:03:28.155905 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8ps9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98015cea-d753-4e7a-b2b5-5ae7733a81e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d68509b673eecd37725cdcc48e3d0612c3b9c7601433f2b72c21d662140daa59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f84f52a073437543721010dc440339654ffb5e04eeb6cc7afbcc62f00f64c8e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f84f52a073437543721010dc440339654ffb5e04eeb6cc7afbcc62f00f64c8e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f6fe1cb25218cb3ac20488dc47efa4dc6a26a43dbc8306b5ee38cbe1a186c67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f6fe1cb25218cb3ac20488dc47efa4dc6a26a43dbc8306b5ee38cbe1a186c67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14b748a8247f6863e8be61e90057a06626dfd3cd4f01bd3a11629f700f56378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f14b748a8247f6863e8be61e90057a06626dfd3cd4f01bd3a11629f700f56378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ec379c0d573cb9b5ffcfc5ca7c653ba60e2fbdd7e3d37836102e26618ce43e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ec379c0d573cb9b5ffcfc5ca7c653ba60e2fbdd7e3d37836102e26618ce43e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c75bf2a6703cf99865d4ed5fa000e1e3b1330d4946176cf585fec1a2f226df0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c75bf2a6703cf99865d4ed5fa000e1e3b1330d4946176cf585fec1a2f226df0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de45812c9a5a5bfe6b0146828d03fab28362b03d04ef861c150b9ffbf181d345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de45812c9a5a5bfe6b0146828d03fab28362b03d04ef861c150b9ffbf181d345\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8ps9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:28Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:28 crc kubenswrapper[4716]: I1207 16:03:28.168668 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e903f3c-ba8c-477d-b66a-6d1331a7c64e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ff9c456318bbdb54ec3f1ac3e2c2366968f18ad084158b7100d142acb668948\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1358745250b35426602fc7417e4f85dd411cdd0fb3e8f703c432d72128ca7c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d8c9a434c6cfb9e4fee6a8955f1379291f153c007876d1ffb498ce561c73ba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad8df9b5bd5f5a8dd44ce2899cfa29c19435149b170d339d339183cd3df09e99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad8df9b5bd5f5a8dd44ce2899cfa29c19435149b170d339d339183cd3df09e99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:28Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:28 crc kubenswrapper[4716]: I1207 16:03:28.189367 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c54d618a-dfec-4f2a-a288-f5188c8a128a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60324dc7d701bc064e14ae4f474bce10f6ee4ff3c3c942eab749e7166082cf9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ed4f4d3564674bf5cb4e7a76dc0fd95e203ecc8950798e59a1fb6c9ab4a20f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://def9ac840fb8ecf2778bed88ea980eddca7c5b2aba4bd1afb967d62fec3f7bae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b73be664f5412c98b89d971a2cac37c494151c293fdb44a98b390978658a6643\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a847d04e22e98091c0037a89054b9236228b701b8c7ba31940a56df0f37b95c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcae78a6133d97f6a32dd7421ccec4333b970a237404e056a5cc601f8eefdc8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f48167b38ff840f85f11b69d8090ec6c9c634ed26682daef93f5da99fc02ef61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f48167b38ff840f85f11b69d8090ec6c9c634ed26682daef93f5da99fc02ef61\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-07T16:03:08Z\\\",\\\"message\\\":\\\"er, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:08Z is after 2025-08-24T17:21:41Z]\\\\nI1207 16:03:08.453903 6420 services_controller.go:451] Built service openshift-controller-manager-operator/metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-controller-manager-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-controller-manager-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.58\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.Tem\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T16:03:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qvc7p_openshift-ovn-kubernetes(c54d618a-dfec-4f2a-a288-f5188c8a128a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcfdffe822635b0ae13b253e6133449fe00461b2792be43d913c4a77219b8255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9bfe2ffce7a7b7c320474117734fc05246cc2b4fd744833e896e8e7c83f5e4d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9bfe2ffce7a7b7c320474117734fc05246cc2b4fd744833e896e8e7c83f5e4d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qvc7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:28Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:28 crc kubenswrapper[4716]: I1207 16:03:28.196282 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:28 crc kubenswrapper[4716]: I1207 16:03:28.196321 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:28 crc kubenswrapper[4716]: I1207 16:03:28.196332 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:28 crc kubenswrapper[4716]: I1207 16:03:28.196346 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:28 crc kubenswrapper[4716]: I1207 16:03:28.196354 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:28Z","lastTransitionTime":"2025-12-07T16:03:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:28 crc kubenswrapper[4716]: I1207 16:03:28.201456 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mx85f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"118dd9ba-437c-4e5c-b5ff-4239ae6d4a63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bcd6c8deb184d0c713c5e7b7cc82fa38d3cd498517911ba7a0ec3cb88a3437a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gg5nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ffc069d2b21fa80871b1b97f2c72c37880d9a309cbf8b7daf7d240f2c332a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gg5nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mx85f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:28Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:28 crc kubenswrapper[4716]: I1207 16:03:28.211587 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nx2lh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04b37f6c-457b-4f9a-8733-e88cb9644483\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv6td\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv6td\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nx2lh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:28Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:28 crc kubenswrapper[4716]: I1207 16:03:28.225463 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdee9957-48b9-458f-ba71-e0cd39f80596\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73f85248b4d68a9beecdacb3a01f433e97a0ffea2247ce92a41ef96855509377\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6276df652efe6ca935e745f24ef20f87a16465475139230c63b9999234a4d12a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3b1dd202586ee755ca89fbe12b30bbb12f95a62514f6e4d93cbbdbe55181e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd318647816e4a27849666b0ec4fc2ca4905295661c557dc381e1144f22cf284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d31de0326a05216fa10442bdb47151c6e073c97e9a16c34e793028e5e253756\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"le observer\\\\nW1207 16:02:34.645464 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1207 16:02:34.645587 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 16:02:34.646501 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2082283479/tls.crt::/tmp/serving-cert-2082283479/tls.key\\\\\\\"\\\\nI1207 16:02:34.972041 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 16:02:34.973702 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 16:02:34.973717 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 16:02:34.973738 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 16:02:34.973743 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 16:02:34.977185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1207 16:02:34.977204 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1207 16:02:34.977209 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977217 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 16:02:34.977225 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 16:02:34.977228 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 16:02:34.977232 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1207 16:02:34.979033 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27bd4a54f920b2725987c1c7fd2d9fdf16cd08738c246c333a4e9164a3a582a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:28Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:28 crc kubenswrapper[4716]: I1207 16:03:28.235403 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:28Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:28 crc kubenswrapper[4716]: I1207 16:03:28.245386 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2jjld" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dee6f5c-861c-4819-9309-7791574c7b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51c1d4159ab11087839a2c2d192566fb9322447c26df560bf09517c906ececb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmdns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2jjld\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:28Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:28 crc kubenswrapper[4716]: I1207 16:03:28.255657 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4wr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0baf8575-26cf-4a72-8a6d-ae94c078eaad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fca3b2ea503451cfaf7afb00445fe0adb0acc87d436d1eebba5068452b8e7cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0ca2b3cd5a11033afd219b0125be7ab5acd750935d7eb89a470eff3d93a0838\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-07T16:03:26Z\\\",\\\"message\\\":\\\"2025-12-07T16:02:41+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_09397239-db79-454b-b301-47a1e4cdfc48\\\\n2025-12-07T16:02:41+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_09397239-db79-454b-b301-47a1e4cdfc48 to /host/opt/cni/bin/\\\\n2025-12-07T16:02:41Z [verbose] multus-daemon started\\\\n2025-12-07T16:02:41Z [verbose] Readiness Indicator file check\\\\n2025-12-07T16:03:26Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:03:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p2trd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4wr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:28Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:28 crc kubenswrapper[4716]: I1207 16:03:28.263464 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-m5jl5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"824ee59f-1474-4994-a70e-462878521514\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a15fe50cc75515704779646cc19aa50004f0db9aa26fe2535110df749bbfb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhwzn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-m5jl5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:28Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:28 crc kubenswrapper[4716]: I1207 16:03:28.272880 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef3335ae3875a75ab64685fc7e3590177280d651e8f8292c6f5ea6f472e55789\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:28Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:28 crc kubenswrapper[4716]: I1207 16:03:28.283220 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:28Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:28 crc kubenswrapper[4716]: I1207 16:03:28.297746 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f95ef61779507288a03bda8eddf745e105f705998a117a30789675bfa452d88f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b955b23163e7c8c458a3bbe2f681633e6a663ce76748febcc0b3a3c37f3ca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:28Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:28 crc kubenswrapper[4716]: I1207 16:03:28.298051 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:28 crc kubenswrapper[4716]: I1207 16:03:28.298127 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:28 crc kubenswrapper[4716]: I1207 16:03:28.298139 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:28 crc kubenswrapper[4716]: I1207 16:03:28.298155 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:28 crc kubenswrapper[4716]: I1207 16:03:28.298165 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:28Z","lastTransitionTime":"2025-12-07T16:03:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:28 crc kubenswrapper[4716]: I1207 16:03:28.399594 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:28 crc kubenswrapper[4716]: I1207 16:03:28.399884 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:28 crc kubenswrapper[4716]: I1207 16:03:28.400014 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:28 crc kubenswrapper[4716]: I1207 16:03:28.400198 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:28 crc kubenswrapper[4716]: I1207 16:03:28.400305 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:28Z","lastTransitionTime":"2025-12-07T16:03:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:28 crc kubenswrapper[4716]: I1207 16:03:28.502011 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:28 crc kubenswrapper[4716]: I1207 16:03:28.502037 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:28 crc kubenswrapper[4716]: I1207 16:03:28.502049 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:28 crc kubenswrapper[4716]: I1207 16:03:28.502061 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:28 crc kubenswrapper[4716]: I1207 16:03:28.502070 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:28Z","lastTransitionTime":"2025-12-07T16:03:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:28 crc kubenswrapper[4716]: I1207 16:03:28.604688 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:28 crc kubenswrapper[4716]: I1207 16:03:28.604729 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:28 crc kubenswrapper[4716]: I1207 16:03:28.604738 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:28 crc kubenswrapper[4716]: I1207 16:03:28.604753 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:28 crc kubenswrapper[4716]: I1207 16:03:28.604763 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:28Z","lastTransitionTime":"2025-12-07T16:03:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:28 crc kubenswrapper[4716]: I1207 16:03:28.657489 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 16:03:28 crc kubenswrapper[4716]: E1207 16:03:28.657576 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 16:03:28 crc kubenswrapper[4716]: I1207 16:03:28.707035 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:28 crc kubenswrapper[4716]: I1207 16:03:28.707090 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:28 crc kubenswrapper[4716]: I1207 16:03:28.707100 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:28 crc kubenswrapper[4716]: I1207 16:03:28.707111 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:28 crc kubenswrapper[4716]: I1207 16:03:28.707121 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:28Z","lastTransitionTime":"2025-12-07T16:03:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:28 crc kubenswrapper[4716]: I1207 16:03:28.813799 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:28 crc kubenswrapper[4716]: I1207 16:03:28.813851 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:28 crc kubenswrapper[4716]: I1207 16:03:28.813876 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:28 crc kubenswrapper[4716]: I1207 16:03:28.813902 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:28 crc kubenswrapper[4716]: I1207 16:03:28.813919 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:28Z","lastTransitionTime":"2025-12-07T16:03:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:28 crc kubenswrapper[4716]: I1207 16:03:28.916601 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:28 crc kubenswrapper[4716]: I1207 16:03:28.916671 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:28 crc kubenswrapper[4716]: I1207 16:03:28.916695 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:28 crc kubenswrapper[4716]: I1207 16:03:28.916722 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:28 crc kubenswrapper[4716]: I1207 16:03:28.916738 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:28Z","lastTransitionTime":"2025-12-07T16:03:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:29 crc kubenswrapper[4716]: I1207 16:03:29.019174 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:29 crc kubenswrapper[4716]: I1207 16:03:29.019205 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:29 crc kubenswrapper[4716]: I1207 16:03:29.019214 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:29 crc kubenswrapper[4716]: I1207 16:03:29.019227 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:29 crc kubenswrapper[4716]: I1207 16:03:29.019236 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:29Z","lastTransitionTime":"2025-12-07T16:03:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:29 crc kubenswrapper[4716]: I1207 16:03:29.121520 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:29 crc kubenswrapper[4716]: I1207 16:03:29.121558 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:29 crc kubenswrapper[4716]: I1207 16:03:29.121569 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:29 crc kubenswrapper[4716]: I1207 16:03:29.121583 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:29 crc kubenswrapper[4716]: I1207 16:03:29.121593 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:29Z","lastTransitionTime":"2025-12-07T16:03:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:29 crc kubenswrapper[4716]: I1207 16:03:29.223877 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:29 crc kubenswrapper[4716]: I1207 16:03:29.223913 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:29 crc kubenswrapper[4716]: I1207 16:03:29.223921 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:29 crc kubenswrapper[4716]: I1207 16:03:29.223932 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:29 crc kubenswrapper[4716]: I1207 16:03:29.223940 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:29Z","lastTransitionTime":"2025-12-07T16:03:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:29 crc kubenswrapper[4716]: I1207 16:03:29.325964 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:29 crc kubenswrapper[4716]: I1207 16:03:29.326349 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:29 crc kubenswrapper[4716]: I1207 16:03:29.326488 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:29 crc kubenswrapper[4716]: I1207 16:03:29.326630 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:29 crc kubenswrapper[4716]: I1207 16:03:29.326762 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:29Z","lastTransitionTime":"2025-12-07T16:03:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:29 crc kubenswrapper[4716]: I1207 16:03:29.428380 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:29 crc kubenswrapper[4716]: I1207 16:03:29.428402 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:29 crc kubenswrapper[4716]: I1207 16:03:29.428424 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:29 crc kubenswrapper[4716]: I1207 16:03:29.428437 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:29 crc kubenswrapper[4716]: I1207 16:03:29.428446 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:29Z","lastTransitionTime":"2025-12-07T16:03:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:29 crc kubenswrapper[4716]: I1207 16:03:29.530516 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:29 crc kubenswrapper[4716]: I1207 16:03:29.530549 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:29 crc kubenswrapper[4716]: I1207 16:03:29.530563 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:29 crc kubenswrapper[4716]: I1207 16:03:29.530578 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:29 crc kubenswrapper[4716]: I1207 16:03:29.530589 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:29Z","lastTransitionTime":"2025-12-07T16:03:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:29 crc kubenswrapper[4716]: I1207 16:03:29.633114 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:29 crc kubenswrapper[4716]: I1207 16:03:29.633151 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:29 crc kubenswrapper[4716]: I1207 16:03:29.633170 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:29 crc kubenswrapper[4716]: I1207 16:03:29.633186 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:29 crc kubenswrapper[4716]: I1207 16:03:29.633197 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:29Z","lastTransitionTime":"2025-12-07T16:03:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:29 crc kubenswrapper[4716]: I1207 16:03:29.657472 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 16:03:29 crc kubenswrapper[4716]: E1207 16:03:29.657567 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 16:03:29 crc kubenswrapper[4716]: I1207 16:03:29.657604 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:03:29 crc kubenswrapper[4716]: I1207 16:03:29.657472 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nx2lh" Dec 07 16:03:29 crc kubenswrapper[4716]: E1207 16:03:29.657763 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 16:03:29 crc kubenswrapper[4716]: E1207 16:03:29.657903 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nx2lh" podUID="04b37f6c-457b-4f9a-8733-e88cb9644483" Dec 07 16:03:29 crc kubenswrapper[4716]: I1207 16:03:29.735093 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:29 crc kubenswrapper[4716]: I1207 16:03:29.735141 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:29 crc kubenswrapper[4716]: I1207 16:03:29.735151 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:29 crc kubenswrapper[4716]: I1207 16:03:29.735166 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:29 crc kubenswrapper[4716]: I1207 16:03:29.735202 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:29Z","lastTransitionTime":"2025-12-07T16:03:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:29 crc kubenswrapper[4716]: I1207 16:03:29.837701 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:29 crc kubenswrapper[4716]: I1207 16:03:29.837770 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:29 crc kubenswrapper[4716]: I1207 16:03:29.837798 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:29 crc kubenswrapper[4716]: I1207 16:03:29.837827 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:29 crc kubenswrapper[4716]: I1207 16:03:29.837848 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:29Z","lastTransitionTime":"2025-12-07T16:03:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:29 crc kubenswrapper[4716]: I1207 16:03:29.940897 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:29 crc kubenswrapper[4716]: I1207 16:03:29.940963 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:29 crc kubenswrapper[4716]: I1207 16:03:29.940985 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:29 crc kubenswrapper[4716]: I1207 16:03:29.941012 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:29 crc kubenswrapper[4716]: I1207 16:03:29.941034 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:29Z","lastTransitionTime":"2025-12-07T16:03:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:30 crc kubenswrapper[4716]: I1207 16:03:30.042730 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:30 crc kubenswrapper[4716]: I1207 16:03:30.042785 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:30 crc kubenswrapper[4716]: I1207 16:03:30.042802 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:30 crc kubenswrapper[4716]: I1207 16:03:30.042822 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:30 crc kubenswrapper[4716]: I1207 16:03:30.042839 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:30Z","lastTransitionTime":"2025-12-07T16:03:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:30 crc kubenswrapper[4716]: I1207 16:03:30.145490 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:30 crc kubenswrapper[4716]: I1207 16:03:30.145527 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:30 crc kubenswrapper[4716]: I1207 16:03:30.145537 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:30 crc kubenswrapper[4716]: I1207 16:03:30.145555 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:30 crc kubenswrapper[4716]: I1207 16:03:30.145565 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:30Z","lastTransitionTime":"2025-12-07T16:03:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:30 crc kubenswrapper[4716]: I1207 16:03:30.248474 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:30 crc kubenswrapper[4716]: I1207 16:03:30.248529 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:30 crc kubenswrapper[4716]: I1207 16:03:30.248545 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:30 crc kubenswrapper[4716]: I1207 16:03:30.248569 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:30 crc kubenswrapper[4716]: I1207 16:03:30.248587 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:30Z","lastTransitionTime":"2025-12-07T16:03:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:30 crc kubenswrapper[4716]: I1207 16:03:30.351389 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:30 crc kubenswrapper[4716]: I1207 16:03:30.351447 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:30 crc kubenswrapper[4716]: I1207 16:03:30.351457 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:30 crc kubenswrapper[4716]: I1207 16:03:30.351470 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:30 crc kubenswrapper[4716]: I1207 16:03:30.351479 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:30Z","lastTransitionTime":"2025-12-07T16:03:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:30 crc kubenswrapper[4716]: I1207 16:03:30.454558 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:30 crc kubenswrapper[4716]: I1207 16:03:30.454627 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:30 crc kubenswrapper[4716]: I1207 16:03:30.454645 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:30 crc kubenswrapper[4716]: I1207 16:03:30.454669 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:30 crc kubenswrapper[4716]: I1207 16:03:30.454686 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:30Z","lastTransitionTime":"2025-12-07T16:03:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:30 crc kubenswrapper[4716]: I1207 16:03:30.557179 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:30 crc kubenswrapper[4716]: I1207 16:03:30.557243 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:30 crc kubenswrapper[4716]: I1207 16:03:30.557260 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:30 crc kubenswrapper[4716]: I1207 16:03:30.557283 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:30 crc kubenswrapper[4716]: I1207 16:03:30.557324 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:30Z","lastTransitionTime":"2025-12-07T16:03:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:30 crc kubenswrapper[4716]: I1207 16:03:30.657517 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 16:03:30 crc kubenswrapper[4716]: E1207 16:03:30.657669 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 16:03:30 crc kubenswrapper[4716]: I1207 16:03:30.660167 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:30 crc kubenswrapper[4716]: I1207 16:03:30.660227 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:30 crc kubenswrapper[4716]: I1207 16:03:30.660244 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:30 crc kubenswrapper[4716]: I1207 16:03:30.660268 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:30 crc kubenswrapper[4716]: I1207 16:03:30.660286 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:30Z","lastTransitionTime":"2025-12-07T16:03:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:30 crc kubenswrapper[4716]: I1207 16:03:30.762954 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:30 crc kubenswrapper[4716]: I1207 16:03:30.763015 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:30 crc kubenswrapper[4716]: I1207 16:03:30.763032 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:30 crc kubenswrapper[4716]: I1207 16:03:30.763056 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:30 crc kubenswrapper[4716]: I1207 16:03:30.763072 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:30Z","lastTransitionTime":"2025-12-07T16:03:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:30 crc kubenswrapper[4716]: I1207 16:03:30.865950 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:30 crc kubenswrapper[4716]: I1207 16:03:30.865981 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:30 crc kubenswrapper[4716]: I1207 16:03:30.865989 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:30 crc kubenswrapper[4716]: I1207 16:03:30.866003 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:30 crc kubenswrapper[4716]: I1207 16:03:30.866011 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:30Z","lastTransitionTime":"2025-12-07T16:03:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:30 crc kubenswrapper[4716]: I1207 16:03:30.968652 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:30 crc kubenswrapper[4716]: I1207 16:03:30.968687 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:30 crc kubenswrapper[4716]: I1207 16:03:30.968698 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:30 crc kubenswrapper[4716]: I1207 16:03:30.968714 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:30 crc kubenswrapper[4716]: I1207 16:03:30.968725 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:30Z","lastTransitionTime":"2025-12-07T16:03:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:31 crc kubenswrapper[4716]: I1207 16:03:31.071567 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:31 crc kubenswrapper[4716]: I1207 16:03:31.071595 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:31 crc kubenswrapper[4716]: I1207 16:03:31.071605 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:31 crc kubenswrapper[4716]: I1207 16:03:31.071620 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:31 crc kubenswrapper[4716]: I1207 16:03:31.071635 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:31Z","lastTransitionTime":"2025-12-07T16:03:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:31 crc kubenswrapper[4716]: I1207 16:03:31.176712 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:31 crc kubenswrapper[4716]: I1207 16:03:31.176769 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:31 crc kubenswrapper[4716]: I1207 16:03:31.176787 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:31 crc kubenswrapper[4716]: I1207 16:03:31.176814 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:31 crc kubenswrapper[4716]: I1207 16:03:31.176832 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:31Z","lastTransitionTime":"2025-12-07T16:03:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:31 crc kubenswrapper[4716]: I1207 16:03:31.279826 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:31 crc kubenswrapper[4716]: I1207 16:03:31.279850 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:31 crc kubenswrapper[4716]: I1207 16:03:31.279857 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:31 crc kubenswrapper[4716]: I1207 16:03:31.279869 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:31 crc kubenswrapper[4716]: I1207 16:03:31.279876 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:31Z","lastTransitionTime":"2025-12-07T16:03:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:31 crc kubenswrapper[4716]: I1207 16:03:31.382656 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:31 crc kubenswrapper[4716]: I1207 16:03:31.382776 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:31 crc kubenswrapper[4716]: I1207 16:03:31.382806 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:31 crc kubenswrapper[4716]: I1207 16:03:31.382859 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:31 crc kubenswrapper[4716]: I1207 16:03:31.382889 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:31Z","lastTransitionTime":"2025-12-07T16:03:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:31 crc kubenswrapper[4716]: I1207 16:03:31.485539 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:31 crc kubenswrapper[4716]: I1207 16:03:31.485575 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:31 crc kubenswrapper[4716]: I1207 16:03:31.485586 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:31 crc kubenswrapper[4716]: I1207 16:03:31.485602 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:31 crc kubenswrapper[4716]: I1207 16:03:31.485614 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:31Z","lastTransitionTime":"2025-12-07T16:03:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:31 crc kubenswrapper[4716]: I1207 16:03:31.588172 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:31 crc kubenswrapper[4716]: I1207 16:03:31.588337 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:31 crc kubenswrapper[4716]: I1207 16:03:31.588412 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:31 crc kubenswrapper[4716]: I1207 16:03:31.588436 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:31 crc kubenswrapper[4716]: I1207 16:03:31.588449 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:31Z","lastTransitionTime":"2025-12-07T16:03:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:31 crc kubenswrapper[4716]: I1207 16:03:31.657597 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:03:31 crc kubenswrapper[4716]: I1207 16:03:31.657637 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 16:03:31 crc kubenswrapper[4716]: I1207 16:03:31.657667 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nx2lh" Dec 07 16:03:31 crc kubenswrapper[4716]: E1207 16:03:31.657827 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 16:03:31 crc kubenswrapper[4716]: E1207 16:03:31.657916 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 16:03:31 crc kubenswrapper[4716]: E1207 16:03:31.658673 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nx2lh" podUID="04b37f6c-457b-4f9a-8733-e88cb9644483" Dec 07 16:03:31 crc kubenswrapper[4716]: I1207 16:03:31.690703 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:31 crc kubenswrapper[4716]: I1207 16:03:31.690769 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:31 crc kubenswrapper[4716]: I1207 16:03:31.690788 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:31 crc kubenswrapper[4716]: I1207 16:03:31.690812 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:31 crc kubenswrapper[4716]: I1207 16:03:31.690830 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:31Z","lastTransitionTime":"2025-12-07T16:03:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:31 crc kubenswrapper[4716]: I1207 16:03:31.793316 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:31 crc kubenswrapper[4716]: I1207 16:03:31.793359 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:31 crc kubenswrapper[4716]: I1207 16:03:31.793370 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:31 crc kubenswrapper[4716]: I1207 16:03:31.793390 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:31 crc kubenswrapper[4716]: I1207 16:03:31.793402 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:31Z","lastTransitionTime":"2025-12-07T16:03:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:31 crc kubenswrapper[4716]: I1207 16:03:31.895931 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:31 crc kubenswrapper[4716]: I1207 16:03:31.896011 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:31 crc kubenswrapper[4716]: I1207 16:03:31.896035 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:31 crc kubenswrapper[4716]: I1207 16:03:31.896064 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:31 crc kubenswrapper[4716]: I1207 16:03:31.896128 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:31Z","lastTransitionTime":"2025-12-07T16:03:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:31 crc kubenswrapper[4716]: I1207 16:03:31.999336 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:31 crc kubenswrapper[4716]: I1207 16:03:31.999407 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:31 crc kubenswrapper[4716]: I1207 16:03:31.999428 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:31 crc kubenswrapper[4716]: I1207 16:03:31.999449 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:32 crc kubenswrapper[4716]: I1207 16:03:31.999466 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:31Z","lastTransitionTime":"2025-12-07T16:03:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:32 crc kubenswrapper[4716]: I1207 16:03:32.101480 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:32 crc kubenswrapper[4716]: I1207 16:03:32.101523 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:32 crc kubenswrapper[4716]: I1207 16:03:32.101535 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:32 crc kubenswrapper[4716]: I1207 16:03:32.101551 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:32 crc kubenswrapper[4716]: I1207 16:03:32.101561 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:32Z","lastTransitionTime":"2025-12-07T16:03:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:32 crc kubenswrapper[4716]: I1207 16:03:32.203472 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:32 crc kubenswrapper[4716]: I1207 16:03:32.203542 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:32 crc kubenswrapper[4716]: I1207 16:03:32.203565 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:32 crc kubenswrapper[4716]: I1207 16:03:32.203598 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:32 crc kubenswrapper[4716]: I1207 16:03:32.203618 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:32Z","lastTransitionTime":"2025-12-07T16:03:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:32 crc kubenswrapper[4716]: I1207 16:03:32.306211 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:32 crc kubenswrapper[4716]: I1207 16:03:32.306248 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:32 crc kubenswrapper[4716]: I1207 16:03:32.306256 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:32 crc kubenswrapper[4716]: I1207 16:03:32.306270 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:32 crc kubenswrapper[4716]: I1207 16:03:32.306280 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:32Z","lastTransitionTime":"2025-12-07T16:03:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:32 crc kubenswrapper[4716]: I1207 16:03:32.408284 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:32 crc kubenswrapper[4716]: I1207 16:03:32.408343 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:32 crc kubenswrapper[4716]: I1207 16:03:32.408359 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:32 crc kubenswrapper[4716]: I1207 16:03:32.408389 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:32 crc kubenswrapper[4716]: I1207 16:03:32.408409 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:32Z","lastTransitionTime":"2025-12-07T16:03:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:32 crc kubenswrapper[4716]: I1207 16:03:32.510181 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:32 crc kubenswrapper[4716]: I1207 16:03:32.510218 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:32 crc kubenswrapper[4716]: I1207 16:03:32.510227 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:32 crc kubenswrapper[4716]: I1207 16:03:32.510240 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:32 crc kubenswrapper[4716]: I1207 16:03:32.510249 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:32Z","lastTransitionTime":"2025-12-07T16:03:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:32 crc kubenswrapper[4716]: I1207 16:03:32.613427 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:32 crc kubenswrapper[4716]: I1207 16:03:32.613508 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:32 crc kubenswrapper[4716]: I1207 16:03:32.613532 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:32 crc kubenswrapper[4716]: I1207 16:03:32.613561 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:32 crc kubenswrapper[4716]: I1207 16:03:32.613584 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:32Z","lastTransitionTime":"2025-12-07T16:03:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:32 crc kubenswrapper[4716]: I1207 16:03:32.657532 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 16:03:32 crc kubenswrapper[4716]: E1207 16:03:32.657659 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 16:03:32 crc kubenswrapper[4716]: I1207 16:03:32.715886 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:32 crc kubenswrapper[4716]: I1207 16:03:32.715943 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:32 crc kubenswrapper[4716]: I1207 16:03:32.715959 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:32 crc kubenswrapper[4716]: I1207 16:03:32.715980 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:32 crc kubenswrapper[4716]: I1207 16:03:32.715993 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:32Z","lastTransitionTime":"2025-12-07T16:03:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:32 crc kubenswrapper[4716]: I1207 16:03:32.818123 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:32 crc kubenswrapper[4716]: I1207 16:03:32.818168 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:32 crc kubenswrapper[4716]: I1207 16:03:32.818180 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:32 crc kubenswrapper[4716]: I1207 16:03:32.818197 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:32 crc kubenswrapper[4716]: I1207 16:03:32.818210 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:32Z","lastTransitionTime":"2025-12-07T16:03:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:32 crc kubenswrapper[4716]: I1207 16:03:32.920407 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:32 crc kubenswrapper[4716]: I1207 16:03:32.920457 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:32 crc kubenswrapper[4716]: I1207 16:03:32.920470 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:32 crc kubenswrapper[4716]: I1207 16:03:32.920490 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:32 crc kubenswrapper[4716]: I1207 16:03:32.920504 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:32Z","lastTransitionTime":"2025-12-07T16:03:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:33 crc kubenswrapper[4716]: I1207 16:03:33.023576 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:33 crc kubenswrapper[4716]: I1207 16:03:33.023632 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:33 crc kubenswrapper[4716]: I1207 16:03:33.023651 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:33 crc kubenswrapper[4716]: I1207 16:03:33.023681 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:33 crc kubenswrapper[4716]: I1207 16:03:33.023703 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:33Z","lastTransitionTime":"2025-12-07T16:03:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:33 crc kubenswrapper[4716]: I1207 16:03:33.125833 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:33 crc kubenswrapper[4716]: I1207 16:03:33.125884 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:33 crc kubenswrapper[4716]: I1207 16:03:33.125901 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:33 crc kubenswrapper[4716]: I1207 16:03:33.125924 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:33 crc kubenswrapper[4716]: I1207 16:03:33.125935 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:33Z","lastTransitionTime":"2025-12-07T16:03:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:33 crc kubenswrapper[4716]: I1207 16:03:33.227790 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:33 crc kubenswrapper[4716]: I1207 16:03:33.227865 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:33 crc kubenswrapper[4716]: I1207 16:03:33.227885 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:33 crc kubenswrapper[4716]: I1207 16:03:33.227910 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:33 crc kubenswrapper[4716]: I1207 16:03:33.227926 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:33Z","lastTransitionTime":"2025-12-07T16:03:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:33 crc kubenswrapper[4716]: I1207 16:03:33.330963 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:33 crc kubenswrapper[4716]: I1207 16:03:33.331000 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:33 crc kubenswrapper[4716]: I1207 16:03:33.331011 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:33 crc kubenswrapper[4716]: I1207 16:03:33.331026 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:33 crc kubenswrapper[4716]: I1207 16:03:33.331037 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:33Z","lastTransitionTime":"2025-12-07T16:03:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:33 crc kubenswrapper[4716]: I1207 16:03:33.434013 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:33 crc kubenswrapper[4716]: I1207 16:03:33.434089 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:33 crc kubenswrapper[4716]: I1207 16:03:33.434102 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:33 crc kubenswrapper[4716]: I1207 16:03:33.434118 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:33 crc kubenswrapper[4716]: I1207 16:03:33.434128 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:33Z","lastTransitionTime":"2025-12-07T16:03:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:33 crc kubenswrapper[4716]: I1207 16:03:33.536554 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:33 crc kubenswrapper[4716]: I1207 16:03:33.536629 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:33 crc kubenswrapper[4716]: I1207 16:03:33.536648 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:33 crc kubenswrapper[4716]: I1207 16:03:33.536679 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:33 crc kubenswrapper[4716]: I1207 16:03:33.536697 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:33Z","lastTransitionTime":"2025-12-07T16:03:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:33 crc kubenswrapper[4716]: I1207 16:03:33.639985 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:33 crc kubenswrapper[4716]: I1207 16:03:33.640144 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:33 crc kubenswrapper[4716]: I1207 16:03:33.640163 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:33 crc kubenswrapper[4716]: I1207 16:03:33.640185 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:33 crc kubenswrapper[4716]: I1207 16:03:33.640201 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:33Z","lastTransitionTime":"2025-12-07T16:03:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:33 crc kubenswrapper[4716]: I1207 16:03:33.657758 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nx2lh" Dec 07 16:03:33 crc kubenswrapper[4716]: I1207 16:03:33.657863 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 16:03:33 crc kubenswrapper[4716]: I1207 16:03:33.657992 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:03:33 crc kubenswrapper[4716]: E1207 16:03:33.657980 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nx2lh" podUID="04b37f6c-457b-4f9a-8733-e88cb9644483" Dec 07 16:03:33 crc kubenswrapper[4716]: E1207 16:03:33.658218 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 16:03:33 crc kubenswrapper[4716]: E1207 16:03:33.658396 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 16:03:33 crc kubenswrapper[4716]: I1207 16:03:33.742618 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:33 crc kubenswrapper[4716]: I1207 16:03:33.742685 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:33 crc kubenswrapper[4716]: I1207 16:03:33.742705 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:33 crc kubenswrapper[4716]: I1207 16:03:33.742729 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:33 crc kubenswrapper[4716]: I1207 16:03:33.742746 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:33Z","lastTransitionTime":"2025-12-07T16:03:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:33 crc kubenswrapper[4716]: I1207 16:03:33.845685 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:33 crc kubenswrapper[4716]: I1207 16:03:33.845760 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:33 crc kubenswrapper[4716]: I1207 16:03:33.845784 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:33 crc kubenswrapper[4716]: I1207 16:03:33.845813 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:33 crc kubenswrapper[4716]: I1207 16:03:33.845836 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:33Z","lastTransitionTime":"2025-12-07T16:03:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:33 crc kubenswrapper[4716]: I1207 16:03:33.948811 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:33 crc kubenswrapper[4716]: I1207 16:03:33.948962 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:33 crc kubenswrapper[4716]: I1207 16:03:33.948983 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:33 crc kubenswrapper[4716]: I1207 16:03:33.949005 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:33 crc kubenswrapper[4716]: I1207 16:03:33.949021 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:33Z","lastTransitionTime":"2025-12-07T16:03:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:34 crc kubenswrapper[4716]: I1207 16:03:34.052100 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:34 crc kubenswrapper[4716]: I1207 16:03:34.052152 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:34 crc kubenswrapper[4716]: I1207 16:03:34.052169 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:34 crc kubenswrapper[4716]: I1207 16:03:34.052190 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:34 crc kubenswrapper[4716]: I1207 16:03:34.052205 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:34Z","lastTransitionTime":"2025-12-07T16:03:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:34 crc kubenswrapper[4716]: I1207 16:03:34.155177 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:34 crc kubenswrapper[4716]: I1207 16:03:34.155220 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:34 crc kubenswrapper[4716]: I1207 16:03:34.155229 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:34 crc kubenswrapper[4716]: I1207 16:03:34.155244 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:34 crc kubenswrapper[4716]: I1207 16:03:34.155259 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:34Z","lastTransitionTime":"2025-12-07T16:03:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:34 crc kubenswrapper[4716]: I1207 16:03:34.258835 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:34 crc kubenswrapper[4716]: I1207 16:03:34.258945 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:34 crc kubenswrapper[4716]: I1207 16:03:34.258971 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:34 crc kubenswrapper[4716]: I1207 16:03:34.259005 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:34 crc kubenswrapper[4716]: I1207 16:03:34.259039 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:34Z","lastTransitionTime":"2025-12-07T16:03:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:34 crc kubenswrapper[4716]: I1207 16:03:34.361885 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:34 crc kubenswrapper[4716]: I1207 16:03:34.361921 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:34 crc kubenswrapper[4716]: I1207 16:03:34.361932 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:34 crc kubenswrapper[4716]: I1207 16:03:34.361948 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:34 crc kubenswrapper[4716]: I1207 16:03:34.361958 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:34Z","lastTransitionTime":"2025-12-07T16:03:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:34 crc kubenswrapper[4716]: I1207 16:03:34.465286 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:34 crc kubenswrapper[4716]: I1207 16:03:34.465361 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:34 crc kubenswrapper[4716]: I1207 16:03:34.465383 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:34 crc kubenswrapper[4716]: I1207 16:03:34.465412 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:34 crc kubenswrapper[4716]: I1207 16:03:34.465434 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:34Z","lastTransitionTime":"2025-12-07T16:03:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:34 crc kubenswrapper[4716]: I1207 16:03:34.567187 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:34 crc kubenswrapper[4716]: I1207 16:03:34.567239 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:34 crc kubenswrapper[4716]: I1207 16:03:34.567265 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:34 crc kubenswrapper[4716]: I1207 16:03:34.567296 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:34 crc kubenswrapper[4716]: I1207 16:03:34.567310 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:34Z","lastTransitionTime":"2025-12-07T16:03:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:34 crc kubenswrapper[4716]: I1207 16:03:34.656653 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 16:03:34 crc kubenswrapper[4716]: E1207 16:03:34.656842 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 16:03:34 crc kubenswrapper[4716]: I1207 16:03:34.666962 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Dec 07 16:03:34 crc kubenswrapper[4716]: I1207 16:03:34.672691 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:34 crc kubenswrapper[4716]: I1207 16:03:34.672723 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:34 crc kubenswrapper[4716]: I1207 16:03:34.672730 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:34 crc kubenswrapper[4716]: I1207 16:03:34.672744 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:34 crc kubenswrapper[4716]: I1207 16:03:34.672755 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:34Z","lastTransitionTime":"2025-12-07T16:03:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:34 crc kubenswrapper[4716]: I1207 16:03:34.776157 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:34 crc kubenswrapper[4716]: I1207 16:03:34.776189 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:34 crc kubenswrapper[4716]: I1207 16:03:34.776201 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:34 crc kubenswrapper[4716]: I1207 16:03:34.776217 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:34 crc kubenswrapper[4716]: I1207 16:03:34.776228 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:34Z","lastTransitionTime":"2025-12-07T16:03:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:34 crc kubenswrapper[4716]: I1207 16:03:34.879266 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:34 crc kubenswrapper[4716]: I1207 16:03:34.879311 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:34 crc kubenswrapper[4716]: I1207 16:03:34.879326 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:34 crc kubenswrapper[4716]: I1207 16:03:34.879345 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:34 crc kubenswrapper[4716]: I1207 16:03:34.879359 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:34Z","lastTransitionTime":"2025-12-07T16:03:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:34 crc kubenswrapper[4716]: I1207 16:03:34.981639 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:34 crc kubenswrapper[4716]: I1207 16:03:34.981676 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:34 crc kubenswrapper[4716]: I1207 16:03:34.981693 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:34 crc kubenswrapper[4716]: I1207 16:03:34.981709 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:34 crc kubenswrapper[4716]: I1207 16:03:34.981719 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:34Z","lastTransitionTime":"2025-12-07T16:03:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:35 crc kubenswrapper[4716]: I1207 16:03:35.084571 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:35 crc kubenswrapper[4716]: I1207 16:03:35.084651 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:35 crc kubenswrapper[4716]: I1207 16:03:35.084671 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:35 crc kubenswrapper[4716]: I1207 16:03:35.084699 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:35 crc kubenswrapper[4716]: I1207 16:03:35.084719 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:35Z","lastTransitionTime":"2025-12-07T16:03:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:35 crc kubenswrapper[4716]: I1207 16:03:35.187520 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:35 crc kubenswrapper[4716]: I1207 16:03:35.187575 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:35 crc kubenswrapper[4716]: I1207 16:03:35.187589 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:35 crc kubenswrapper[4716]: I1207 16:03:35.187606 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:35 crc kubenswrapper[4716]: I1207 16:03:35.187618 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:35Z","lastTransitionTime":"2025-12-07T16:03:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:35 crc kubenswrapper[4716]: I1207 16:03:35.310302 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:35 crc kubenswrapper[4716]: I1207 16:03:35.310358 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:35 crc kubenswrapper[4716]: I1207 16:03:35.310374 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:35 crc kubenswrapper[4716]: I1207 16:03:35.310400 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:35 crc kubenswrapper[4716]: I1207 16:03:35.310417 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:35Z","lastTransitionTime":"2025-12-07T16:03:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:35 crc kubenswrapper[4716]: I1207 16:03:35.412891 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:35 crc kubenswrapper[4716]: I1207 16:03:35.412938 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:35 crc kubenswrapper[4716]: I1207 16:03:35.412950 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:35 crc kubenswrapper[4716]: I1207 16:03:35.412968 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:35 crc kubenswrapper[4716]: I1207 16:03:35.412980 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:35Z","lastTransitionTime":"2025-12-07T16:03:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:35 crc kubenswrapper[4716]: I1207 16:03:35.515024 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:35 crc kubenswrapper[4716]: I1207 16:03:35.515066 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:35 crc kubenswrapper[4716]: I1207 16:03:35.515116 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:35 crc kubenswrapper[4716]: I1207 16:03:35.515139 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:35 crc kubenswrapper[4716]: I1207 16:03:35.515156 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:35Z","lastTransitionTime":"2025-12-07T16:03:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:35 crc kubenswrapper[4716]: I1207 16:03:35.617491 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:35 crc kubenswrapper[4716]: I1207 16:03:35.617520 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:35 crc kubenswrapper[4716]: I1207 16:03:35.617529 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:35 crc kubenswrapper[4716]: I1207 16:03:35.617544 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:35 crc kubenswrapper[4716]: I1207 16:03:35.617552 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:35Z","lastTransitionTime":"2025-12-07T16:03:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:35 crc kubenswrapper[4716]: I1207 16:03:35.657357 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nx2lh" Dec 07 16:03:35 crc kubenswrapper[4716]: I1207 16:03:35.657355 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 16:03:35 crc kubenswrapper[4716]: I1207 16:03:35.657435 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:03:35 crc kubenswrapper[4716]: E1207 16:03:35.657599 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nx2lh" podUID="04b37f6c-457b-4f9a-8733-e88cb9644483" Dec 07 16:03:35 crc kubenswrapper[4716]: E1207 16:03:35.657715 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 16:03:35 crc kubenswrapper[4716]: E1207 16:03:35.657820 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 16:03:35 crc kubenswrapper[4716]: I1207 16:03:35.720616 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:35 crc kubenswrapper[4716]: I1207 16:03:35.720657 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:35 crc kubenswrapper[4716]: I1207 16:03:35.720670 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:35 crc kubenswrapper[4716]: I1207 16:03:35.720686 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:35 crc kubenswrapper[4716]: I1207 16:03:35.720698 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:35Z","lastTransitionTime":"2025-12-07T16:03:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:35 crc kubenswrapper[4716]: I1207 16:03:35.823480 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:35 crc kubenswrapper[4716]: I1207 16:03:35.823505 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:35 crc kubenswrapper[4716]: I1207 16:03:35.823513 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:35 crc kubenswrapper[4716]: I1207 16:03:35.823526 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:35 crc kubenswrapper[4716]: I1207 16:03:35.823535 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:35Z","lastTransitionTime":"2025-12-07T16:03:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:35 crc kubenswrapper[4716]: I1207 16:03:35.925797 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:35 crc kubenswrapper[4716]: I1207 16:03:35.925840 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:35 crc kubenswrapper[4716]: I1207 16:03:35.925856 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:35 crc kubenswrapper[4716]: I1207 16:03:35.925877 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:35 crc kubenswrapper[4716]: I1207 16:03:35.925893 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:35Z","lastTransitionTime":"2025-12-07T16:03:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:36 crc kubenswrapper[4716]: I1207 16:03:36.027738 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:36 crc kubenswrapper[4716]: I1207 16:03:36.027775 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:36 crc kubenswrapper[4716]: I1207 16:03:36.027784 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:36 crc kubenswrapper[4716]: I1207 16:03:36.027798 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:36 crc kubenswrapper[4716]: I1207 16:03:36.027807 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:36Z","lastTransitionTime":"2025-12-07T16:03:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:36 crc kubenswrapper[4716]: I1207 16:03:36.129747 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:36 crc kubenswrapper[4716]: I1207 16:03:36.129819 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:36 crc kubenswrapper[4716]: I1207 16:03:36.129843 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:36 crc kubenswrapper[4716]: I1207 16:03:36.129873 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:36 crc kubenswrapper[4716]: I1207 16:03:36.129895 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:36Z","lastTransitionTime":"2025-12-07T16:03:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:36 crc kubenswrapper[4716]: I1207 16:03:36.231960 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:36 crc kubenswrapper[4716]: I1207 16:03:36.231996 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:36 crc kubenswrapper[4716]: I1207 16:03:36.232007 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:36 crc kubenswrapper[4716]: I1207 16:03:36.232023 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:36 crc kubenswrapper[4716]: I1207 16:03:36.232034 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:36Z","lastTransitionTime":"2025-12-07T16:03:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:36 crc kubenswrapper[4716]: I1207 16:03:36.334686 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:36 crc kubenswrapper[4716]: I1207 16:03:36.334743 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:36 crc kubenswrapper[4716]: I1207 16:03:36.334761 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:36 crc kubenswrapper[4716]: I1207 16:03:36.334786 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:36 crc kubenswrapper[4716]: I1207 16:03:36.334807 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:36Z","lastTransitionTime":"2025-12-07T16:03:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:36 crc kubenswrapper[4716]: I1207 16:03:36.437245 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:36 crc kubenswrapper[4716]: I1207 16:03:36.437321 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:36 crc kubenswrapper[4716]: I1207 16:03:36.437340 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:36 crc kubenswrapper[4716]: I1207 16:03:36.437365 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:36 crc kubenswrapper[4716]: I1207 16:03:36.437386 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:36Z","lastTransitionTime":"2025-12-07T16:03:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:36 crc kubenswrapper[4716]: I1207 16:03:36.540271 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:36 crc kubenswrapper[4716]: I1207 16:03:36.540341 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:36 crc kubenswrapper[4716]: I1207 16:03:36.540380 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:36 crc kubenswrapper[4716]: I1207 16:03:36.540414 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:36 crc kubenswrapper[4716]: I1207 16:03:36.540437 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:36Z","lastTransitionTime":"2025-12-07T16:03:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:36 crc kubenswrapper[4716]: I1207 16:03:36.643135 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:36 crc kubenswrapper[4716]: I1207 16:03:36.643201 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:36 crc kubenswrapper[4716]: I1207 16:03:36.643225 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:36 crc kubenswrapper[4716]: I1207 16:03:36.643256 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:36 crc kubenswrapper[4716]: I1207 16:03:36.643279 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:36Z","lastTransitionTime":"2025-12-07T16:03:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:36 crc kubenswrapper[4716]: I1207 16:03:36.657483 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 16:03:36 crc kubenswrapper[4716]: E1207 16:03:36.657747 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 16:03:36 crc kubenswrapper[4716]: I1207 16:03:36.744935 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:36 crc kubenswrapper[4716]: I1207 16:03:36.744974 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:36 crc kubenswrapper[4716]: I1207 16:03:36.744986 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:36 crc kubenswrapper[4716]: I1207 16:03:36.745003 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:36 crc kubenswrapper[4716]: I1207 16:03:36.745014 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:36Z","lastTransitionTime":"2025-12-07T16:03:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:36 crc kubenswrapper[4716]: I1207 16:03:36.847508 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:36 crc kubenswrapper[4716]: I1207 16:03:36.847544 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:36 crc kubenswrapper[4716]: I1207 16:03:36.847552 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:36 crc kubenswrapper[4716]: I1207 16:03:36.847566 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:36 crc kubenswrapper[4716]: I1207 16:03:36.847575 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:36Z","lastTransitionTime":"2025-12-07T16:03:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:36 crc kubenswrapper[4716]: I1207 16:03:36.949978 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:36 crc kubenswrapper[4716]: I1207 16:03:36.950036 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:36 crc kubenswrapper[4716]: I1207 16:03:36.950053 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:36 crc kubenswrapper[4716]: I1207 16:03:36.950112 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:36 crc kubenswrapper[4716]: I1207 16:03:36.950131 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:36Z","lastTransitionTime":"2025-12-07T16:03:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.052868 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.052912 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.052922 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.052935 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.052944 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:37Z","lastTransitionTime":"2025-12-07T16:03:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.154708 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.154779 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.154803 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.154831 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.154849 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:37Z","lastTransitionTime":"2025-12-07T16:03:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.257656 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.257724 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.257742 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.257766 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.257783 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:37Z","lastTransitionTime":"2025-12-07T16:03:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.360975 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.361042 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.361059 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.361105 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.361122 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:37Z","lastTransitionTime":"2025-12-07T16:03:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.464526 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.464592 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.464610 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.464634 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.464655 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:37Z","lastTransitionTime":"2025-12-07T16:03:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.492707 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.492761 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.492777 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.492801 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.492818 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:37Z","lastTransitionTime":"2025-12-07T16:03:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:37 crc kubenswrapper[4716]: E1207 16:03:37.504728 4716 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a8d1b7d4-7ec4-40c7-86b0-291db1f38894\\\",\\\"systemUUID\\\":\\\"a57d09b0-0224-46eb-a3ac-a2b60dad2250\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:37Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.509040 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.509101 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.509113 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.509136 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.509147 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:37Z","lastTransitionTime":"2025-12-07T16:03:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:37 crc kubenswrapper[4716]: E1207 16:03:37.528760 4716 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a8d1b7d4-7ec4-40c7-86b0-291db1f38894\\\",\\\"systemUUID\\\":\\\"a57d09b0-0224-46eb-a3ac-a2b60dad2250\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:37Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.532663 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.532722 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.532740 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.532765 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.532782 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:37Z","lastTransitionTime":"2025-12-07T16:03:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:37 crc kubenswrapper[4716]: E1207 16:03:37.545457 4716 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a8d1b7d4-7ec4-40c7-86b0-291db1f38894\\\",\\\"systemUUID\\\":\\\"a57d09b0-0224-46eb-a3ac-a2b60dad2250\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:37Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.549359 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.549405 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.549418 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.549435 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.549446 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:37Z","lastTransitionTime":"2025-12-07T16:03:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:37 crc kubenswrapper[4716]: E1207 16:03:37.561947 4716 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a8d1b7d4-7ec4-40c7-86b0-291db1f38894\\\",\\\"systemUUID\\\":\\\"a57d09b0-0224-46eb-a3ac-a2b60dad2250\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:37Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.566491 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.566563 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.566585 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.566613 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.566635 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:37Z","lastTransitionTime":"2025-12-07T16:03:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:37 crc kubenswrapper[4716]: E1207 16:03:37.580164 4716 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:03:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a8d1b7d4-7ec4-40c7-86b0-291db1f38894\\\",\\\"systemUUID\\\":\\\"a57d09b0-0224-46eb-a3ac-a2b60dad2250\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:37Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:37 crc kubenswrapper[4716]: E1207 16:03:37.580380 4716 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.582198 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.582257 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.582280 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.582309 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.582341 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:37Z","lastTransitionTime":"2025-12-07T16:03:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.657552 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.657632 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 16:03:37 crc kubenswrapper[4716]: E1207 16:03:37.657762 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.657783 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nx2lh" Dec 07 16:03:37 crc kubenswrapper[4716]: E1207 16:03:37.657929 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 16:03:37 crc kubenswrapper[4716]: E1207 16:03:37.657970 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nx2lh" podUID="04b37f6c-457b-4f9a-8733-e88cb9644483" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.658640 4716 scope.go:117] "RemoveContainer" containerID="f48167b38ff840f85f11b69d8090ec6c9c634ed26682daef93f5da99fc02ef61" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.677411 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7961cc9a-734e-42c2-b1ff-7075f3f20091\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84a36b76f20618c7885d611ecad9db29bac5101c27572be31d3acbb90c581291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0a0eca3c2087a457c0e448e649345ec4595a9c8a019aed2cddc5f30c77871e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://404426a171b8676eddd4a00996c76afab7e269643fa9608f1e1b6bdba385071d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fd3c2941782921f99a6cdc3de4e892234548bde527fec15a8ac92e513566e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e23abfdca314ed43d594333c9f9e401cc8ef7dad39b8ebac84a836b06ee5c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09b38cd4b389ea477c802a842ee4f2297a64d64684624117163e819539c257c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09b38cd4b389ea477c802a842ee4f2297a64d64684624117163e819539c257c6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c00f7a8e024e0e0dfd7b73e16d1d5437602e79ea95b73c21f24b209edef60e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c00f7a8e024e0e0dfd7b73e16d1d5437602e79ea95b73c21f24b209edef60e98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3f6392d426fa2cd788ec3a5c3bff4ac39483a54dfb2624a1c4370522413a350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3f6392d426fa2cd788ec3a5c3bff4ac39483a54dfb2624a1c4370522413a350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:37Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.685930 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.685966 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.685975 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.685988 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.685999 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:37Z","lastTransitionTime":"2025-12-07T16:03:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.695399 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3067de7c-c11c-4be1-bd94-d6572c266795\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d14d018fa044e9e80fe81e7e38df935992d14c773ed2b2193fd5cb873e5d0947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a9e9cf048220ca14c996847e9ced9f2367b9331dc6c2836dd023658855cd79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6d3c911874095b3c7cb062d03070e5090a7bb2456cb20cb4ce078f502515cda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ac6ce7701f1c6aa1ea8fcf7a3035e3b0eaa28ae3cb6afc442372bd9117bdde4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:37Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.708478 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4742ea84b3a2c9461d9b900beee7354a2ded24c39ad23e007cbf270ae12377c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:37Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.729609 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:37Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.747340 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c15b59eb-565d-4556-a4ce-75afdf159dc8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://530a2935826940767ea33fb4f37fa11f7dee4950d867fbcd6af72836d80f7db3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jks8z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ccaf6c1e6a978eb978a1c207b6c4c00b23737f78b743f966579aca781e22e0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jks8z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zcxxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:37Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.763341 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8ps9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98015cea-d753-4e7a-b2b5-5ae7733a81e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d68509b673eecd37725cdcc48e3d0612c3b9c7601433f2b72c21d662140daa59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f84f52a073437543721010dc440339654ffb5e04eeb6cc7afbcc62f00f64c8e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f84f52a073437543721010dc440339654ffb5e04eeb6cc7afbcc62f00f64c8e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f6fe1cb25218cb3ac20488dc47efa4dc6a26a43dbc8306b5ee38cbe1a186c67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f6fe1cb25218cb3ac20488dc47efa4dc6a26a43dbc8306b5ee38cbe1a186c67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14b748a8247f6863e8be61e90057a06626dfd3cd4f01bd3a11629f700f56378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f14b748a8247f6863e8be61e90057a06626dfd3cd4f01bd3a11629f700f56378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ec379c0d573cb9b5ffcfc5ca7c653ba60e2fbdd7e3d37836102e26618ce43e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ec379c0d573cb9b5ffcfc5ca7c653ba60e2fbdd7e3d37836102e26618ce43e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c75bf2a6703cf99865d4ed5fa000e1e3b1330d4946176cf585fec1a2f226df0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c75bf2a6703cf99865d4ed5fa000e1e3b1330d4946176cf585fec1a2f226df0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de45812c9a5a5bfe6b0146828d03fab28362b03d04ef861c150b9ffbf181d345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de45812c9a5a5bfe6b0146828d03fab28362b03d04ef861c150b9ffbf181d345\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8ps9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:37Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.775388 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e903f3c-ba8c-477d-b66a-6d1331a7c64e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ff9c456318bbdb54ec3f1ac3e2c2366968f18ad084158b7100d142acb668948\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1358745250b35426602fc7417e4f85dd411cdd0fb3e8f703c432d72128ca7c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d8c9a434c6cfb9e4fee6a8955f1379291f153c007876d1ffb498ce561c73ba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad8df9b5bd5f5a8dd44ce2899cfa29c19435149b170d339d339183cd3df09e99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad8df9b5bd5f5a8dd44ce2899cfa29c19435149b170d339d339183cd3df09e99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:37Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.788639 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.788696 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.788715 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.788738 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.788754 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:37Z","lastTransitionTime":"2025-12-07T16:03:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.798523 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c54d618a-dfec-4f2a-a288-f5188c8a128a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60324dc7d701bc064e14ae4f474bce10f6ee4ff3c3c942eab749e7166082cf9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ed4f4d3564674bf5cb4e7a76dc0fd95e203ecc8950798e59a1fb6c9ab4a20f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://def9ac840fb8ecf2778bed88ea980eddca7c5b2aba4bd1afb967d62fec3f7bae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b73be664f5412c98b89d971a2cac37c494151c293fdb44a98b390978658a6643\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a847d04e22e98091c0037a89054b9236228b701b8c7ba31940a56df0f37b95c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcae78a6133d97f6a32dd7421ccec4333b970a237404e056a5cc601f8eefdc8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f48167b38ff840f85f11b69d8090ec6c9c634ed26682daef93f5da99fc02ef61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f48167b38ff840f85f11b69d8090ec6c9c634ed26682daef93f5da99fc02ef61\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-07T16:03:08Z\\\",\\\"message\\\":\\\"er, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:08Z is after 2025-08-24T17:21:41Z]\\\\nI1207 16:03:08.453903 6420 services_controller.go:451] Built service openshift-controller-manager-operator/metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-controller-manager-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-controller-manager-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.58\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.Tem\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T16:03:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qvc7p_openshift-ovn-kubernetes(c54d618a-dfec-4f2a-a288-f5188c8a128a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcfdffe822635b0ae13b253e6133449fe00461b2792be43d913c4a77219b8255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9bfe2ffce7a7b7c320474117734fc05246cc2b4fd744833e896e8e7c83f5e4d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9bfe2ffce7a7b7c320474117734fc05246cc2b4fd744833e896e8e7c83f5e4d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qvc7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:37Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.810012 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mx85f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"118dd9ba-437c-4e5c-b5ff-4239ae6d4a63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bcd6c8deb184d0c713c5e7b7cc82fa38d3cd498517911ba7a0ec3cb88a3437a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gg5nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ffc069d2b21fa80871b1b97f2c72c37880d9a309cbf8b7daf7d240f2c332a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gg5nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mx85f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:37Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.875332 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nx2lh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04b37f6c-457b-4f9a-8733-e88cb9644483\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv6td\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv6td\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nx2lh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:37Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.888293 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdee9957-48b9-458f-ba71-e0cd39f80596\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73f85248b4d68a9beecdacb3a01f433e97a0ffea2247ce92a41ef96855509377\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6276df652efe6ca935e745f24ef20f87a16465475139230c63b9999234a4d12a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3b1dd202586ee755ca89fbe12b30bbb12f95a62514f6e4d93cbbdbe55181e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd318647816e4a27849666b0ec4fc2ca4905295661c557dc381e1144f22cf284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d31de0326a05216fa10442bdb47151c6e073c97e9a16c34e793028e5e253756\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"le observer\\\\nW1207 16:02:34.645464 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1207 16:02:34.645587 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 16:02:34.646501 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2082283479/tls.crt::/tmp/serving-cert-2082283479/tls.key\\\\\\\"\\\\nI1207 16:02:34.972041 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 16:02:34.973702 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 16:02:34.973717 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 16:02:34.973738 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 16:02:34.973743 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 16:02:34.977185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1207 16:02:34.977204 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1207 16:02:34.977209 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977217 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 16:02:34.977225 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 16:02:34.977228 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 16:02:34.977232 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1207 16:02:34.979033 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27bd4a54f920b2725987c1c7fd2d9fdf16cd08738c246c333a4e9164a3a582a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:37Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.891741 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.891771 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.891779 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.891793 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.891802 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:37Z","lastTransitionTime":"2025-12-07T16:03:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.899931 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:37Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.907732 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2jjld" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dee6f5c-861c-4819-9309-7791574c7b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51c1d4159ab11087839a2c2d192566fb9322447c26df560bf09517c906ececb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmdns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2jjld\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:37Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.918831 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4wr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0baf8575-26cf-4a72-8a6d-ae94c078eaad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fca3b2ea503451cfaf7afb00445fe0adb0acc87d436d1eebba5068452b8e7cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0ca2b3cd5a11033afd219b0125be7ab5acd750935d7eb89a470eff3d93a0838\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-07T16:03:26Z\\\",\\\"message\\\":\\\"2025-12-07T16:02:41+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_09397239-db79-454b-b301-47a1e4cdfc48\\\\n2025-12-07T16:02:41+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_09397239-db79-454b-b301-47a1e4cdfc48 to /host/opt/cni/bin/\\\\n2025-12-07T16:02:41Z [verbose] multus-daemon started\\\\n2025-12-07T16:02:41Z [verbose] Readiness Indicator file check\\\\n2025-12-07T16:03:26Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:03:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p2trd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4wr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:37Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.926982 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-m5jl5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"824ee59f-1474-4994-a70e-462878521514\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a15fe50cc75515704779646cc19aa50004f0db9aa26fe2535110df749bbfb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhwzn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-m5jl5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:37Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.935115 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71d1a4e4-5c15-4aa2-b24c-75c1f7729297\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1ab9067e4d9988fd94f93d7798954e4c8888875f47244d797cb6b0557ce5582\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3d77f99006bada2a3a676398c5ddb366c0c727f3837f10674a9d582dd83e2cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3d77f99006bada2a3a676398c5ddb366c0c727f3837f10674a9d582dd83e2cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:37Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.945037 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef3335ae3875a75ab64685fc7e3590177280d651e8f8292c6f5ea6f472e55789\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:37Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.956703 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:37Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.968205 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f95ef61779507288a03bda8eddf745e105f705998a117a30789675bfa452d88f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b955b23163e7c8c458a3bbe2f681633e6a663ce76748febcc0b3a3c37f3ca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:37Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.993510 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.993539 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.993548 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.993560 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:37 crc kubenswrapper[4716]: I1207 16:03:37.993569 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:37Z","lastTransitionTime":"2025-12-07T16:03:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:38 crc kubenswrapper[4716]: I1207 16:03:38.095552 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:38 crc kubenswrapper[4716]: I1207 16:03:38.095594 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:38 crc kubenswrapper[4716]: I1207 16:03:38.095606 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:38 crc kubenswrapper[4716]: I1207 16:03:38.095622 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:38 crc kubenswrapper[4716]: I1207 16:03:38.095633 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:38Z","lastTransitionTime":"2025-12-07T16:03:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:38 crc kubenswrapper[4716]: I1207 16:03:38.197291 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:38 crc kubenswrapper[4716]: I1207 16:03:38.197331 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:38 crc kubenswrapper[4716]: I1207 16:03:38.197340 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:38 crc kubenswrapper[4716]: I1207 16:03:38.197357 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:38 crc kubenswrapper[4716]: I1207 16:03:38.197367 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:38Z","lastTransitionTime":"2025-12-07T16:03:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:38 crc kubenswrapper[4716]: I1207 16:03:38.299964 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:38 crc kubenswrapper[4716]: I1207 16:03:38.300004 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:38 crc kubenswrapper[4716]: I1207 16:03:38.300016 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:38 crc kubenswrapper[4716]: I1207 16:03:38.300034 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:38 crc kubenswrapper[4716]: I1207 16:03:38.300044 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:38Z","lastTransitionTime":"2025-12-07T16:03:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:38 crc kubenswrapper[4716]: I1207 16:03:38.402513 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:38 crc kubenswrapper[4716]: I1207 16:03:38.402545 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:38 crc kubenswrapper[4716]: I1207 16:03:38.402555 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:38 crc kubenswrapper[4716]: I1207 16:03:38.402568 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:38 crc kubenswrapper[4716]: I1207 16:03:38.402577 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:38Z","lastTransitionTime":"2025-12-07T16:03:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:38 crc kubenswrapper[4716]: I1207 16:03:38.505073 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:38 crc kubenswrapper[4716]: I1207 16:03:38.505157 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:38 crc kubenswrapper[4716]: I1207 16:03:38.505199 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:38 crc kubenswrapper[4716]: I1207 16:03:38.505230 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:38 crc kubenswrapper[4716]: I1207 16:03:38.505251 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:38Z","lastTransitionTime":"2025-12-07T16:03:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:38 crc kubenswrapper[4716]: I1207 16:03:38.607348 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:38 crc kubenswrapper[4716]: I1207 16:03:38.607383 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:38 crc kubenswrapper[4716]: I1207 16:03:38.607392 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:38 crc kubenswrapper[4716]: I1207 16:03:38.607405 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:38 crc kubenswrapper[4716]: I1207 16:03:38.607415 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:38Z","lastTransitionTime":"2025-12-07T16:03:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:38 crc kubenswrapper[4716]: I1207 16:03:38.657399 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 16:03:38 crc kubenswrapper[4716]: E1207 16:03:38.657546 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 16:03:38 crc kubenswrapper[4716]: I1207 16:03:38.710233 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:38 crc kubenswrapper[4716]: I1207 16:03:38.710301 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:38 crc kubenswrapper[4716]: I1207 16:03:38.710317 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:38 crc kubenswrapper[4716]: I1207 16:03:38.710331 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:38 crc kubenswrapper[4716]: I1207 16:03:38.710344 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:38Z","lastTransitionTime":"2025-12-07T16:03:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:38 crc kubenswrapper[4716]: I1207 16:03:38.812998 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:38 crc kubenswrapper[4716]: I1207 16:03:38.813063 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:38 crc kubenswrapper[4716]: I1207 16:03:38.813118 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:38 crc kubenswrapper[4716]: I1207 16:03:38.813151 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:38 crc kubenswrapper[4716]: I1207 16:03:38.813176 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:38Z","lastTransitionTime":"2025-12-07T16:03:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:38 crc kubenswrapper[4716]: I1207 16:03:38.932636 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:38 crc kubenswrapper[4716]: I1207 16:03:38.932677 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:38 crc kubenswrapper[4716]: I1207 16:03:38.932687 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:38 crc kubenswrapper[4716]: I1207 16:03:38.932700 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:38 crc kubenswrapper[4716]: I1207 16:03:38.932712 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:38Z","lastTransitionTime":"2025-12-07T16:03:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.035214 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.035493 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.035507 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.035524 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.035534 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:39Z","lastTransitionTime":"2025-12-07T16:03:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.120413 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qvc7p_c54d618a-dfec-4f2a-a288-f5188c8a128a/ovnkube-controller/2.log" Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.122816 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" event={"ID":"c54d618a-dfec-4f2a-a288-f5188c8a128a","Type":"ContainerStarted","Data":"c60794c12cb6e7647d4c1d0bee76e214a71d9cb818d9c98dbf9880fb030ed54b"} Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.123282 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.137751 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.137800 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.137810 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.137824 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.137832 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:39Z","lastTransitionTime":"2025-12-07T16:03:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.140650 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71d1a4e4-5c15-4aa2-b24c-75c1f7729297\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1ab9067e4d9988fd94f93d7798954e4c8888875f47244d797cb6b0557ce5582\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3d77f99006bada2a3a676398c5ddb366c0c727f3837f10674a9d582dd83e2cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3d77f99006bada2a3a676398c5ddb366c0c727f3837f10674a9d582dd83e2cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:39Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.155259 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef3335ae3875a75ab64685fc7e3590177280d651e8f8292c6f5ea6f472e55789\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:39Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.169311 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:39Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.184280 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f95ef61779507288a03bda8eddf745e105f705998a117a30789675bfa452d88f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b955b23163e7c8c458a3bbe2f681633e6a663ce76748febcc0b3a3c37f3ca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:39Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.202393 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7961cc9a-734e-42c2-b1ff-7075f3f20091\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84a36b76f20618c7885d611ecad9db29bac5101c27572be31d3acbb90c581291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0a0eca3c2087a457c0e448e649345ec4595a9c8a019aed2cddc5f30c77871e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://404426a171b8676eddd4a00996c76afab7e269643fa9608f1e1b6bdba385071d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fd3c2941782921f99a6cdc3de4e892234548bde527fec15a8ac92e513566e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e23abfdca314ed43d594333c9f9e401cc8ef7dad39b8ebac84a836b06ee5c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09b38cd4b389ea477c802a842ee4f2297a64d64684624117163e819539c257c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09b38cd4b389ea477c802a842ee4f2297a64d64684624117163e819539c257c6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c00f7a8e024e0e0dfd7b73e16d1d5437602e79ea95b73c21f24b209edef60e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c00f7a8e024e0e0dfd7b73e16d1d5437602e79ea95b73c21f24b209edef60e98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3f6392d426fa2cd788ec3a5c3bff4ac39483a54dfb2624a1c4370522413a350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3f6392d426fa2cd788ec3a5c3bff4ac39483a54dfb2624a1c4370522413a350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:39Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.215097 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3067de7c-c11c-4be1-bd94-d6572c266795\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d14d018fa044e9e80fe81e7e38df935992d14c773ed2b2193fd5cb873e5d0947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a9e9cf048220ca14c996847e9ced9f2367b9331dc6c2836dd023658855cd79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6d3c911874095b3c7cb062d03070e5090a7bb2456cb20cb4ce078f502515cda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ac6ce7701f1c6aa1ea8fcf7a3035e3b0eaa28ae3cb6afc442372bd9117bdde4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:39Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.227382 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4742ea84b3a2c9461d9b900beee7354a2ded24c39ad23e007cbf270ae12377c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:39Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.238523 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:39Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.240126 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.240177 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.240187 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.240204 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.240218 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:39Z","lastTransitionTime":"2025-12-07T16:03:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.250329 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c15b59eb-565d-4556-a4ce-75afdf159dc8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://530a2935826940767ea33fb4f37fa11f7dee4950d867fbcd6af72836d80f7db3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jks8z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ccaf6c1e6a978eb978a1c207b6c4c00b23737f78b743f966579aca781e22e0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jks8z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zcxxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:39Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.267997 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8ps9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98015cea-d753-4e7a-b2b5-5ae7733a81e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d68509b673eecd37725cdcc48e3d0612c3b9c7601433f2b72c21d662140daa59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f84f52a073437543721010dc440339654ffb5e04eeb6cc7afbcc62f00f64c8e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f84f52a073437543721010dc440339654ffb5e04eeb6cc7afbcc62f00f64c8e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f6fe1cb25218cb3ac20488dc47efa4dc6a26a43dbc8306b5ee38cbe1a186c67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f6fe1cb25218cb3ac20488dc47efa4dc6a26a43dbc8306b5ee38cbe1a186c67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14b748a8247f6863e8be61e90057a06626dfd3cd4f01bd3a11629f700f56378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f14b748a8247f6863e8be61e90057a06626dfd3cd4f01bd3a11629f700f56378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ec379c0d573cb9b5ffcfc5ca7c653ba60e2fbdd7e3d37836102e26618ce43e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ec379c0d573cb9b5ffcfc5ca7c653ba60e2fbdd7e3d37836102e26618ce43e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c75bf2a6703cf99865d4ed5fa000e1e3b1330d4946176cf585fec1a2f226df0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c75bf2a6703cf99865d4ed5fa000e1e3b1330d4946176cf585fec1a2f226df0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de45812c9a5a5bfe6b0146828d03fab28362b03d04ef861c150b9ffbf181d345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de45812c9a5a5bfe6b0146828d03fab28362b03d04ef861c150b9ffbf181d345\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8ps9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:39Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.280366 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e903f3c-ba8c-477d-b66a-6d1331a7c64e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ff9c456318bbdb54ec3f1ac3e2c2366968f18ad084158b7100d142acb668948\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1358745250b35426602fc7417e4f85dd411cdd0fb3e8f703c432d72128ca7c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d8c9a434c6cfb9e4fee6a8955f1379291f153c007876d1ffb498ce561c73ba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad8df9b5bd5f5a8dd44ce2899cfa29c19435149b170d339d339183cd3df09e99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad8df9b5bd5f5a8dd44ce2899cfa29c19435149b170d339d339183cd3df09e99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:39Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.299858 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c54d618a-dfec-4f2a-a288-f5188c8a128a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60324dc7d701bc064e14ae4f474bce10f6ee4ff3c3c942eab749e7166082cf9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ed4f4d3564674bf5cb4e7a76dc0fd95e203ecc8950798e59a1fb6c9ab4a20f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://def9ac840fb8ecf2778bed88ea980eddca7c5b2aba4bd1afb967d62fec3f7bae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b73be664f5412c98b89d971a2cac37c494151c293fdb44a98b390978658a6643\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a847d04e22e98091c0037a89054b9236228b701b8c7ba31940a56df0f37b95c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcae78a6133d97f6a32dd7421ccec4333b970a237404e056a5cc601f8eefdc8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c60794c12cb6e7647d4c1d0bee76e214a71d9cb818d9c98dbf9880fb030ed54b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f48167b38ff840f85f11b69d8090ec6c9c634ed26682daef93f5da99fc02ef61\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-07T16:03:08Z\\\",\\\"message\\\":\\\"er, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:08Z is after 2025-08-24T17:21:41Z]\\\\nI1207 16:03:08.453903 6420 services_controller.go:451] Built service openshift-controller-manager-operator/metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-controller-manager-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-controller-manager-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.58\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.Tem\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T16:03:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:03:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcfdffe822635b0ae13b253e6133449fe00461b2792be43d913c4a77219b8255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9bfe2ffce7a7b7c320474117734fc05246cc2b4fd744833e896e8e7c83f5e4d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9bfe2ffce7a7b7c320474117734fc05246cc2b4fd744833e896e8e7c83f5e4d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qvc7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:39Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.314177 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mx85f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"118dd9ba-437c-4e5c-b5ff-4239ae6d4a63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bcd6c8deb184d0c713c5e7b7cc82fa38d3cd498517911ba7a0ec3cb88a3437a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gg5nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ffc069d2b21fa80871b1b97f2c72c37880d9a309cbf8b7daf7d240f2c332a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gg5nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mx85f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:39Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.327702 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nx2lh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04b37f6c-457b-4f9a-8733-e88cb9644483\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv6td\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv6td\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nx2lh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:39Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.342778 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.342816 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.342827 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.342841 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.342850 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:39Z","lastTransitionTime":"2025-12-07T16:03:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.342808 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdee9957-48b9-458f-ba71-e0cd39f80596\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73f85248b4d68a9beecdacb3a01f433e97a0ffea2247ce92a41ef96855509377\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6276df652efe6ca935e745f24ef20f87a16465475139230c63b9999234a4d12a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3b1dd202586ee755ca89fbe12b30bbb12f95a62514f6e4d93cbbdbe55181e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd318647816e4a27849666b0ec4fc2ca4905295661c557dc381e1144f22cf284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d31de0326a05216fa10442bdb47151c6e073c97e9a16c34e793028e5e253756\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"le observer\\\\nW1207 16:02:34.645464 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1207 16:02:34.645587 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 16:02:34.646501 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2082283479/tls.crt::/tmp/serving-cert-2082283479/tls.key\\\\\\\"\\\\nI1207 16:02:34.972041 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 16:02:34.973702 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 16:02:34.973717 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 16:02:34.973738 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 16:02:34.973743 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 16:02:34.977185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1207 16:02:34.977204 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1207 16:02:34.977209 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977217 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 16:02:34.977225 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 16:02:34.977228 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 16:02:34.977232 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1207 16:02:34.979033 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27bd4a54f920b2725987c1c7fd2d9fdf16cd08738c246c333a4e9164a3a582a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:39Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.359002 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:39Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.370447 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2jjld" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dee6f5c-861c-4819-9309-7791574c7b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51c1d4159ab11087839a2c2d192566fb9322447c26df560bf09517c906ececb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmdns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2jjld\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:39Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.385138 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4wr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0baf8575-26cf-4a72-8a6d-ae94c078eaad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fca3b2ea503451cfaf7afb00445fe0adb0acc87d436d1eebba5068452b8e7cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0ca2b3cd5a11033afd219b0125be7ab5acd750935d7eb89a470eff3d93a0838\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-07T16:03:26Z\\\",\\\"message\\\":\\\"2025-12-07T16:02:41+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_09397239-db79-454b-b301-47a1e4cdfc48\\\\n2025-12-07T16:02:41+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_09397239-db79-454b-b301-47a1e4cdfc48 to /host/opt/cni/bin/\\\\n2025-12-07T16:02:41Z [verbose] multus-daemon started\\\\n2025-12-07T16:02:41Z [verbose] Readiness Indicator file check\\\\n2025-12-07T16:03:26Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:03:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p2trd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4wr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:39Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.394998 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-m5jl5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"824ee59f-1474-4994-a70e-462878521514\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a15fe50cc75515704779646cc19aa50004f0db9aa26fe2535110df749bbfb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhwzn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-m5jl5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:39Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.444809 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.444847 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.444859 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.444876 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.444888 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:39Z","lastTransitionTime":"2025-12-07T16:03:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.490198 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 16:03:39 crc kubenswrapper[4716]: E1207 16:03:39.490370 4716 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 07 16:03:39 crc kubenswrapper[4716]: E1207 16:03:39.490405 4716 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.490416 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 16:03:39 crc kubenswrapper[4716]: E1207 16:03:39.490425 4716 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 07 16:03:39 crc kubenswrapper[4716]: E1207 16:03:39.490495 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-07 16:04:43.490472725 +0000 UTC m=+146.180757687 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 07 16:03:39 crc kubenswrapper[4716]: E1207 16:03:39.490560 4716 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 07 16:03:39 crc kubenswrapper[4716]: E1207 16:03:39.490581 4716 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 07 16:03:39 crc kubenswrapper[4716]: E1207 16:03:39.490596 4716 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 07 16:03:39 crc kubenswrapper[4716]: E1207 16:03:39.490650 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-07 16:04:43.49063281 +0000 UTC m=+146.180917732 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.547888 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.547947 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.547957 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.547973 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.547984 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:39Z","lastTransitionTime":"2025-12-07T16:03:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.591027 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.591208 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.591248 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:03:39 crc kubenswrapper[4716]: E1207 16:03:39.591296 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 16:04:43.59125176 +0000 UTC m=+146.281536712 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:03:39 crc kubenswrapper[4716]: E1207 16:03:39.591355 4716 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 07 16:03:39 crc kubenswrapper[4716]: E1207 16:03:39.591390 4716 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 07 16:03:39 crc kubenswrapper[4716]: E1207 16:03:39.591422 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-07 16:04:43.591403984 +0000 UTC m=+146.281689016 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 07 16:03:39 crc kubenswrapper[4716]: E1207 16:03:39.591487 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-07 16:04:43.591460676 +0000 UTC m=+146.281745648 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.651310 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.651383 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.651394 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.651410 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.651419 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:39Z","lastTransitionTime":"2025-12-07T16:03:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.656624 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.656686 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nx2lh" Dec 07 16:03:39 crc kubenswrapper[4716]: E1207 16:03:39.656714 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.656765 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:03:39 crc kubenswrapper[4716]: E1207 16:03:39.656811 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nx2lh" podUID="04b37f6c-457b-4f9a-8733-e88cb9644483" Dec 07 16:03:39 crc kubenswrapper[4716]: E1207 16:03:39.656880 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.754462 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.754505 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.754520 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.754539 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.754554 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:39Z","lastTransitionTime":"2025-12-07T16:03:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.857609 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.857830 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.857842 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.857856 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.857867 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:39Z","lastTransitionTime":"2025-12-07T16:03:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.960846 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.960886 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.960898 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.960915 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:39 crc kubenswrapper[4716]: I1207 16:03:39.960926 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:39Z","lastTransitionTime":"2025-12-07T16:03:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:40 crc kubenswrapper[4716]: I1207 16:03:40.064181 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:40 crc kubenswrapper[4716]: I1207 16:03:40.064244 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:40 crc kubenswrapper[4716]: I1207 16:03:40.064259 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:40 crc kubenswrapper[4716]: I1207 16:03:40.064276 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:40 crc kubenswrapper[4716]: I1207 16:03:40.064289 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:40Z","lastTransitionTime":"2025-12-07T16:03:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:40 crc kubenswrapper[4716]: I1207 16:03:40.127651 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qvc7p_c54d618a-dfec-4f2a-a288-f5188c8a128a/ovnkube-controller/3.log" Dec 07 16:03:40 crc kubenswrapper[4716]: I1207 16:03:40.128676 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qvc7p_c54d618a-dfec-4f2a-a288-f5188c8a128a/ovnkube-controller/2.log" Dec 07 16:03:40 crc kubenswrapper[4716]: I1207 16:03:40.132266 4716 generic.go:334] "Generic (PLEG): container finished" podID="c54d618a-dfec-4f2a-a288-f5188c8a128a" containerID="c60794c12cb6e7647d4c1d0bee76e214a71d9cb818d9c98dbf9880fb030ed54b" exitCode=1 Dec 07 16:03:40 crc kubenswrapper[4716]: I1207 16:03:40.132317 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" event={"ID":"c54d618a-dfec-4f2a-a288-f5188c8a128a","Type":"ContainerDied","Data":"c60794c12cb6e7647d4c1d0bee76e214a71d9cb818d9c98dbf9880fb030ed54b"} Dec 07 16:03:40 crc kubenswrapper[4716]: I1207 16:03:40.132367 4716 scope.go:117] "RemoveContainer" containerID="f48167b38ff840f85f11b69d8090ec6c9c634ed26682daef93f5da99fc02ef61" Dec 07 16:03:40 crc kubenswrapper[4716]: I1207 16:03:40.132921 4716 scope.go:117] "RemoveContainer" containerID="c60794c12cb6e7647d4c1d0bee76e214a71d9cb818d9c98dbf9880fb030ed54b" Dec 07 16:03:40 crc kubenswrapper[4716]: E1207 16:03:40.133115 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-qvc7p_openshift-ovn-kubernetes(c54d618a-dfec-4f2a-a288-f5188c8a128a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" podUID="c54d618a-dfec-4f2a-a288-f5188c8a128a" Dec 07 16:03:40 crc kubenswrapper[4716]: I1207 16:03:40.154254 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8ps9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98015cea-d753-4e7a-b2b5-5ae7733a81e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d68509b673eecd37725cdcc48e3d0612c3b9c7601433f2b72c21d662140daa59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f84f52a073437543721010dc440339654ffb5e04eeb6cc7afbcc62f00f64c8e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f84f52a073437543721010dc440339654ffb5e04eeb6cc7afbcc62f00f64c8e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f6fe1cb25218cb3ac20488dc47efa4dc6a26a43dbc8306b5ee38cbe1a186c67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f6fe1cb25218cb3ac20488dc47efa4dc6a26a43dbc8306b5ee38cbe1a186c67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14b748a8247f6863e8be61e90057a06626dfd3cd4f01bd3a11629f700f56378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f14b748a8247f6863e8be61e90057a06626dfd3cd4f01bd3a11629f700f56378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ec379c0d573cb9b5ffcfc5ca7c653ba60e2fbdd7e3d37836102e26618ce43e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ec379c0d573cb9b5ffcfc5ca7c653ba60e2fbdd7e3d37836102e26618ce43e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c75bf2a6703cf99865d4ed5fa000e1e3b1330d4946176cf585fec1a2f226df0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c75bf2a6703cf99865d4ed5fa000e1e3b1330d4946176cf585fec1a2f226df0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de45812c9a5a5bfe6b0146828d03fab28362b03d04ef861c150b9ffbf181d345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de45812c9a5a5bfe6b0146828d03fab28362b03d04ef861c150b9ffbf181d345\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8ps9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:40Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:40 crc kubenswrapper[4716]: I1207 16:03:40.169234 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:40 crc kubenswrapper[4716]: I1207 16:03:40.169284 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:40 crc kubenswrapper[4716]: I1207 16:03:40.169296 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:40 crc kubenswrapper[4716]: I1207 16:03:40.169313 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:40 crc kubenswrapper[4716]: I1207 16:03:40.169329 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:40Z","lastTransitionTime":"2025-12-07T16:03:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:40 crc kubenswrapper[4716]: I1207 16:03:40.179729 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7961cc9a-734e-42c2-b1ff-7075f3f20091\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84a36b76f20618c7885d611ecad9db29bac5101c27572be31d3acbb90c581291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0a0eca3c2087a457c0e448e649345ec4595a9c8a019aed2cddc5f30c77871e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://404426a171b8676eddd4a00996c76afab7e269643fa9608f1e1b6bdba385071d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fd3c2941782921f99a6cdc3de4e892234548bde527fec15a8ac92e513566e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e23abfdca314ed43d594333c9f9e401cc8ef7dad39b8ebac84a836b06ee5c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09b38cd4b389ea477c802a842ee4f2297a64d64684624117163e819539c257c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09b38cd4b389ea477c802a842ee4f2297a64d64684624117163e819539c257c6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c00f7a8e024e0e0dfd7b73e16d1d5437602e79ea95b73c21f24b209edef60e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c00f7a8e024e0e0dfd7b73e16d1d5437602e79ea95b73c21f24b209edef60e98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3f6392d426fa2cd788ec3a5c3bff4ac39483a54dfb2624a1c4370522413a350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3f6392d426fa2cd788ec3a5c3bff4ac39483a54dfb2624a1c4370522413a350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:40Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:40 crc kubenswrapper[4716]: I1207 16:03:40.200577 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3067de7c-c11c-4be1-bd94-d6572c266795\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d14d018fa044e9e80fe81e7e38df935992d14c773ed2b2193fd5cb873e5d0947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a9e9cf048220ca14c996847e9ced9f2367b9331dc6c2836dd023658855cd79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6d3c911874095b3c7cb062d03070e5090a7bb2456cb20cb4ce078f502515cda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ac6ce7701f1c6aa1ea8fcf7a3035e3b0eaa28ae3cb6afc442372bd9117bdde4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:40Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:40 crc kubenswrapper[4716]: I1207 16:03:40.218617 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4742ea84b3a2c9461d9b900beee7354a2ded24c39ad23e007cbf270ae12377c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:40Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:40 crc kubenswrapper[4716]: I1207 16:03:40.232271 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:40Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:40 crc kubenswrapper[4716]: I1207 16:03:40.249633 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c15b59eb-565d-4556-a4ce-75afdf159dc8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://530a2935826940767ea33fb4f37fa11f7dee4950d867fbcd6af72836d80f7db3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jks8z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ccaf6c1e6a978eb978a1c207b6c4c00b23737f78b743f966579aca781e22e0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jks8z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zcxxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:40Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:40 crc kubenswrapper[4716]: I1207 16:03:40.265369 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e903f3c-ba8c-477d-b66a-6d1331a7c64e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ff9c456318bbdb54ec3f1ac3e2c2366968f18ad084158b7100d142acb668948\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1358745250b35426602fc7417e4f85dd411cdd0fb3e8f703c432d72128ca7c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d8c9a434c6cfb9e4fee6a8955f1379291f153c007876d1ffb498ce561c73ba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad8df9b5bd5f5a8dd44ce2899cfa29c19435149b170d339d339183cd3df09e99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad8df9b5bd5f5a8dd44ce2899cfa29c19435149b170d339d339183cd3df09e99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:40Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:40 crc kubenswrapper[4716]: I1207 16:03:40.272407 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:40 crc kubenswrapper[4716]: I1207 16:03:40.272452 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:40 crc kubenswrapper[4716]: I1207 16:03:40.272465 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:40 crc kubenswrapper[4716]: I1207 16:03:40.272484 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:40 crc kubenswrapper[4716]: I1207 16:03:40.272496 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:40Z","lastTransitionTime":"2025-12-07T16:03:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:40 crc kubenswrapper[4716]: I1207 16:03:40.296976 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c54d618a-dfec-4f2a-a288-f5188c8a128a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60324dc7d701bc064e14ae4f474bce10f6ee4ff3c3c942eab749e7166082cf9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ed4f4d3564674bf5cb4e7a76dc0fd95e203ecc8950798e59a1fb6c9ab4a20f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://def9ac840fb8ecf2778bed88ea980eddca7c5b2aba4bd1afb967d62fec3f7bae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b73be664f5412c98b89d971a2cac37c494151c293fdb44a98b390978658a6643\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a847d04e22e98091c0037a89054b9236228b701b8c7ba31940a56df0f37b95c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcae78a6133d97f6a32dd7421ccec4333b970a237404e056a5cc601f8eefdc8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c60794c12cb6e7647d4c1d0bee76e214a71d9cb818d9c98dbf9880fb030ed54b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f48167b38ff840f85f11b69d8090ec6c9c634ed26682daef93f5da99fc02ef61\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-07T16:03:08Z\\\",\\\"message\\\":\\\"er, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:08Z is after 2025-08-24T17:21:41Z]\\\\nI1207 16:03:08.453903 6420 services_controller.go:451] Built service openshift-controller-manager-operator/metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-controller-manager-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-controller-manager-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.58\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.Tem\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T16:03:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c60794c12cb6e7647d4c1d0bee76e214a71d9cb818d9c98dbf9880fb030ed54b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-07T16:03:39Z\\\",\\\"message\\\":\\\"reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.88:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {ebd4748e-0473-49fb-88ad-83dbb221791a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1207 16:03:39.306886 6826 obj_retry.go:303] Retry object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mx85f\\\\nI1207 16:03:39.308511 6826 obj_retry.go:365] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mx85f\\\\nI1207 16:03:39.308518 6826 ovn.go:134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mx85f in node crc\\\\nI1207 16:03:39.308525 6826 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mx85f after 0 failed attempt(s)\\\\nI1207 16:03:39.308531 6826 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mx85f\\\\nF1207 16:03:39.307073 6826 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T16:03:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcfdffe822635b0ae13b253e6133449fe00461b2792be43d913c4a77219b8255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9bfe2ffce7a7b7c320474117734fc05246cc2b4fd744833e896e8e7c83f5e4d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9bfe2ffce7a7b7c320474117734fc05246cc2b4fd744833e896e8e7c83f5e4d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qvc7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:40Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:40 crc kubenswrapper[4716]: I1207 16:03:40.314240 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mx85f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"118dd9ba-437c-4e5c-b5ff-4239ae6d4a63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bcd6c8deb184d0c713c5e7b7cc82fa38d3cd498517911ba7a0ec3cb88a3437a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gg5nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ffc069d2b21fa80871b1b97f2c72c37880d9a309cbf8b7daf7d240f2c332a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gg5nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mx85f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:40Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:40 crc kubenswrapper[4716]: I1207 16:03:40.330311 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nx2lh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04b37f6c-457b-4f9a-8733-e88cb9644483\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv6td\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv6td\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nx2lh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:40Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:40 crc kubenswrapper[4716]: I1207 16:03:40.349676 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdee9957-48b9-458f-ba71-e0cd39f80596\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73f85248b4d68a9beecdacb3a01f433e97a0ffea2247ce92a41ef96855509377\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6276df652efe6ca935e745f24ef20f87a16465475139230c63b9999234a4d12a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3b1dd202586ee755ca89fbe12b30bbb12f95a62514f6e4d93cbbdbe55181e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd318647816e4a27849666b0ec4fc2ca4905295661c557dc381e1144f22cf284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d31de0326a05216fa10442bdb47151c6e073c97e9a16c34e793028e5e253756\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"le observer\\\\nW1207 16:02:34.645464 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1207 16:02:34.645587 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 16:02:34.646501 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2082283479/tls.crt::/tmp/serving-cert-2082283479/tls.key\\\\\\\"\\\\nI1207 16:02:34.972041 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 16:02:34.973702 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 16:02:34.973717 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 16:02:34.973738 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 16:02:34.973743 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 16:02:34.977185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1207 16:02:34.977204 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1207 16:02:34.977209 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977217 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 16:02:34.977225 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 16:02:34.977228 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 16:02:34.977232 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1207 16:02:34.979033 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27bd4a54f920b2725987c1c7fd2d9fdf16cd08738c246c333a4e9164a3a582a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:40Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:40 crc kubenswrapper[4716]: I1207 16:03:40.367573 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:40Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:40 crc kubenswrapper[4716]: I1207 16:03:40.375317 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:40 crc kubenswrapper[4716]: I1207 16:03:40.375365 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:40 crc kubenswrapper[4716]: I1207 16:03:40.375379 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:40 crc kubenswrapper[4716]: I1207 16:03:40.375395 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:40 crc kubenswrapper[4716]: I1207 16:03:40.375407 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:40Z","lastTransitionTime":"2025-12-07T16:03:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:40 crc kubenswrapper[4716]: I1207 16:03:40.378593 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2jjld" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dee6f5c-861c-4819-9309-7791574c7b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51c1d4159ab11087839a2c2d192566fb9322447c26df560bf09517c906ececb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmdns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2jjld\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:40Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:40 crc kubenswrapper[4716]: I1207 16:03:40.394143 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4wr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0baf8575-26cf-4a72-8a6d-ae94c078eaad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fca3b2ea503451cfaf7afb00445fe0adb0acc87d436d1eebba5068452b8e7cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0ca2b3cd5a11033afd219b0125be7ab5acd750935d7eb89a470eff3d93a0838\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-07T16:03:26Z\\\",\\\"message\\\":\\\"2025-12-07T16:02:41+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_09397239-db79-454b-b301-47a1e4cdfc48\\\\n2025-12-07T16:02:41+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_09397239-db79-454b-b301-47a1e4cdfc48 to /host/opt/cni/bin/\\\\n2025-12-07T16:02:41Z [verbose] multus-daemon started\\\\n2025-12-07T16:02:41Z [verbose] Readiness Indicator file check\\\\n2025-12-07T16:03:26Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:03:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p2trd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4wr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:40Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:40 crc kubenswrapper[4716]: I1207 16:03:40.405160 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-m5jl5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"824ee59f-1474-4994-a70e-462878521514\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a15fe50cc75515704779646cc19aa50004f0db9aa26fe2535110df749bbfb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhwzn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-m5jl5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:40Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:40 crc kubenswrapper[4716]: I1207 16:03:40.416700 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71d1a4e4-5c15-4aa2-b24c-75c1f7729297\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1ab9067e4d9988fd94f93d7798954e4c8888875f47244d797cb6b0557ce5582\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3d77f99006bada2a3a676398c5ddb366c0c727f3837f10674a9d582dd83e2cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3d77f99006bada2a3a676398c5ddb366c0c727f3837f10674a9d582dd83e2cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:40Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:40 crc kubenswrapper[4716]: I1207 16:03:40.432973 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef3335ae3875a75ab64685fc7e3590177280d651e8f8292c6f5ea6f472e55789\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:40Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:40 crc kubenswrapper[4716]: I1207 16:03:40.450838 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:40Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:40 crc kubenswrapper[4716]: I1207 16:03:40.464424 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f95ef61779507288a03bda8eddf745e105f705998a117a30789675bfa452d88f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b955b23163e7c8c458a3bbe2f681633e6a663ce76748febcc0b3a3c37f3ca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:40Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:40 crc kubenswrapper[4716]: I1207 16:03:40.478042 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:40 crc kubenswrapper[4716]: I1207 16:03:40.478124 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:40 crc kubenswrapper[4716]: I1207 16:03:40.478139 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:40 crc kubenswrapper[4716]: I1207 16:03:40.478161 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:40 crc kubenswrapper[4716]: I1207 16:03:40.478177 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:40Z","lastTransitionTime":"2025-12-07T16:03:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:40 crc kubenswrapper[4716]: I1207 16:03:40.580232 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:40 crc kubenswrapper[4716]: I1207 16:03:40.580311 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:40 crc kubenswrapper[4716]: I1207 16:03:40.580333 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:40 crc kubenswrapper[4716]: I1207 16:03:40.580362 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:40 crc kubenswrapper[4716]: I1207 16:03:40.580383 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:40Z","lastTransitionTime":"2025-12-07T16:03:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:40 crc kubenswrapper[4716]: I1207 16:03:40.657260 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 16:03:40 crc kubenswrapper[4716]: E1207 16:03:40.657400 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 16:03:40 crc kubenswrapper[4716]: I1207 16:03:40.683716 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:40 crc kubenswrapper[4716]: I1207 16:03:40.683784 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:40 crc kubenswrapper[4716]: I1207 16:03:40.683806 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:40 crc kubenswrapper[4716]: I1207 16:03:40.683839 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:40 crc kubenswrapper[4716]: I1207 16:03:40.683860 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:40Z","lastTransitionTime":"2025-12-07T16:03:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:40 crc kubenswrapper[4716]: I1207 16:03:40.787424 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:40 crc kubenswrapper[4716]: I1207 16:03:40.787493 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:40 crc kubenswrapper[4716]: I1207 16:03:40.787517 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:40 crc kubenswrapper[4716]: I1207 16:03:40.787545 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:40 crc kubenswrapper[4716]: I1207 16:03:40.787565 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:40Z","lastTransitionTime":"2025-12-07T16:03:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:40 crc kubenswrapper[4716]: I1207 16:03:40.890323 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:40 crc kubenswrapper[4716]: I1207 16:03:40.890361 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:40 crc kubenswrapper[4716]: I1207 16:03:40.890370 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:40 crc kubenswrapper[4716]: I1207 16:03:40.890403 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:40 crc kubenswrapper[4716]: I1207 16:03:40.890413 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:40Z","lastTransitionTime":"2025-12-07T16:03:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:40 crc kubenswrapper[4716]: I1207 16:03:40.992962 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:40 crc kubenswrapper[4716]: I1207 16:03:40.993030 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:40 crc kubenswrapper[4716]: I1207 16:03:40.993050 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:40 crc kubenswrapper[4716]: I1207 16:03:40.993121 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:40 crc kubenswrapper[4716]: I1207 16:03:40.993148 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:40Z","lastTransitionTime":"2025-12-07T16:03:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:41 crc kubenswrapper[4716]: I1207 16:03:41.096044 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:41 crc kubenswrapper[4716]: I1207 16:03:41.096161 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:41 crc kubenswrapper[4716]: I1207 16:03:41.096184 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:41 crc kubenswrapper[4716]: I1207 16:03:41.096318 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:41 crc kubenswrapper[4716]: I1207 16:03:41.096370 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:41Z","lastTransitionTime":"2025-12-07T16:03:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:41 crc kubenswrapper[4716]: I1207 16:03:41.137866 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qvc7p_c54d618a-dfec-4f2a-a288-f5188c8a128a/ovnkube-controller/3.log" Dec 07 16:03:41 crc kubenswrapper[4716]: I1207 16:03:41.140813 4716 scope.go:117] "RemoveContainer" containerID="c60794c12cb6e7647d4c1d0bee76e214a71d9cb818d9c98dbf9880fb030ed54b" Dec 07 16:03:41 crc kubenswrapper[4716]: E1207 16:03:41.140997 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-qvc7p_openshift-ovn-kubernetes(c54d618a-dfec-4f2a-a288-f5188c8a128a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" podUID="c54d618a-dfec-4f2a-a288-f5188c8a128a" Dec 07 16:03:41 crc kubenswrapper[4716]: I1207 16:03:41.158448 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e903f3c-ba8c-477d-b66a-6d1331a7c64e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ff9c456318bbdb54ec3f1ac3e2c2366968f18ad084158b7100d142acb668948\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1358745250b35426602fc7417e4f85dd411cdd0fb3e8f703c432d72128ca7c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d8c9a434c6cfb9e4fee6a8955f1379291f153c007876d1ffb498ce561c73ba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad8df9b5bd5f5a8dd44ce2899cfa29c19435149b170d339d339183cd3df09e99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad8df9b5bd5f5a8dd44ce2899cfa29c19435149b170d339d339183cd3df09e99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:41Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:41 crc kubenswrapper[4716]: I1207 16:03:41.180462 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c54d618a-dfec-4f2a-a288-f5188c8a128a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60324dc7d701bc064e14ae4f474bce10f6ee4ff3c3c942eab749e7166082cf9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ed4f4d3564674bf5cb4e7a76dc0fd95e203ecc8950798e59a1fb6c9ab4a20f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://def9ac840fb8ecf2778bed88ea980eddca7c5b2aba4bd1afb967d62fec3f7bae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b73be664f5412c98b89d971a2cac37c494151c293fdb44a98b390978658a6643\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a847d04e22e98091c0037a89054b9236228b701b8c7ba31940a56df0f37b95c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcae78a6133d97f6a32dd7421ccec4333b970a237404e056a5cc601f8eefdc8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c60794c12cb6e7647d4c1d0bee76e214a71d9cb818d9c98dbf9880fb030ed54b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c60794c12cb6e7647d4c1d0bee76e214a71d9cb818d9c98dbf9880fb030ed54b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-07T16:03:39Z\\\",\\\"message\\\":\\\"reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.88:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {ebd4748e-0473-49fb-88ad-83dbb221791a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1207 16:03:39.306886 6826 obj_retry.go:303] Retry object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mx85f\\\\nI1207 16:03:39.308511 6826 obj_retry.go:365] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mx85f\\\\nI1207 16:03:39.308518 6826 ovn.go:134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mx85f in node crc\\\\nI1207 16:03:39.308525 6826 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mx85f after 0 failed attempt(s)\\\\nI1207 16:03:39.308531 6826 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mx85f\\\\nF1207 16:03:39.307073 6826 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T16:03:38Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-qvc7p_openshift-ovn-kubernetes(c54d618a-dfec-4f2a-a288-f5188c8a128a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcfdffe822635b0ae13b253e6133449fe00461b2792be43d913c4a77219b8255\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9bfe2ffce7a7b7c320474117734fc05246cc2b4fd744833e896e8e7c83f5e4d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9bfe2ffce7a7b7c320474117734fc05246cc2b4fd744833e896e8e7c83f5e4d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qphz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qvc7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:41Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:41 crc kubenswrapper[4716]: I1207 16:03:41.195042 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mx85f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"118dd9ba-437c-4e5c-b5ff-4239ae6d4a63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bcd6c8deb184d0c713c5e7b7cc82fa38d3cd498517911ba7a0ec3cb88a3437a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gg5nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ffc069d2b21fa80871b1b97f2c72c37880d9a309cbf8b7daf7d240f2c332a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gg5nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mx85f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:41Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:41 crc kubenswrapper[4716]: I1207 16:03:41.199356 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:41 crc kubenswrapper[4716]: I1207 16:03:41.199385 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:41 crc kubenswrapper[4716]: I1207 16:03:41.199394 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:41 crc kubenswrapper[4716]: I1207 16:03:41.199407 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:41 crc kubenswrapper[4716]: I1207 16:03:41.199432 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:41Z","lastTransitionTime":"2025-12-07T16:03:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:41 crc kubenswrapper[4716]: I1207 16:03:41.209701 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nx2lh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04b37f6c-457b-4f9a-8733-e88cb9644483\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv6td\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bv6td\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:54Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nx2lh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:41Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:41 crc kubenswrapper[4716]: I1207 16:03:41.222019 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-m5jl5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"824ee59f-1474-4994-a70e-462878521514\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a15fe50cc75515704779646cc19aa50004f0db9aa26fe2535110df749bbfb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhwzn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-m5jl5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:41Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:41 crc kubenswrapper[4716]: I1207 16:03:41.237978 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cdee9957-48b9-458f-ba71-e0cd39f80596\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73f85248b4d68a9beecdacb3a01f433e97a0ffea2247ce92a41ef96855509377\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6276df652efe6ca935e745f24ef20f87a16465475139230c63b9999234a4d12a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9c3b1dd202586ee755ca89fbe12b30bbb12f95a62514f6e4d93cbbdbe55181e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd318647816e4a27849666b0ec4fc2ca4905295661c557dc381e1144f22cf284\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d31de0326a05216fa10442bdb47151c6e073c97e9a16c34e793028e5e253756\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"le observer\\\\nW1207 16:02:34.645464 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1207 16:02:34.645587 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1207 16:02:34.646501 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2082283479/tls.crt::/tmp/serving-cert-2082283479/tls.key\\\\\\\"\\\\nI1207 16:02:34.972041 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1207 16:02:34.973702 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1207 16:02:34.973717 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1207 16:02:34.973738 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1207 16:02:34.973743 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1207 16:02:34.977185 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1207 16:02:34.977204 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1207 16:02:34.977209 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977217 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1207 16:02:34.977221 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1207 16:02:34.977225 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1207 16:02:34.977228 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1207 16:02:34.977232 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1207 16:02:34.979033 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27bd4a54f920b2725987c1c7fd2d9fdf16cd08738c246c333a4e9164a3a582a0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:41Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:41 crc kubenswrapper[4716]: I1207 16:03:41.255472 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:41Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:41 crc kubenswrapper[4716]: I1207 16:03:41.266856 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2jjld" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dee6f5c-861c-4819-9309-7791574c7b80\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51c1d4159ab11087839a2c2d192566fb9322447c26df560bf09517c906ececb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rmdns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2jjld\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:41Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:41 crc kubenswrapper[4716]: I1207 16:03:41.285507 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k4wr2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0baf8575-26cf-4a72-8a6d-ae94c078eaad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:03:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fca3b2ea503451cfaf7afb00445fe0adb0acc87d436d1eebba5068452b8e7cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0ca2b3cd5a11033afd219b0125be7ab5acd750935d7eb89a470eff3d93a0838\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-07T16:03:26Z\\\",\\\"message\\\":\\\"2025-12-07T16:02:41+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_09397239-db79-454b-b301-47a1e4cdfc48\\\\n2025-12-07T16:02:41+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_09397239-db79-454b-b301-47a1e4cdfc48 to /host/opt/cni/bin/\\\\n2025-12-07T16:02:41Z [verbose] multus-daemon started\\\\n2025-12-07T16:02:41Z [verbose] Readiness Indicator file check\\\\n2025-12-07T16:03:26Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:03:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p2trd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k4wr2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:41Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:41 crc kubenswrapper[4716]: I1207 16:03:41.297921 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71d1a4e4-5c15-4aa2-b24c-75c1f7729297\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1ab9067e4d9988fd94f93d7798954e4c8888875f47244d797cb6b0557ce5582\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3d77f99006bada2a3a676398c5ddb366c0c727f3837f10674a9d582dd83e2cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3d77f99006bada2a3a676398c5ddb366c0c727f3837f10674a9d582dd83e2cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:41Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:41 crc kubenswrapper[4716]: I1207 16:03:41.301998 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:41 crc kubenswrapper[4716]: I1207 16:03:41.302052 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:41 crc kubenswrapper[4716]: I1207 16:03:41.302071 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:41 crc kubenswrapper[4716]: I1207 16:03:41.302134 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:41 crc kubenswrapper[4716]: I1207 16:03:41.302152 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:41Z","lastTransitionTime":"2025-12-07T16:03:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:41 crc kubenswrapper[4716]: I1207 16:03:41.312768 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef3335ae3875a75ab64685fc7e3590177280d651e8f8292c6f5ea6f472e55789\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:41Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:41 crc kubenswrapper[4716]: I1207 16:03:41.330189 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:41Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:41 crc kubenswrapper[4716]: I1207 16:03:41.345245 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f95ef61779507288a03bda8eddf745e105f705998a117a30789675bfa452d88f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b955b23163e7c8c458a3bbe2f681633e6a663ce76748febcc0b3a3c37f3ca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:41Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:41 crc kubenswrapper[4716]: I1207 16:03:41.360582 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c15b59eb-565d-4556-a4ce-75afdf159dc8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://530a2935826940767ea33fb4f37fa11f7dee4950d867fbcd6af72836d80f7db3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jks8z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ccaf6c1e6a978eb978a1c207b6c4c00b23737f78b743f966579aca781e22e0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jks8z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zcxxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:41Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:41 crc kubenswrapper[4716]: I1207 16:03:41.384565 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-8ps9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98015cea-d753-4e7a-b2b5-5ae7733a81e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d68509b673eecd37725cdcc48e3d0612c3b9c7601433f2b72c21d662140daa59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f84f52a073437543721010dc440339654ffb5e04eeb6cc7afbcc62f00f64c8e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f84f52a073437543721010dc440339654ffb5e04eeb6cc7afbcc62f00f64c8e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f6fe1cb25218cb3ac20488dc47efa4dc6a26a43dbc8306b5ee38cbe1a186c67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f6fe1cb25218cb3ac20488dc47efa4dc6a26a43dbc8306b5ee38cbe1a186c67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f14b748a8247f6863e8be61e90057a06626dfd3cd4f01bd3a11629f700f56378\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f14b748a8247f6863e8be61e90057a06626dfd3cd4f01bd3a11629f700f56378\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ec379c0d573cb9b5ffcfc5ca7c653ba60e2fbdd7e3d37836102e26618ce43e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ec379c0d573cb9b5ffcfc5ca7c653ba60e2fbdd7e3d37836102e26618ce43e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c75bf2a6703cf99865d4ed5fa000e1e3b1330d4946176cf585fec1a2f226df0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c75bf2a6703cf99865d4ed5fa000e1e3b1330d4946176cf585fec1a2f226df0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de45812c9a5a5bfe6b0146828d03fab28362b03d04ef861c150b9ffbf181d345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de45812c9a5a5bfe6b0146828d03fab28362b03d04ef861c150b9ffbf181d345\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-llw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-8ps9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:41Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:41 crc kubenswrapper[4716]: I1207 16:03:41.405117 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:41 crc kubenswrapper[4716]: I1207 16:03:41.405171 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:41 crc kubenswrapper[4716]: I1207 16:03:41.405216 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:41 crc kubenswrapper[4716]: I1207 16:03:41.405233 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:41 crc kubenswrapper[4716]: I1207 16:03:41.405246 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:41Z","lastTransitionTime":"2025-12-07T16:03:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:41 crc kubenswrapper[4716]: I1207 16:03:41.407603 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7961cc9a-734e-42c2-b1ff-7075f3f20091\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84a36b76f20618c7885d611ecad9db29bac5101c27572be31d3acbb90c581291\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0a0eca3c2087a457c0e448e649345ec4595a9c8a019aed2cddc5f30c77871e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://404426a171b8676eddd4a00996c76afab7e269643fa9608f1e1b6bdba385071d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54fd3c2941782921f99a6cdc3de4e892234548bde527fec15a8ac92e513566e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e23abfdca314ed43d594333c9f9e401cc8ef7dad39b8ebac84a836b06ee5c41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09b38cd4b389ea477c802a842ee4f2297a64d64684624117163e819539c257c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09b38cd4b389ea477c802a842ee4f2297a64d64684624117163e819539c257c6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c00f7a8e024e0e0dfd7b73e16d1d5437602e79ea95b73c21f24b209edef60e98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c00f7a8e024e0e0dfd7b73e16d1d5437602e79ea95b73c21f24b209edef60e98\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3f6392d426fa2cd788ec3a5c3bff4ac39483a54dfb2624a1c4370522413a350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3f6392d426fa2cd788ec3a5c3bff4ac39483a54dfb2624a1c4370522413a350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-07T16:02:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-07T16:02:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:41Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:41 crc kubenswrapper[4716]: I1207 16:03:41.421061 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3067de7c-c11c-4be1-bd94-d6572c266795\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d14d018fa044e9e80fe81e7e38df935992d14c773ed2b2193fd5cb873e5d0947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a9e9cf048220ca14c996847e9ced9f2367b9331dc6c2836dd023658855cd79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6d3c911874095b3c7cb062d03070e5090a7bb2456cb20cb4ce078f502515cda\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ac6ce7701f1c6aa1ea8fcf7a3035e3b0eaa28ae3cb6afc442372bd9117bdde4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-07T16:02:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:41Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:41 crc kubenswrapper[4716]: I1207 16:03:41.438303 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4742ea84b3a2c9461d9b900beee7354a2ded24c39ad23e007cbf270ae12377c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-07T16:02:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:41Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:41 crc kubenswrapper[4716]: I1207 16:03:41.451951 4716 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-07T16:02:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-07T16:03:41Z is after 2025-08-24T17:21:41Z" Dec 07 16:03:41 crc kubenswrapper[4716]: I1207 16:03:41.507971 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:41 crc kubenswrapper[4716]: I1207 16:03:41.508016 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:41 crc kubenswrapper[4716]: I1207 16:03:41.508031 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:41 crc kubenswrapper[4716]: I1207 16:03:41.508063 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:41 crc kubenswrapper[4716]: I1207 16:03:41.508103 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:41Z","lastTransitionTime":"2025-12-07T16:03:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:41 crc kubenswrapper[4716]: I1207 16:03:41.612741 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:41 crc kubenswrapper[4716]: I1207 16:03:41.612797 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:41 crc kubenswrapper[4716]: I1207 16:03:41.612813 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:41 crc kubenswrapper[4716]: I1207 16:03:41.612836 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:41 crc kubenswrapper[4716]: I1207 16:03:41.612851 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:41Z","lastTransitionTime":"2025-12-07T16:03:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:41 crc kubenswrapper[4716]: I1207 16:03:41.657549 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 16:03:41 crc kubenswrapper[4716]: I1207 16:03:41.657672 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nx2lh" Dec 07 16:03:41 crc kubenswrapper[4716]: E1207 16:03:41.657844 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 16:03:41 crc kubenswrapper[4716]: I1207 16:03:41.658181 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:03:41 crc kubenswrapper[4716]: E1207 16:03:41.658364 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nx2lh" podUID="04b37f6c-457b-4f9a-8733-e88cb9644483" Dec 07 16:03:41 crc kubenswrapper[4716]: E1207 16:03:41.658480 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 16:03:41 crc kubenswrapper[4716]: I1207 16:03:41.715689 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:41 crc kubenswrapper[4716]: I1207 16:03:41.715741 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:41 crc kubenswrapper[4716]: I1207 16:03:41.715758 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:41 crc kubenswrapper[4716]: I1207 16:03:41.715779 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:41 crc kubenswrapper[4716]: I1207 16:03:41.715796 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:41Z","lastTransitionTime":"2025-12-07T16:03:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:41 crc kubenswrapper[4716]: I1207 16:03:41.818718 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:41 crc kubenswrapper[4716]: I1207 16:03:41.818760 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:41 crc kubenswrapper[4716]: I1207 16:03:41.818776 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:41 crc kubenswrapper[4716]: I1207 16:03:41.818798 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:41 crc kubenswrapper[4716]: I1207 16:03:41.818814 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:41Z","lastTransitionTime":"2025-12-07T16:03:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:41 crc kubenswrapper[4716]: I1207 16:03:41.921330 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:41 crc kubenswrapper[4716]: I1207 16:03:41.921365 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:41 crc kubenswrapper[4716]: I1207 16:03:41.921374 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:41 crc kubenswrapper[4716]: I1207 16:03:41.921387 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:41 crc kubenswrapper[4716]: I1207 16:03:41.921396 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:41Z","lastTransitionTime":"2025-12-07T16:03:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:42 crc kubenswrapper[4716]: I1207 16:03:42.024446 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:42 crc kubenswrapper[4716]: I1207 16:03:42.024515 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:42 crc kubenswrapper[4716]: I1207 16:03:42.024538 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:42 crc kubenswrapper[4716]: I1207 16:03:42.024567 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:42 crc kubenswrapper[4716]: I1207 16:03:42.024589 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:42Z","lastTransitionTime":"2025-12-07T16:03:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:42 crc kubenswrapper[4716]: I1207 16:03:42.126932 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:42 crc kubenswrapper[4716]: I1207 16:03:42.126975 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:42 crc kubenswrapper[4716]: I1207 16:03:42.126984 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:42 crc kubenswrapper[4716]: I1207 16:03:42.126999 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:42 crc kubenswrapper[4716]: I1207 16:03:42.127010 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:42Z","lastTransitionTime":"2025-12-07T16:03:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:42 crc kubenswrapper[4716]: I1207 16:03:42.228667 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:42 crc kubenswrapper[4716]: I1207 16:03:42.228714 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:42 crc kubenswrapper[4716]: I1207 16:03:42.228726 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:42 crc kubenswrapper[4716]: I1207 16:03:42.228743 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:42 crc kubenswrapper[4716]: I1207 16:03:42.228756 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:42Z","lastTransitionTime":"2025-12-07T16:03:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:42 crc kubenswrapper[4716]: I1207 16:03:42.331306 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:42 crc kubenswrapper[4716]: I1207 16:03:42.331415 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:42 crc kubenswrapper[4716]: I1207 16:03:42.331437 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:42 crc kubenswrapper[4716]: I1207 16:03:42.331461 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:42 crc kubenswrapper[4716]: I1207 16:03:42.331479 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:42Z","lastTransitionTime":"2025-12-07T16:03:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:42 crc kubenswrapper[4716]: I1207 16:03:42.434902 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:42 crc kubenswrapper[4716]: I1207 16:03:42.434945 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:42 crc kubenswrapper[4716]: I1207 16:03:42.434957 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:42 crc kubenswrapper[4716]: I1207 16:03:42.434979 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:42 crc kubenswrapper[4716]: I1207 16:03:42.434992 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:42Z","lastTransitionTime":"2025-12-07T16:03:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:42 crc kubenswrapper[4716]: I1207 16:03:42.536513 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:42 crc kubenswrapper[4716]: I1207 16:03:42.536545 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:42 crc kubenswrapper[4716]: I1207 16:03:42.536553 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:42 crc kubenswrapper[4716]: I1207 16:03:42.536565 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:42 crc kubenswrapper[4716]: I1207 16:03:42.536576 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:42Z","lastTransitionTime":"2025-12-07T16:03:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:42 crc kubenswrapper[4716]: I1207 16:03:42.638489 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:42 crc kubenswrapper[4716]: I1207 16:03:42.638530 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:42 crc kubenswrapper[4716]: I1207 16:03:42.638541 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:42 crc kubenswrapper[4716]: I1207 16:03:42.638555 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:42 crc kubenswrapper[4716]: I1207 16:03:42.638564 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:42Z","lastTransitionTime":"2025-12-07T16:03:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:42 crc kubenswrapper[4716]: I1207 16:03:42.656843 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 16:03:42 crc kubenswrapper[4716]: E1207 16:03:42.657182 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 16:03:42 crc kubenswrapper[4716]: I1207 16:03:42.740531 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:42 crc kubenswrapper[4716]: I1207 16:03:42.740795 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:42 crc kubenswrapper[4716]: I1207 16:03:42.740809 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:42 crc kubenswrapper[4716]: I1207 16:03:42.740825 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:42 crc kubenswrapper[4716]: I1207 16:03:42.740837 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:42Z","lastTransitionTime":"2025-12-07T16:03:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:42 crc kubenswrapper[4716]: I1207 16:03:42.844029 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:42 crc kubenswrapper[4716]: I1207 16:03:42.844062 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:42 crc kubenswrapper[4716]: I1207 16:03:42.844070 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:42 crc kubenswrapper[4716]: I1207 16:03:42.844099 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:42 crc kubenswrapper[4716]: I1207 16:03:42.844108 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:42Z","lastTransitionTime":"2025-12-07T16:03:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:42 crc kubenswrapper[4716]: I1207 16:03:42.947239 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:42 crc kubenswrapper[4716]: I1207 16:03:42.947321 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:42 crc kubenswrapper[4716]: I1207 16:03:42.947338 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:42 crc kubenswrapper[4716]: I1207 16:03:42.947361 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:42 crc kubenswrapper[4716]: I1207 16:03:42.947378 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:42Z","lastTransitionTime":"2025-12-07T16:03:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:43 crc kubenswrapper[4716]: I1207 16:03:43.049194 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:43 crc kubenswrapper[4716]: I1207 16:03:43.049233 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:43 crc kubenswrapper[4716]: I1207 16:03:43.049241 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:43 crc kubenswrapper[4716]: I1207 16:03:43.049257 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:43 crc kubenswrapper[4716]: I1207 16:03:43.049267 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:43Z","lastTransitionTime":"2025-12-07T16:03:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:43 crc kubenswrapper[4716]: I1207 16:03:43.151328 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:43 crc kubenswrapper[4716]: I1207 16:03:43.151355 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:43 crc kubenswrapper[4716]: I1207 16:03:43.151363 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:43 crc kubenswrapper[4716]: I1207 16:03:43.151374 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:43 crc kubenswrapper[4716]: I1207 16:03:43.151383 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:43Z","lastTransitionTime":"2025-12-07T16:03:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:43 crc kubenswrapper[4716]: I1207 16:03:43.253276 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:43 crc kubenswrapper[4716]: I1207 16:03:43.253302 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:43 crc kubenswrapper[4716]: I1207 16:03:43.253310 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:43 crc kubenswrapper[4716]: I1207 16:03:43.253321 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:43 crc kubenswrapper[4716]: I1207 16:03:43.253328 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:43Z","lastTransitionTime":"2025-12-07T16:03:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:43 crc kubenswrapper[4716]: I1207 16:03:43.355867 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:43 crc kubenswrapper[4716]: I1207 16:03:43.355902 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:43 crc kubenswrapper[4716]: I1207 16:03:43.355913 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:43 crc kubenswrapper[4716]: I1207 16:03:43.355929 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:43 crc kubenswrapper[4716]: I1207 16:03:43.355940 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:43Z","lastTransitionTime":"2025-12-07T16:03:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:43 crc kubenswrapper[4716]: I1207 16:03:43.459348 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:43 crc kubenswrapper[4716]: I1207 16:03:43.459386 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:43 crc kubenswrapper[4716]: I1207 16:03:43.459397 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:43 crc kubenswrapper[4716]: I1207 16:03:43.459427 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:43 crc kubenswrapper[4716]: I1207 16:03:43.459438 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:43Z","lastTransitionTime":"2025-12-07T16:03:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:43 crc kubenswrapper[4716]: I1207 16:03:43.563169 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:43 crc kubenswrapper[4716]: I1207 16:03:43.563241 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:43 crc kubenswrapper[4716]: I1207 16:03:43.563263 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:43 crc kubenswrapper[4716]: I1207 16:03:43.563297 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:43 crc kubenswrapper[4716]: I1207 16:03:43.563320 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:43Z","lastTransitionTime":"2025-12-07T16:03:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:43 crc kubenswrapper[4716]: I1207 16:03:43.657294 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 16:03:43 crc kubenswrapper[4716]: I1207 16:03:43.657358 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:03:43 crc kubenswrapper[4716]: I1207 16:03:43.657364 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nx2lh" Dec 07 16:03:43 crc kubenswrapper[4716]: E1207 16:03:43.657440 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 16:03:43 crc kubenswrapper[4716]: E1207 16:03:43.657570 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 16:03:43 crc kubenswrapper[4716]: E1207 16:03:43.657745 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nx2lh" podUID="04b37f6c-457b-4f9a-8733-e88cb9644483" Dec 07 16:03:43 crc kubenswrapper[4716]: I1207 16:03:43.666767 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:43 crc kubenswrapper[4716]: I1207 16:03:43.666810 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:43 crc kubenswrapper[4716]: I1207 16:03:43.666823 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:43 crc kubenswrapper[4716]: I1207 16:03:43.666839 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:43 crc kubenswrapper[4716]: I1207 16:03:43.666853 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:43Z","lastTransitionTime":"2025-12-07T16:03:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:43 crc kubenswrapper[4716]: I1207 16:03:43.769923 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:43 crc kubenswrapper[4716]: I1207 16:03:43.769987 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:43 crc kubenswrapper[4716]: I1207 16:03:43.770005 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:43 crc kubenswrapper[4716]: I1207 16:03:43.770029 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:43 crc kubenswrapper[4716]: I1207 16:03:43.770045 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:43Z","lastTransitionTime":"2025-12-07T16:03:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:43 crc kubenswrapper[4716]: I1207 16:03:43.873741 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:43 crc kubenswrapper[4716]: I1207 16:03:43.873873 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:43 crc kubenswrapper[4716]: I1207 16:03:43.873892 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:43 crc kubenswrapper[4716]: I1207 16:03:43.873915 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:43 crc kubenswrapper[4716]: I1207 16:03:43.873933 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:43Z","lastTransitionTime":"2025-12-07T16:03:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:43 crc kubenswrapper[4716]: I1207 16:03:43.977247 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:43 crc kubenswrapper[4716]: I1207 16:03:43.977291 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:43 crc kubenswrapper[4716]: I1207 16:03:43.977303 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:43 crc kubenswrapper[4716]: I1207 16:03:43.977318 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:43 crc kubenswrapper[4716]: I1207 16:03:43.977328 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:43Z","lastTransitionTime":"2025-12-07T16:03:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:44 crc kubenswrapper[4716]: I1207 16:03:44.080057 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:44 crc kubenswrapper[4716]: I1207 16:03:44.080169 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:44 crc kubenswrapper[4716]: I1207 16:03:44.080193 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:44 crc kubenswrapper[4716]: I1207 16:03:44.080222 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:44 crc kubenswrapper[4716]: I1207 16:03:44.080244 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:44Z","lastTransitionTime":"2025-12-07T16:03:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:44 crc kubenswrapper[4716]: I1207 16:03:44.182949 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:44 crc kubenswrapper[4716]: I1207 16:03:44.183027 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:44 crc kubenswrapper[4716]: I1207 16:03:44.183050 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:44 crc kubenswrapper[4716]: I1207 16:03:44.183116 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:44 crc kubenswrapper[4716]: I1207 16:03:44.183143 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:44Z","lastTransitionTime":"2025-12-07T16:03:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:44 crc kubenswrapper[4716]: I1207 16:03:44.285693 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:44 crc kubenswrapper[4716]: I1207 16:03:44.285768 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:44 crc kubenswrapper[4716]: I1207 16:03:44.285792 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:44 crc kubenswrapper[4716]: I1207 16:03:44.285823 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:44 crc kubenswrapper[4716]: I1207 16:03:44.285847 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:44Z","lastTransitionTime":"2025-12-07T16:03:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:44 crc kubenswrapper[4716]: I1207 16:03:44.389049 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:44 crc kubenswrapper[4716]: I1207 16:03:44.389171 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:44 crc kubenswrapper[4716]: I1207 16:03:44.389192 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:44 crc kubenswrapper[4716]: I1207 16:03:44.389216 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:44 crc kubenswrapper[4716]: I1207 16:03:44.389233 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:44Z","lastTransitionTime":"2025-12-07T16:03:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:44 crc kubenswrapper[4716]: I1207 16:03:44.492069 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:44 crc kubenswrapper[4716]: I1207 16:03:44.492170 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:44 crc kubenswrapper[4716]: I1207 16:03:44.492187 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:44 crc kubenswrapper[4716]: I1207 16:03:44.492211 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:44 crc kubenswrapper[4716]: I1207 16:03:44.492228 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:44Z","lastTransitionTime":"2025-12-07T16:03:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:44 crc kubenswrapper[4716]: I1207 16:03:44.595014 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:44 crc kubenswrapper[4716]: I1207 16:03:44.595130 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:44 crc kubenswrapper[4716]: I1207 16:03:44.595152 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:44 crc kubenswrapper[4716]: I1207 16:03:44.595180 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:44 crc kubenswrapper[4716]: I1207 16:03:44.595200 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:44Z","lastTransitionTime":"2025-12-07T16:03:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:44 crc kubenswrapper[4716]: I1207 16:03:44.656878 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 16:03:44 crc kubenswrapper[4716]: E1207 16:03:44.657067 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 16:03:44 crc kubenswrapper[4716]: I1207 16:03:44.698187 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:44 crc kubenswrapper[4716]: I1207 16:03:44.698243 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:44 crc kubenswrapper[4716]: I1207 16:03:44.698264 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:44 crc kubenswrapper[4716]: I1207 16:03:44.698286 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:44 crc kubenswrapper[4716]: I1207 16:03:44.698304 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:44Z","lastTransitionTime":"2025-12-07T16:03:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:44 crc kubenswrapper[4716]: I1207 16:03:44.800882 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:44 crc kubenswrapper[4716]: I1207 16:03:44.800961 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:44 crc kubenswrapper[4716]: I1207 16:03:44.800980 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:44 crc kubenswrapper[4716]: I1207 16:03:44.801005 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:44 crc kubenswrapper[4716]: I1207 16:03:44.801022 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:44Z","lastTransitionTime":"2025-12-07T16:03:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:44 crc kubenswrapper[4716]: I1207 16:03:44.903667 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:44 crc kubenswrapper[4716]: I1207 16:03:44.903706 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:44 crc kubenswrapper[4716]: I1207 16:03:44.903715 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:44 crc kubenswrapper[4716]: I1207 16:03:44.903730 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:44 crc kubenswrapper[4716]: I1207 16:03:44.903740 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:44Z","lastTransitionTime":"2025-12-07T16:03:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:45 crc kubenswrapper[4716]: I1207 16:03:45.005725 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:45 crc kubenswrapper[4716]: I1207 16:03:45.005770 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:45 crc kubenswrapper[4716]: I1207 16:03:45.005783 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:45 crc kubenswrapper[4716]: I1207 16:03:45.005800 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:45 crc kubenswrapper[4716]: I1207 16:03:45.005812 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:45Z","lastTransitionTime":"2025-12-07T16:03:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:45 crc kubenswrapper[4716]: I1207 16:03:45.108017 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:45 crc kubenswrapper[4716]: I1207 16:03:45.108111 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:45 crc kubenswrapper[4716]: I1207 16:03:45.108131 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:45 crc kubenswrapper[4716]: I1207 16:03:45.108154 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:45 crc kubenswrapper[4716]: I1207 16:03:45.108172 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:45Z","lastTransitionTime":"2025-12-07T16:03:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:45 crc kubenswrapper[4716]: I1207 16:03:45.211375 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:45 crc kubenswrapper[4716]: I1207 16:03:45.211440 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:45 crc kubenswrapper[4716]: I1207 16:03:45.211460 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:45 crc kubenswrapper[4716]: I1207 16:03:45.211485 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:45 crc kubenswrapper[4716]: I1207 16:03:45.211502 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:45Z","lastTransitionTime":"2025-12-07T16:03:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:45 crc kubenswrapper[4716]: I1207 16:03:45.314751 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:45 crc kubenswrapper[4716]: I1207 16:03:45.314829 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:45 crc kubenswrapper[4716]: I1207 16:03:45.314853 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:45 crc kubenswrapper[4716]: I1207 16:03:45.314886 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:45 crc kubenswrapper[4716]: I1207 16:03:45.314912 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:45Z","lastTransitionTime":"2025-12-07T16:03:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:45 crc kubenswrapper[4716]: I1207 16:03:45.418296 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:45 crc kubenswrapper[4716]: I1207 16:03:45.418365 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:45 crc kubenswrapper[4716]: I1207 16:03:45.418388 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:45 crc kubenswrapper[4716]: I1207 16:03:45.418420 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:45 crc kubenswrapper[4716]: I1207 16:03:45.418445 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:45Z","lastTransitionTime":"2025-12-07T16:03:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:45 crc kubenswrapper[4716]: I1207 16:03:45.521760 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:45 crc kubenswrapper[4716]: I1207 16:03:45.521859 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:45 crc kubenswrapper[4716]: I1207 16:03:45.521878 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:45 crc kubenswrapper[4716]: I1207 16:03:45.521902 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:45 crc kubenswrapper[4716]: I1207 16:03:45.521919 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:45Z","lastTransitionTime":"2025-12-07T16:03:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:45 crc kubenswrapper[4716]: I1207 16:03:45.625260 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:45 crc kubenswrapper[4716]: I1207 16:03:45.625294 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:45 crc kubenswrapper[4716]: I1207 16:03:45.625305 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:45 crc kubenswrapper[4716]: I1207 16:03:45.625321 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:45 crc kubenswrapper[4716]: I1207 16:03:45.625332 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:45Z","lastTransitionTime":"2025-12-07T16:03:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:45 crc kubenswrapper[4716]: I1207 16:03:45.656644 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:03:45 crc kubenswrapper[4716]: I1207 16:03:45.656710 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 16:03:45 crc kubenswrapper[4716]: E1207 16:03:45.656789 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 16:03:45 crc kubenswrapper[4716]: I1207 16:03:45.656885 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nx2lh" Dec 07 16:03:45 crc kubenswrapper[4716]: E1207 16:03:45.657013 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 16:03:45 crc kubenswrapper[4716]: E1207 16:03:45.657273 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nx2lh" podUID="04b37f6c-457b-4f9a-8733-e88cb9644483" Dec 07 16:03:45 crc kubenswrapper[4716]: I1207 16:03:45.728905 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:45 crc kubenswrapper[4716]: I1207 16:03:45.728969 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:45 crc kubenswrapper[4716]: I1207 16:03:45.728989 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:45 crc kubenswrapper[4716]: I1207 16:03:45.729015 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:45 crc kubenswrapper[4716]: I1207 16:03:45.729033 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:45Z","lastTransitionTime":"2025-12-07T16:03:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:45 crc kubenswrapper[4716]: I1207 16:03:45.831595 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:45 crc kubenswrapper[4716]: I1207 16:03:45.831640 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:45 crc kubenswrapper[4716]: I1207 16:03:45.831654 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:45 crc kubenswrapper[4716]: I1207 16:03:45.831673 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:45 crc kubenswrapper[4716]: I1207 16:03:45.831692 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:45Z","lastTransitionTime":"2025-12-07T16:03:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:45 crc kubenswrapper[4716]: I1207 16:03:45.935420 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:45 crc kubenswrapper[4716]: I1207 16:03:45.935499 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:45 crc kubenswrapper[4716]: I1207 16:03:45.935521 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:45 crc kubenswrapper[4716]: I1207 16:03:45.935552 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:45 crc kubenswrapper[4716]: I1207 16:03:45.935574 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:45Z","lastTransitionTime":"2025-12-07T16:03:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:46 crc kubenswrapper[4716]: I1207 16:03:46.038827 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:46 crc kubenswrapper[4716]: I1207 16:03:46.038898 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:46 crc kubenswrapper[4716]: I1207 16:03:46.038915 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:46 crc kubenswrapper[4716]: I1207 16:03:46.038944 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:46 crc kubenswrapper[4716]: I1207 16:03:46.038962 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:46Z","lastTransitionTime":"2025-12-07T16:03:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:46 crc kubenswrapper[4716]: I1207 16:03:46.141806 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:46 crc kubenswrapper[4716]: I1207 16:03:46.141931 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:46 crc kubenswrapper[4716]: I1207 16:03:46.141956 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:46 crc kubenswrapper[4716]: I1207 16:03:46.141990 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:46 crc kubenswrapper[4716]: I1207 16:03:46.142015 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:46Z","lastTransitionTime":"2025-12-07T16:03:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:46 crc kubenswrapper[4716]: I1207 16:03:46.250595 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:46 crc kubenswrapper[4716]: I1207 16:03:46.250680 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:46 crc kubenswrapper[4716]: I1207 16:03:46.250702 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:46 crc kubenswrapper[4716]: I1207 16:03:46.250728 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:46 crc kubenswrapper[4716]: I1207 16:03:46.250746 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:46Z","lastTransitionTime":"2025-12-07T16:03:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:46 crc kubenswrapper[4716]: I1207 16:03:46.353725 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:46 crc kubenswrapper[4716]: I1207 16:03:46.353770 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:46 crc kubenswrapper[4716]: I1207 16:03:46.353780 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:46 crc kubenswrapper[4716]: I1207 16:03:46.353795 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:46 crc kubenswrapper[4716]: I1207 16:03:46.353805 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:46Z","lastTransitionTime":"2025-12-07T16:03:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:46 crc kubenswrapper[4716]: I1207 16:03:46.456352 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:46 crc kubenswrapper[4716]: I1207 16:03:46.456397 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:46 crc kubenswrapper[4716]: I1207 16:03:46.456409 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:46 crc kubenswrapper[4716]: I1207 16:03:46.456426 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:46 crc kubenswrapper[4716]: I1207 16:03:46.456439 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:46Z","lastTransitionTime":"2025-12-07T16:03:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:46 crc kubenswrapper[4716]: I1207 16:03:46.558799 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:46 crc kubenswrapper[4716]: I1207 16:03:46.558882 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:46 crc kubenswrapper[4716]: I1207 16:03:46.558908 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:46 crc kubenswrapper[4716]: I1207 16:03:46.558940 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:46 crc kubenswrapper[4716]: I1207 16:03:46.558986 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:46Z","lastTransitionTime":"2025-12-07T16:03:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:46 crc kubenswrapper[4716]: I1207 16:03:46.656855 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 16:03:46 crc kubenswrapper[4716]: E1207 16:03:46.656965 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 16:03:46 crc kubenswrapper[4716]: I1207 16:03:46.661480 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:46 crc kubenswrapper[4716]: I1207 16:03:46.661525 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:46 crc kubenswrapper[4716]: I1207 16:03:46.661536 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:46 crc kubenswrapper[4716]: I1207 16:03:46.661551 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:46 crc kubenswrapper[4716]: I1207 16:03:46.661560 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:46Z","lastTransitionTime":"2025-12-07T16:03:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:46 crc kubenswrapper[4716]: I1207 16:03:46.764949 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:46 crc kubenswrapper[4716]: I1207 16:03:46.764979 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:46 crc kubenswrapper[4716]: I1207 16:03:46.764988 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:46 crc kubenswrapper[4716]: I1207 16:03:46.765001 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:46 crc kubenswrapper[4716]: I1207 16:03:46.765010 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:46Z","lastTransitionTime":"2025-12-07T16:03:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:46 crc kubenswrapper[4716]: I1207 16:03:46.867919 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:46 crc kubenswrapper[4716]: I1207 16:03:46.867965 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:46 crc kubenswrapper[4716]: I1207 16:03:46.867976 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:46 crc kubenswrapper[4716]: I1207 16:03:46.867994 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:46 crc kubenswrapper[4716]: I1207 16:03:46.868006 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:46Z","lastTransitionTime":"2025-12-07T16:03:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:46 crc kubenswrapper[4716]: I1207 16:03:46.970532 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:46 crc kubenswrapper[4716]: I1207 16:03:46.970587 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:46 crc kubenswrapper[4716]: I1207 16:03:46.970600 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:46 crc kubenswrapper[4716]: I1207 16:03:46.970619 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:46 crc kubenswrapper[4716]: I1207 16:03:46.970633 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:46Z","lastTransitionTime":"2025-12-07T16:03:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:47 crc kubenswrapper[4716]: I1207 16:03:47.073091 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:47 crc kubenswrapper[4716]: I1207 16:03:47.073123 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:47 crc kubenswrapper[4716]: I1207 16:03:47.073132 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:47 crc kubenswrapper[4716]: I1207 16:03:47.073146 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:47 crc kubenswrapper[4716]: I1207 16:03:47.073155 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:47Z","lastTransitionTime":"2025-12-07T16:03:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:47 crc kubenswrapper[4716]: I1207 16:03:47.175308 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:47 crc kubenswrapper[4716]: I1207 16:03:47.175351 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:47 crc kubenswrapper[4716]: I1207 16:03:47.175362 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:47 crc kubenswrapper[4716]: I1207 16:03:47.175377 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:47 crc kubenswrapper[4716]: I1207 16:03:47.175390 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:47Z","lastTransitionTime":"2025-12-07T16:03:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:47 crc kubenswrapper[4716]: I1207 16:03:47.278297 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:47 crc kubenswrapper[4716]: I1207 16:03:47.278333 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:47 crc kubenswrapper[4716]: I1207 16:03:47.278344 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:47 crc kubenswrapper[4716]: I1207 16:03:47.278359 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:47 crc kubenswrapper[4716]: I1207 16:03:47.278373 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:47Z","lastTransitionTime":"2025-12-07T16:03:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:47 crc kubenswrapper[4716]: I1207 16:03:47.381206 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:47 crc kubenswrapper[4716]: I1207 16:03:47.381254 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:47 crc kubenswrapper[4716]: I1207 16:03:47.381268 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:47 crc kubenswrapper[4716]: I1207 16:03:47.381285 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:47 crc kubenswrapper[4716]: I1207 16:03:47.381297 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:47Z","lastTransitionTime":"2025-12-07T16:03:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:47 crc kubenswrapper[4716]: I1207 16:03:47.484588 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:47 crc kubenswrapper[4716]: I1207 16:03:47.484641 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:47 crc kubenswrapper[4716]: I1207 16:03:47.484657 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:47 crc kubenswrapper[4716]: I1207 16:03:47.484683 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:47 crc kubenswrapper[4716]: I1207 16:03:47.484702 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:47Z","lastTransitionTime":"2025-12-07T16:03:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:47 crc kubenswrapper[4716]: I1207 16:03:47.587185 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:47 crc kubenswrapper[4716]: I1207 16:03:47.587258 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:47 crc kubenswrapper[4716]: I1207 16:03:47.587283 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:47 crc kubenswrapper[4716]: I1207 16:03:47.587314 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:47 crc kubenswrapper[4716]: I1207 16:03:47.587338 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:47Z","lastTransitionTime":"2025-12-07T16:03:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:47 crc kubenswrapper[4716]: I1207 16:03:47.657347 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 16:03:47 crc kubenswrapper[4716]: I1207 16:03:47.657378 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nx2lh" Dec 07 16:03:47 crc kubenswrapper[4716]: I1207 16:03:47.657378 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:03:47 crc kubenswrapper[4716]: E1207 16:03:47.657609 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 16:03:47 crc kubenswrapper[4716]: E1207 16:03:47.657965 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nx2lh" podUID="04b37f6c-457b-4f9a-8733-e88cb9644483" Dec 07 16:03:47 crc kubenswrapper[4716]: E1207 16:03:47.658105 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 16:03:47 crc kubenswrapper[4716]: I1207 16:03:47.689862 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:47 crc kubenswrapper[4716]: I1207 16:03:47.689928 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:47 crc kubenswrapper[4716]: I1207 16:03:47.689949 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:47 crc kubenswrapper[4716]: I1207 16:03:47.689975 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:47 crc kubenswrapper[4716]: I1207 16:03:47.689993 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:47Z","lastTransitionTime":"2025-12-07T16:03:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:47 crc kubenswrapper[4716]: I1207 16:03:47.700514 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=13.700495039 podStartE2EDuration="13.700495039s" podCreationTimestamp="2025-12-07 16:03:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:03:47.684042185 +0000 UTC m=+90.374327107" watchObservedRunningTime="2025-12-07 16:03:47.700495039 +0000 UTC m=+90.390779961" Dec 07 16:03:47 crc kubenswrapper[4716]: I1207 16:03:47.769816 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:47 crc kubenswrapper[4716]: I1207 16:03:47.769870 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:47 crc kubenswrapper[4716]: I1207 16:03:47.769887 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:47 crc kubenswrapper[4716]: I1207 16:03:47.769913 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:47 crc kubenswrapper[4716]: I1207 16:03:47.769934 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:47Z","lastTransitionTime":"2025-12-07T16:03:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:47 crc kubenswrapper[4716]: I1207 16:03:47.781392 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=67.78136985 podStartE2EDuration="1m7.78136985s" podCreationTimestamp="2025-12-07 16:02:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:03:47.779835456 +0000 UTC m=+90.470120408" watchObservedRunningTime="2025-12-07 16:03:47.78136985 +0000 UTC m=+90.471654772" Dec 07 16:03:47 crc kubenswrapper[4716]: I1207 16:03:47.792824 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 07 16:03:47 crc kubenswrapper[4716]: I1207 16:03:47.792859 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 07 16:03:47 crc kubenswrapper[4716]: I1207 16:03:47.792869 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 07 16:03:47 crc kubenswrapper[4716]: I1207 16:03:47.792887 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 07 16:03:47 crc kubenswrapper[4716]: I1207 16:03:47.792898 4716 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-07T16:03:47Z","lastTransitionTime":"2025-12-07T16:03:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 07 16:03:47 crc kubenswrapper[4716]: I1207 16:03:47.807008 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=67.806986988 podStartE2EDuration="1m7.806986988s" podCreationTimestamp="2025-12-07 16:02:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:03:47.801629794 +0000 UTC m=+90.491914746" watchObservedRunningTime="2025-12-07 16:03:47.806986988 +0000 UTC m=+90.497271900" Dec 07 16:03:47 crc kubenswrapper[4716]: I1207 16:03:47.821887 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-kr2j6"] Dec 07 16:03:47 crc kubenswrapper[4716]: I1207 16:03:47.822258 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kr2j6" Dec 07 16:03:47 crc kubenswrapper[4716]: I1207 16:03:47.824479 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 07 16:03:47 crc kubenswrapper[4716]: I1207 16:03:47.824579 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 07 16:03:47 crc kubenswrapper[4716]: I1207 16:03:47.824663 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 07 16:03:47 crc kubenswrapper[4716]: I1207 16:03:47.824849 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 07 16:03:47 crc kubenswrapper[4716]: I1207 16:03:47.853566 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podStartSLOduration=68.85354076 podStartE2EDuration="1m8.85354076s" podCreationTimestamp="2025-12-07 16:02:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:03:47.853152789 +0000 UTC m=+90.543437721" watchObservedRunningTime="2025-12-07 16:03:47.85354076 +0000 UTC m=+90.543825712" Dec 07 16:03:47 crc kubenswrapper[4716]: I1207 16:03:47.874155 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-8ps9j" podStartSLOduration=68.874128203 podStartE2EDuration="1m8.874128203s" podCreationTimestamp="2025-12-07 16:02:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:03:47.873920847 +0000 UTC m=+90.564205799" watchObservedRunningTime="2025-12-07 16:03:47.874128203 +0000 UTC m=+90.564413165" Dec 07 16:03:47 crc kubenswrapper[4716]: I1207 16:03:47.880937 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1cbaca95-7667-4abd-b5c9-8468782263b7-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-kr2j6\" (UID: \"1cbaca95-7667-4abd-b5c9-8468782263b7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kr2j6" Dec 07 16:03:47 crc kubenswrapper[4716]: I1207 16:03:47.880994 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/1cbaca95-7667-4abd-b5c9-8468782263b7-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-kr2j6\" (UID: \"1cbaca95-7667-4abd-b5c9-8468782263b7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kr2j6" Dec 07 16:03:47 crc kubenswrapper[4716]: I1207 16:03:47.881018 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1cbaca95-7667-4abd-b5c9-8468782263b7-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-kr2j6\" (UID: \"1cbaca95-7667-4abd-b5c9-8468782263b7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kr2j6" Dec 07 16:03:47 crc kubenswrapper[4716]: I1207 16:03:47.881067 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/1cbaca95-7667-4abd-b5c9-8468782263b7-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-kr2j6\" (UID: \"1cbaca95-7667-4abd-b5c9-8468782263b7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kr2j6" Dec 07 16:03:47 crc kubenswrapper[4716]: I1207 16:03:47.881139 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/1cbaca95-7667-4abd-b5c9-8468782263b7-service-ca\") pod \"cluster-version-operator-5c965bbfc6-kr2j6\" (UID: \"1cbaca95-7667-4abd-b5c9-8468782263b7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kr2j6" Dec 07 16:03:47 crc kubenswrapper[4716]: I1207 16:03:47.891906 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=43.891881085 podStartE2EDuration="43.891881085s" podCreationTimestamp="2025-12-07 16:03:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:03:47.891494963 +0000 UTC m=+90.581779915" watchObservedRunningTime="2025-12-07 16:03:47.891881085 +0000 UTC m=+90.582166037" Dec 07 16:03:47 crc kubenswrapper[4716]: I1207 16:03:47.962913 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mx85f" podStartSLOduration=67.962888022 podStartE2EDuration="1m7.962888022s" podCreationTimestamp="2025-12-07 16:02:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:03:47.951637507 +0000 UTC m=+90.641922459" watchObservedRunningTime="2025-12-07 16:03:47.962888022 +0000 UTC m=+90.653172944" Dec 07 16:03:47 crc kubenswrapper[4716]: I1207 16:03:47.978955 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=72.978934744 podStartE2EDuration="1m12.978934744s" podCreationTimestamp="2025-12-07 16:02:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:03:47.978054688 +0000 UTC m=+90.668339610" watchObservedRunningTime="2025-12-07 16:03:47.978934744 +0000 UTC m=+90.669219666" Dec 07 16:03:47 crc kubenswrapper[4716]: I1207 16:03:47.981953 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/1cbaca95-7667-4abd-b5c9-8468782263b7-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-kr2j6\" (UID: \"1cbaca95-7667-4abd-b5c9-8468782263b7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kr2j6" Dec 07 16:03:47 crc kubenswrapper[4716]: I1207 16:03:47.982037 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1cbaca95-7667-4abd-b5c9-8468782263b7-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-kr2j6\" (UID: \"1cbaca95-7667-4abd-b5c9-8468782263b7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kr2j6" Dec 07 16:03:47 crc kubenswrapper[4716]: I1207 16:03:47.982065 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/1cbaca95-7667-4abd-b5c9-8468782263b7-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-kr2j6\" (UID: \"1cbaca95-7667-4abd-b5c9-8468782263b7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kr2j6" Dec 07 16:03:47 crc kubenswrapper[4716]: I1207 16:03:47.982193 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/1cbaca95-7667-4abd-b5c9-8468782263b7-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-kr2j6\" (UID: \"1cbaca95-7667-4abd-b5c9-8468782263b7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kr2j6" Dec 07 16:03:47 crc kubenswrapper[4716]: I1207 16:03:47.982281 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/1cbaca95-7667-4abd-b5c9-8468782263b7-service-ca\") pod \"cluster-version-operator-5c965bbfc6-kr2j6\" (UID: \"1cbaca95-7667-4abd-b5c9-8468782263b7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kr2j6" Dec 07 16:03:47 crc kubenswrapper[4716]: I1207 16:03:47.982349 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/1cbaca95-7667-4abd-b5c9-8468782263b7-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-kr2j6\" (UID: \"1cbaca95-7667-4abd-b5c9-8468782263b7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kr2j6" Dec 07 16:03:47 crc kubenswrapper[4716]: I1207 16:03:47.982355 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1cbaca95-7667-4abd-b5c9-8468782263b7-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-kr2j6\" (UID: \"1cbaca95-7667-4abd-b5c9-8468782263b7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kr2j6" Dec 07 16:03:47 crc kubenswrapper[4716]: I1207 16:03:47.983050 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/1cbaca95-7667-4abd-b5c9-8468782263b7-service-ca\") pod \"cluster-version-operator-5c965bbfc6-kr2j6\" (UID: \"1cbaca95-7667-4abd-b5c9-8468782263b7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kr2j6" Dec 07 16:03:47 crc kubenswrapper[4716]: I1207 16:03:47.989286 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1cbaca95-7667-4abd-b5c9-8468782263b7-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-kr2j6\" (UID: \"1cbaca95-7667-4abd-b5c9-8468782263b7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kr2j6" Dec 07 16:03:48 crc kubenswrapper[4716]: I1207 16:03:48.000661 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1cbaca95-7667-4abd-b5c9-8468782263b7-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-kr2j6\" (UID: \"1cbaca95-7667-4abd-b5c9-8468782263b7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kr2j6" Dec 07 16:03:48 crc kubenswrapper[4716]: I1207 16:03:48.011144 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-2jjld" podStartSLOduration=69.011129742 podStartE2EDuration="1m9.011129742s" podCreationTimestamp="2025-12-07 16:02:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:03:48.010681929 +0000 UTC m=+90.700966851" watchObservedRunningTime="2025-12-07 16:03:48.011129742 +0000 UTC m=+90.701414654" Dec 07 16:03:48 crc kubenswrapper[4716]: I1207 16:03:48.023357 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-k4wr2" podStartSLOduration=69.023336143 podStartE2EDuration="1m9.023336143s" podCreationTimestamp="2025-12-07 16:02:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:03:48.023112997 +0000 UTC m=+90.713397909" watchObservedRunningTime="2025-12-07 16:03:48.023336143 +0000 UTC m=+90.713621065" Dec 07 16:03:48 crc kubenswrapper[4716]: I1207 16:03:48.043140 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-m5jl5" podStartSLOduration=69.043119624 podStartE2EDuration="1m9.043119624s" podCreationTimestamp="2025-12-07 16:02:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:03:48.03255525 +0000 UTC m=+90.722840162" watchObservedRunningTime="2025-12-07 16:03:48.043119624 +0000 UTC m=+90.733404546" Dec 07 16:03:48 crc kubenswrapper[4716]: I1207 16:03:48.141975 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kr2j6" Dec 07 16:03:48 crc kubenswrapper[4716]: I1207 16:03:48.166197 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kr2j6" event={"ID":"1cbaca95-7667-4abd-b5c9-8468782263b7","Type":"ContainerStarted","Data":"29ce433ca3f922b7471c08bccef91c138f57298dcc14dcb3ab0f47d5381af61d"} Dec 07 16:03:48 crc kubenswrapper[4716]: I1207 16:03:48.657122 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 16:03:48 crc kubenswrapper[4716]: E1207 16:03:48.657858 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 16:03:49 crc kubenswrapper[4716]: I1207 16:03:49.170384 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kr2j6" event={"ID":"1cbaca95-7667-4abd-b5c9-8468782263b7","Type":"ContainerStarted","Data":"296bad7f231297ddcc6d40fd47064dfeb996d0d859046e3037d18a3d44614c68"} Dec 07 16:03:49 crc kubenswrapper[4716]: I1207 16:03:49.183977 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kr2j6" podStartSLOduration=70.183954862 podStartE2EDuration="1m10.183954862s" podCreationTimestamp="2025-12-07 16:02:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:03:49.182463919 +0000 UTC m=+91.872748871" watchObservedRunningTime="2025-12-07 16:03:49.183954862 +0000 UTC m=+91.874239794" Dec 07 16:03:49 crc kubenswrapper[4716]: I1207 16:03:49.657551 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 16:03:49 crc kubenswrapper[4716]: I1207 16:03:49.657627 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nx2lh" Dec 07 16:03:49 crc kubenswrapper[4716]: E1207 16:03:49.657688 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 16:03:49 crc kubenswrapper[4716]: E1207 16:03:49.657846 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nx2lh" podUID="04b37f6c-457b-4f9a-8733-e88cb9644483" Dec 07 16:03:49 crc kubenswrapper[4716]: I1207 16:03:49.658054 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:03:49 crc kubenswrapper[4716]: E1207 16:03:49.658148 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 16:03:50 crc kubenswrapper[4716]: I1207 16:03:50.657033 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 16:03:50 crc kubenswrapper[4716]: E1207 16:03:50.657184 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 16:03:51 crc kubenswrapper[4716]: I1207 16:03:51.657167 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 16:03:51 crc kubenswrapper[4716]: I1207 16:03:51.657237 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:03:51 crc kubenswrapper[4716]: E1207 16:03:51.657286 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 16:03:51 crc kubenswrapper[4716]: E1207 16:03:51.657359 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 16:03:51 crc kubenswrapper[4716]: I1207 16:03:51.657437 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nx2lh" Dec 07 16:03:51 crc kubenswrapper[4716]: E1207 16:03:51.657495 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nx2lh" podUID="04b37f6c-457b-4f9a-8733-e88cb9644483" Dec 07 16:03:52 crc kubenswrapper[4716]: I1207 16:03:52.656858 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 16:03:52 crc kubenswrapper[4716]: E1207 16:03:52.657257 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 16:03:53 crc kubenswrapper[4716]: I1207 16:03:53.657088 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 16:03:53 crc kubenswrapper[4716]: E1207 16:03:53.657204 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 16:03:53 crc kubenswrapper[4716]: I1207 16:03:53.657280 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:03:53 crc kubenswrapper[4716]: I1207 16:03:53.657350 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nx2lh" Dec 07 16:03:53 crc kubenswrapper[4716]: E1207 16:03:53.657558 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 16:03:53 crc kubenswrapper[4716]: E1207 16:03:53.657801 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nx2lh" podUID="04b37f6c-457b-4f9a-8733-e88cb9644483" Dec 07 16:03:54 crc kubenswrapper[4716]: I1207 16:03:54.657144 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 16:03:54 crc kubenswrapper[4716]: E1207 16:03:54.657364 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 16:03:54 crc kubenswrapper[4716]: I1207 16:03:54.658800 4716 scope.go:117] "RemoveContainer" containerID="c60794c12cb6e7647d4c1d0bee76e214a71d9cb818d9c98dbf9880fb030ed54b" Dec 07 16:03:54 crc kubenswrapper[4716]: E1207 16:03:54.659149 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-qvc7p_openshift-ovn-kubernetes(c54d618a-dfec-4f2a-a288-f5188c8a128a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" podUID="c54d618a-dfec-4f2a-a288-f5188c8a128a" Dec 07 16:03:55 crc kubenswrapper[4716]: I1207 16:03:55.656760 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 16:03:55 crc kubenswrapper[4716]: I1207 16:03:55.656886 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nx2lh" Dec 07 16:03:55 crc kubenswrapper[4716]: E1207 16:03:55.656930 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 16:03:55 crc kubenswrapper[4716]: E1207 16:03:55.657113 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nx2lh" podUID="04b37f6c-457b-4f9a-8733-e88cb9644483" Dec 07 16:03:55 crc kubenswrapper[4716]: I1207 16:03:55.657197 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:03:55 crc kubenswrapper[4716]: E1207 16:03:55.657370 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 16:03:56 crc kubenswrapper[4716]: I1207 16:03:56.656600 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 16:03:56 crc kubenswrapper[4716]: E1207 16:03:56.656884 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 16:03:57 crc kubenswrapper[4716]: I1207 16:03:57.657057 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 16:03:57 crc kubenswrapper[4716]: I1207 16:03:57.657177 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:03:57 crc kubenswrapper[4716]: E1207 16:03:57.658243 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 16:03:57 crc kubenswrapper[4716]: I1207 16:03:57.658288 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nx2lh" Dec 07 16:03:57 crc kubenswrapper[4716]: E1207 16:03:57.658338 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 16:03:57 crc kubenswrapper[4716]: E1207 16:03:57.658626 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nx2lh" podUID="04b37f6c-457b-4f9a-8733-e88cb9644483" Dec 07 16:03:58 crc kubenswrapper[4716]: I1207 16:03:58.202917 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/04b37f6c-457b-4f9a-8733-e88cb9644483-metrics-certs\") pod \"network-metrics-daemon-nx2lh\" (UID: \"04b37f6c-457b-4f9a-8733-e88cb9644483\") " pod="openshift-multus/network-metrics-daemon-nx2lh" Dec 07 16:03:58 crc kubenswrapper[4716]: E1207 16:03:58.203203 4716 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 07 16:03:58 crc kubenswrapper[4716]: E1207 16:03:58.203360 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/04b37f6c-457b-4f9a-8733-e88cb9644483-metrics-certs podName:04b37f6c-457b-4f9a-8733-e88cb9644483 nodeName:}" failed. No retries permitted until 2025-12-07 16:05:02.203331869 +0000 UTC m=+164.893616821 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/04b37f6c-457b-4f9a-8733-e88cb9644483-metrics-certs") pod "network-metrics-daemon-nx2lh" (UID: "04b37f6c-457b-4f9a-8733-e88cb9644483") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 07 16:03:58 crc kubenswrapper[4716]: I1207 16:03:58.657657 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 16:03:58 crc kubenswrapper[4716]: E1207 16:03:58.657782 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 16:03:59 crc kubenswrapper[4716]: I1207 16:03:59.657741 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nx2lh" Dec 07 16:03:59 crc kubenswrapper[4716]: I1207 16:03:59.657889 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 16:03:59 crc kubenswrapper[4716]: E1207 16:03:59.658155 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nx2lh" podUID="04b37f6c-457b-4f9a-8733-e88cb9644483" Dec 07 16:03:59 crc kubenswrapper[4716]: I1207 16:03:59.658197 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:03:59 crc kubenswrapper[4716]: E1207 16:03:59.658266 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 16:03:59 crc kubenswrapper[4716]: E1207 16:03:59.658298 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 16:04:00 crc kubenswrapper[4716]: I1207 16:04:00.656995 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 16:04:00 crc kubenswrapper[4716]: E1207 16:04:00.657448 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 16:04:01 crc kubenswrapper[4716]: I1207 16:04:01.657778 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 16:04:01 crc kubenswrapper[4716]: I1207 16:04:01.657819 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nx2lh" Dec 07 16:04:01 crc kubenswrapper[4716]: I1207 16:04:01.659336 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:04:01 crc kubenswrapper[4716]: E1207 16:04:01.659654 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 16:04:01 crc kubenswrapper[4716]: E1207 16:04:01.659667 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 16:04:01 crc kubenswrapper[4716]: E1207 16:04:01.659764 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nx2lh" podUID="04b37f6c-457b-4f9a-8733-e88cb9644483" Dec 07 16:04:02 crc kubenswrapper[4716]: I1207 16:04:02.657208 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 16:04:02 crc kubenswrapper[4716]: E1207 16:04:02.657436 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 16:04:03 crc kubenswrapper[4716]: I1207 16:04:03.657602 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 16:04:03 crc kubenswrapper[4716]: I1207 16:04:03.658309 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nx2lh" Dec 07 16:04:03 crc kubenswrapper[4716]: E1207 16:04:03.658495 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 16:04:03 crc kubenswrapper[4716]: I1207 16:04:03.658861 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:04:03 crc kubenswrapper[4716]: E1207 16:04:03.659005 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 16:04:03 crc kubenswrapper[4716]: E1207 16:04:03.659403 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nx2lh" podUID="04b37f6c-457b-4f9a-8733-e88cb9644483" Dec 07 16:04:04 crc kubenswrapper[4716]: I1207 16:04:04.657346 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 16:04:04 crc kubenswrapper[4716]: E1207 16:04:04.657704 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 16:04:05 crc kubenswrapper[4716]: I1207 16:04:05.657233 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 16:04:05 crc kubenswrapper[4716]: I1207 16:04:05.657345 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nx2lh" Dec 07 16:04:05 crc kubenswrapper[4716]: E1207 16:04:05.657605 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 16:04:05 crc kubenswrapper[4716]: I1207 16:04:05.657758 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:04:05 crc kubenswrapper[4716]: E1207 16:04:05.658263 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nx2lh" podUID="04b37f6c-457b-4f9a-8733-e88cb9644483" Dec 07 16:04:05 crc kubenswrapper[4716]: E1207 16:04:05.658539 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 16:04:05 crc kubenswrapper[4716]: I1207 16:04:05.659005 4716 scope.go:117] "RemoveContainer" containerID="c60794c12cb6e7647d4c1d0bee76e214a71d9cb818d9c98dbf9880fb030ed54b" Dec 07 16:04:05 crc kubenswrapper[4716]: E1207 16:04:05.659354 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-qvc7p_openshift-ovn-kubernetes(c54d618a-dfec-4f2a-a288-f5188c8a128a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" podUID="c54d618a-dfec-4f2a-a288-f5188c8a128a" Dec 07 16:04:06 crc kubenswrapper[4716]: I1207 16:04:06.656964 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 16:04:06 crc kubenswrapper[4716]: E1207 16:04:06.657115 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 16:04:07 crc kubenswrapper[4716]: I1207 16:04:07.657647 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 16:04:07 crc kubenswrapper[4716]: I1207 16:04:07.657671 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nx2lh" Dec 07 16:04:07 crc kubenswrapper[4716]: E1207 16:04:07.660747 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 16:04:07 crc kubenswrapper[4716]: I1207 16:04:07.660785 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:04:07 crc kubenswrapper[4716]: E1207 16:04:07.660893 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nx2lh" podUID="04b37f6c-457b-4f9a-8733-e88cb9644483" Dec 07 16:04:07 crc kubenswrapper[4716]: E1207 16:04:07.661009 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 16:04:08 crc kubenswrapper[4716]: I1207 16:04:08.656823 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 16:04:08 crc kubenswrapper[4716]: E1207 16:04:08.657048 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 16:04:09 crc kubenswrapper[4716]: I1207 16:04:09.657783 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nx2lh" Dec 07 16:04:09 crc kubenswrapper[4716]: I1207 16:04:09.657886 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 16:04:09 crc kubenswrapper[4716]: E1207 16:04:09.658753 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 16:04:09 crc kubenswrapper[4716]: I1207 16:04:09.658015 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:04:09 crc kubenswrapper[4716]: E1207 16:04:09.658874 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 16:04:09 crc kubenswrapper[4716]: E1207 16:04:09.658598 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nx2lh" podUID="04b37f6c-457b-4f9a-8733-e88cb9644483" Dec 07 16:04:10 crc kubenswrapper[4716]: I1207 16:04:10.657264 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 16:04:10 crc kubenswrapper[4716]: E1207 16:04:10.657416 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 16:04:11 crc kubenswrapper[4716]: I1207 16:04:11.657260 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 16:04:11 crc kubenswrapper[4716]: I1207 16:04:11.657318 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nx2lh" Dec 07 16:04:11 crc kubenswrapper[4716]: I1207 16:04:11.657318 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:04:11 crc kubenswrapper[4716]: E1207 16:04:11.657481 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 16:04:11 crc kubenswrapper[4716]: E1207 16:04:11.657628 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nx2lh" podUID="04b37f6c-457b-4f9a-8733-e88cb9644483" Dec 07 16:04:11 crc kubenswrapper[4716]: E1207 16:04:11.657774 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 16:04:12 crc kubenswrapper[4716]: I1207 16:04:12.656576 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 16:04:12 crc kubenswrapper[4716]: E1207 16:04:12.656710 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 16:04:13 crc kubenswrapper[4716]: I1207 16:04:13.263364 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-k4wr2_0baf8575-26cf-4a72-8a6d-ae94c078eaad/kube-multus/1.log" Dec 07 16:04:13 crc kubenswrapper[4716]: I1207 16:04:13.264227 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-k4wr2_0baf8575-26cf-4a72-8a6d-ae94c078eaad/kube-multus/0.log" Dec 07 16:04:13 crc kubenswrapper[4716]: I1207 16:04:13.264317 4716 generic.go:334] "Generic (PLEG): container finished" podID="0baf8575-26cf-4a72-8a6d-ae94c078eaad" containerID="6fca3b2ea503451cfaf7afb00445fe0adb0acc87d436d1eebba5068452b8e7cb" exitCode=1 Dec 07 16:04:13 crc kubenswrapper[4716]: I1207 16:04:13.264372 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-k4wr2" event={"ID":"0baf8575-26cf-4a72-8a6d-ae94c078eaad","Type":"ContainerDied","Data":"6fca3b2ea503451cfaf7afb00445fe0adb0acc87d436d1eebba5068452b8e7cb"} Dec 07 16:04:13 crc kubenswrapper[4716]: I1207 16:04:13.264433 4716 scope.go:117] "RemoveContainer" containerID="f0ca2b3cd5a11033afd219b0125be7ab5acd750935d7eb89a470eff3d93a0838" Dec 07 16:04:13 crc kubenswrapper[4716]: I1207 16:04:13.265677 4716 scope.go:117] "RemoveContainer" containerID="6fca3b2ea503451cfaf7afb00445fe0adb0acc87d436d1eebba5068452b8e7cb" Dec 07 16:04:13 crc kubenswrapper[4716]: E1207 16:04:13.266354 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-k4wr2_openshift-multus(0baf8575-26cf-4a72-8a6d-ae94c078eaad)\"" pod="openshift-multus/multus-k4wr2" podUID="0baf8575-26cf-4a72-8a6d-ae94c078eaad" Dec 07 16:04:13 crc kubenswrapper[4716]: I1207 16:04:13.657554 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 16:04:13 crc kubenswrapper[4716]: I1207 16:04:13.657761 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:04:13 crc kubenswrapper[4716]: I1207 16:04:13.658117 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nx2lh" Dec 07 16:04:13 crc kubenswrapper[4716]: E1207 16:04:13.658616 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 16:04:13 crc kubenswrapper[4716]: E1207 16:04:13.659679 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nx2lh" podUID="04b37f6c-457b-4f9a-8733-e88cb9644483" Dec 07 16:04:13 crc kubenswrapper[4716]: E1207 16:04:13.662461 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 16:04:14 crc kubenswrapper[4716]: I1207 16:04:14.269888 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-k4wr2_0baf8575-26cf-4a72-8a6d-ae94c078eaad/kube-multus/1.log" Dec 07 16:04:14 crc kubenswrapper[4716]: I1207 16:04:14.657636 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 16:04:14 crc kubenswrapper[4716]: E1207 16:04:14.657835 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 16:04:15 crc kubenswrapper[4716]: I1207 16:04:15.657042 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 16:04:15 crc kubenswrapper[4716]: I1207 16:04:15.657042 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nx2lh" Dec 07 16:04:15 crc kubenswrapper[4716]: E1207 16:04:15.657249 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 16:04:15 crc kubenswrapper[4716]: I1207 16:04:15.657212 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:04:15 crc kubenswrapper[4716]: E1207 16:04:15.657483 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nx2lh" podUID="04b37f6c-457b-4f9a-8733-e88cb9644483" Dec 07 16:04:15 crc kubenswrapper[4716]: E1207 16:04:15.657751 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 16:04:16 crc kubenswrapper[4716]: I1207 16:04:16.656743 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 16:04:16 crc kubenswrapper[4716]: E1207 16:04:16.656905 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 16:04:17 crc kubenswrapper[4716]: E1207 16:04:17.610718 4716 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Dec 07 16:04:17 crc kubenswrapper[4716]: I1207 16:04:17.657124 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nx2lh" Dec 07 16:04:17 crc kubenswrapper[4716]: I1207 16:04:17.657142 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:04:17 crc kubenswrapper[4716]: I1207 16:04:17.657148 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 16:04:17 crc kubenswrapper[4716]: E1207 16:04:17.659527 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nx2lh" podUID="04b37f6c-457b-4f9a-8733-e88cb9644483" Dec 07 16:04:17 crc kubenswrapper[4716]: E1207 16:04:17.659685 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 16:04:17 crc kubenswrapper[4716]: E1207 16:04:17.659813 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 16:04:17 crc kubenswrapper[4716]: E1207 16:04:17.759529 4716 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 07 16:04:18 crc kubenswrapper[4716]: I1207 16:04:18.657588 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 16:04:18 crc kubenswrapper[4716]: E1207 16:04:18.657698 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 16:04:19 crc kubenswrapper[4716]: I1207 16:04:19.657386 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:04:19 crc kubenswrapper[4716]: I1207 16:04:19.657519 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 16:04:19 crc kubenswrapper[4716]: E1207 16:04:19.657558 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 16:04:19 crc kubenswrapper[4716]: E1207 16:04:19.657707 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 16:04:19 crc kubenswrapper[4716]: I1207 16:04:19.657786 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nx2lh" Dec 07 16:04:19 crc kubenswrapper[4716]: E1207 16:04:19.657895 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nx2lh" podUID="04b37f6c-457b-4f9a-8733-e88cb9644483" Dec 07 16:04:20 crc kubenswrapper[4716]: I1207 16:04:20.657353 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 16:04:20 crc kubenswrapper[4716]: E1207 16:04:20.657625 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 16:04:20 crc kubenswrapper[4716]: I1207 16:04:20.658894 4716 scope.go:117] "RemoveContainer" containerID="c60794c12cb6e7647d4c1d0bee76e214a71d9cb818d9c98dbf9880fb030ed54b" Dec 07 16:04:21 crc kubenswrapper[4716]: I1207 16:04:21.298584 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qvc7p_c54d618a-dfec-4f2a-a288-f5188c8a128a/ovnkube-controller/3.log" Dec 07 16:04:21 crc kubenswrapper[4716]: I1207 16:04:21.301891 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" event={"ID":"c54d618a-dfec-4f2a-a288-f5188c8a128a","Type":"ContainerStarted","Data":"e462e925550f957641f83d75b2990297a4cfd973b9227fb08dc610a75b1d6e79"} Dec 07 16:04:21 crc kubenswrapper[4716]: I1207 16:04:21.302919 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" Dec 07 16:04:21 crc kubenswrapper[4716]: I1207 16:04:21.511306 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" podStartSLOduration=102.511283145 podStartE2EDuration="1m42.511283145s" podCreationTimestamp="2025-12-07 16:02:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:04:21.351508849 +0000 UTC m=+124.041793761" watchObservedRunningTime="2025-12-07 16:04:21.511283145 +0000 UTC m=+124.201568077" Dec 07 16:04:21 crc kubenswrapper[4716]: I1207 16:04:21.512239 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-nx2lh"] Dec 07 16:04:21 crc kubenswrapper[4716]: I1207 16:04:21.512330 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nx2lh" Dec 07 16:04:21 crc kubenswrapper[4716]: E1207 16:04:21.512430 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nx2lh" podUID="04b37f6c-457b-4f9a-8733-e88cb9644483" Dec 07 16:04:21 crc kubenswrapper[4716]: I1207 16:04:21.656740 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:04:21 crc kubenswrapper[4716]: I1207 16:04:21.656781 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 16:04:21 crc kubenswrapper[4716]: E1207 16:04:21.656948 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 16:04:21 crc kubenswrapper[4716]: E1207 16:04:21.657063 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 16:04:22 crc kubenswrapper[4716]: I1207 16:04:22.657373 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 16:04:22 crc kubenswrapper[4716]: E1207 16:04:22.657782 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 16:04:22 crc kubenswrapper[4716]: E1207 16:04:22.761039 4716 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 07 16:04:23 crc kubenswrapper[4716]: I1207 16:04:23.657876 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 16:04:23 crc kubenswrapper[4716]: E1207 16:04:23.658008 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 16:04:23 crc kubenswrapper[4716]: I1207 16:04:23.658216 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nx2lh" Dec 07 16:04:23 crc kubenswrapper[4716]: E1207 16:04:23.658266 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nx2lh" podUID="04b37f6c-457b-4f9a-8733-e88cb9644483" Dec 07 16:04:23 crc kubenswrapper[4716]: I1207 16:04:23.658368 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:04:23 crc kubenswrapper[4716]: E1207 16:04:23.658410 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 16:04:24 crc kubenswrapper[4716]: I1207 16:04:24.657641 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 16:04:24 crc kubenswrapper[4716]: E1207 16:04:24.657790 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 16:04:25 crc kubenswrapper[4716]: I1207 16:04:25.657237 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nx2lh" Dec 07 16:04:25 crc kubenswrapper[4716]: I1207 16:04:25.657295 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:04:25 crc kubenswrapper[4716]: E1207 16:04:25.657386 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nx2lh" podUID="04b37f6c-457b-4f9a-8733-e88cb9644483" Dec 07 16:04:25 crc kubenswrapper[4716]: I1207 16:04:25.657236 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 16:04:25 crc kubenswrapper[4716]: I1207 16:04:25.657761 4716 scope.go:117] "RemoveContainer" containerID="6fca3b2ea503451cfaf7afb00445fe0adb0acc87d436d1eebba5068452b8e7cb" Dec 07 16:04:25 crc kubenswrapper[4716]: E1207 16:04:25.657781 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 16:04:25 crc kubenswrapper[4716]: E1207 16:04:25.657859 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 16:04:26 crc kubenswrapper[4716]: I1207 16:04:26.330024 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-k4wr2_0baf8575-26cf-4a72-8a6d-ae94c078eaad/kube-multus/1.log" Dec 07 16:04:26 crc kubenswrapper[4716]: I1207 16:04:26.330336 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-k4wr2" event={"ID":"0baf8575-26cf-4a72-8a6d-ae94c078eaad","Type":"ContainerStarted","Data":"a7122711012345eed920500d2894cb3e2cb321f8b1c45845be71f106475783c8"} Dec 07 16:04:26 crc kubenswrapper[4716]: I1207 16:04:26.657466 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 16:04:26 crc kubenswrapper[4716]: E1207 16:04:26.657602 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 07 16:04:27 crc kubenswrapper[4716]: I1207 16:04:27.657290 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 16:04:27 crc kubenswrapper[4716]: I1207 16:04:27.657390 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nx2lh" Dec 07 16:04:27 crc kubenswrapper[4716]: I1207 16:04:27.657401 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:04:27 crc kubenswrapper[4716]: E1207 16:04:27.658469 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nx2lh" podUID="04b37f6c-457b-4f9a-8733-e88cb9644483" Dec 07 16:04:27 crc kubenswrapper[4716]: E1207 16:04:27.658560 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 07 16:04:27 crc kubenswrapper[4716]: E1207 16:04:27.658599 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.657778 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.661926 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.662030 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.760174 4716 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.823197 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-qjcjv"] Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.824106 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-qjcjv" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.825357 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-4d2qh"] Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.826301 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4d2qh" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.828582 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-j26br"] Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.829561 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j26br" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.830487 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-j7zhr"] Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.831294 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-j7zhr" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.861676 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.861753 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.862177 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.862640 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.862730 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.862646 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.863005 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-7ff64"] Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.863072 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.863381 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.864134 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-gc4cp"] Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.864648 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.864789 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7ff64" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.864903 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.865174 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.865331 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-gc4cp" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.865602 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.865908 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.866374 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.866417 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.866454 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.866607 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.866617 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.866657 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.866748 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.866833 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.866606 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.866876 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.867029 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.866844 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.867169 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-6xp6f"] Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.867417 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.868247 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-6xp6f" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.869230 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-z587m"] Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.871258 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-z587m" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.872856 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8sxfn"] Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.873995 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8sxfn" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.877731 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-x2nd8"] Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.880000 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-x2nd8" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.883203 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.883208 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.885435 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.885889 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.885935 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.886245 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.889206 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.897152 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.897158 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.897529 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.898016 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.901911 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-ndp5v"] Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.902494 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-ndp5v" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.905270 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-pszqv"] Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.906059 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zjwcl"] Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.906758 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zjwcl" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.906068 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-pszqv" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.909374 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.909521 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.909659 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.909823 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.909970 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.910037 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.909973 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.910260 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.910372 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.910397 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.910410 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.910478 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.910542 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.910599 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.911067 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.911336 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.912701 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.912713 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.912816 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.912843 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.912888 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.912933 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.912948 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.913145 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.914684 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.914819 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.914928 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.915670 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.915692 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.915813 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.917355 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-jjn8w"] Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.917709 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.917984 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-jjn8w" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.920569 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.920739 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-9bs98"] Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.923484 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-b257c"] Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.923654 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-9bs98" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.925255 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.925398 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.925533 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.925618 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.928743 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-b257c" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.925763 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.936151 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.943106 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.943254 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.943320 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.943467 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.943640 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.943660 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.943717 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.943807 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.943886 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.975033 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.975513 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3cef2105-ae00-4224-bf5d-246393f3caa6-serving-cert\") pod \"controller-manager-879f6c89f-qjcjv\" (UID: \"3cef2105-ae00-4224-bf5d-246393f3caa6\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qjcjv" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.975551 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3cef2105-ae00-4224-bf5d-246393f3caa6-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-qjcjv\" (UID: \"3cef2105-ae00-4224-bf5d-246393f3caa6\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qjcjv" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.975597 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/297ee903-c479-460e-80df-a0c508c45fbe-audit-dir\") pod \"apiserver-7bbb656c7d-j26br\" (UID: \"297ee903-c479-460e-80df-a0c508c45fbe\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j26br" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.975630 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9a0a9cf5-14e9-41e1-8c6c-27b7b3b8c7b4-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-8sxfn\" (UID: \"9a0a9cf5-14e9-41e1-8c6c-27b7b3b8c7b4\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8sxfn" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.975657 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/297ee903-c479-460e-80df-a0c508c45fbe-etcd-client\") pod \"apiserver-7bbb656c7d-j26br\" (UID: \"297ee903-c479-460e-80df-a0c508c45fbe\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j26br" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.975690 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f7305ef-d76f-4be3-9151-7247f7590fe0-config\") pod \"apiserver-76f77b778f-gc4cp\" (UID: \"6f7305ef-d76f-4be3-9151-7247f7590fe0\") " pod="openshift-apiserver/apiserver-76f77b778f-gc4cp" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.975724 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-clrp8\" (UniqueName: \"kubernetes.io/projected/9a0a9cf5-14e9-41e1-8c6c-27b7b3b8c7b4-kube-api-access-clrp8\") pod \"cluster-image-registry-operator-dc59b4c8b-8sxfn\" (UID: \"9a0a9cf5-14e9-41e1-8c6c-27b7b3b8c7b4\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8sxfn" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.975758 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/297ee903-c479-460e-80df-a0c508c45fbe-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-j26br\" (UID: \"297ee903-c479-460e-80df-a0c508c45fbe\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j26br" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.975791 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4nhkp\" (UniqueName: \"kubernetes.io/projected/6f7305ef-d76f-4be3-9151-7247f7590fe0-kube-api-access-4nhkp\") pod \"apiserver-76f77b778f-gc4cp\" (UID: \"6f7305ef-d76f-4be3-9151-7247f7590fe0\") " pod="openshift-apiserver/apiserver-76f77b778f-gc4cp" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.975817 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-knwnm\" (UniqueName: \"kubernetes.io/projected/23992ff8-b006-40d1-9048-1950f10fff7d-kube-api-access-knwnm\") pod \"openshift-config-operator-7777fb866f-x2nd8\" (UID: \"23992ff8-b006-40d1-9048-1950f10fff7d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-x2nd8" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.975851 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/4df2ba74-d18e-4e94-bee1-703645b1a463-images\") pod \"machine-api-operator-5694c8668f-j7zhr\" (UID: \"4df2ba74-d18e-4e94-bee1-703645b1a463\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j7zhr" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.975879 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/6e5a490b-b9b3-4f19-aebf-14f6c1efbaa2-machine-approver-tls\") pod \"machine-approver-56656f9798-7ff64\" (UID: \"6e5a490b-b9b3-4f19-aebf-14f6c1efbaa2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7ff64" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.975913 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ffff2831-338e-423e-81f2-c8d3b5de7785-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-z587m\" (UID: \"ffff2831-338e-423e-81f2-c8d3b5de7785\") " pod="openshift-authentication/oauth-openshift-558db77b4-z587m" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.975942 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xlqp9\" (UniqueName: \"kubernetes.io/projected/f086c3a7-c2a4-4066-aed6-270b5dd6eef8-kube-api-access-xlqp9\") pod \"authentication-operator-69f744f599-6xp6f\" (UID: \"f086c3a7-c2a4-4066-aed6-270b5dd6eef8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6xp6f" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.975972 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/4df2ba74-d18e-4e94-bee1-703645b1a463-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-j7zhr\" (UID: \"4df2ba74-d18e-4e94-bee1-703645b1a463\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j7zhr" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.976007 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9a0a9cf5-14e9-41e1-8c6c-27b7b3b8c7b4-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-8sxfn\" (UID: \"9a0a9cf5-14e9-41e1-8c6c-27b7b3b8c7b4\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8sxfn" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.976037 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/6f7305ef-d76f-4be3-9151-7247f7590fe0-etcd-client\") pod \"apiserver-76f77b778f-gc4cp\" (UID: \"6f7305ef-d76f-4be3-9151-7247f7590fe0\") " pod="openshift-apiserver/apiserver-76f77b778f-gc4cp" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.976068 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qlvv6\" (UniqueName: \"kubernetes.io/projected/97d9c075-80c3-4dcb-aebc-649eec930413-kube-api-access-qlvv6\") pod \"route-controller-manager-6576b87f9c-4d2qh\" (UID: \"97d9c075-80c3-4dcb-aebc-649eec930413\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4d2qh" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.976115 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6e5a490b-b9b3-4f19-aebf-14f6c1efbaa2-auth-proxy-config\") pod \"machine-approver-56656f9798-7ff64\" (UID: \"6e5a490b-b9b3-4f19-aebf-14f6c1efbaa2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7ff64" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.976146 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ffff2831-338e-423e-81f2-c8d3b5de7785-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-z587m\" (UID: \"ffff2831-338e-423e-81f2-c8d3b5de7785\") " pod="openshift-authentication/oauth-openshift-558db77b4-z587m" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.976178 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/297ee903-c479-460e-80df-a0c508c45fbe-serving-cert\") pod \"apiserver-7bbb656c7d-j26br\" (UID: \"297ee903-c479-460e-80df-a0c508c45fbe\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j26br" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.976206 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/6f7305ef-d76f-4be3-9151-7247f7590fe0-image-import-ca\") pod \"apiserver-76f77b778f-gc4cp\" (UID: \"6f7305ef-d76f-4be3-9151-7247f7590fe0\") " pod="openshift-apiserver/apiserver-76f77b778f-gc4cp" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.976239 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ffff2831-338e-423e-81f2-c8d3b5de7785-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-z587m\" (UID: \"ffff2831-338e-423e-81f2-c8d3b5de7785\") " pod="openshift-authentication/oauth-openshift-558db77b4-z587m" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.976273 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ffff2831-338e-423e-81f2-c8d3b5de7785-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-z587m\" (UID: \"ffff2831-338e-423e-81f2-c8d3b5de7785\") " pod="openshift-authentication/oauth-openshift-558db77b4-z587m" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.976307 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/9a0a9cf5-14e9-41e1-8c6c-27b7b3b8c7b4-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-8sxfn\" (UID: \"9a0a9cf5-14e9-41e1-8c6c-27b7b3b8c7b4\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8sxfn" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.976338 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rvqph\" (UniqueName: \"kubernetes.io/projected/6e5a490b-b9b3-4f19-aebf-14f6c1efbaa2-kube-api-access-rvqph\") pod \"machine-approver-56656f9798-7ff64\" (UID: \"6e5a490b-b9b3-4f19-aebf-14f6c1efbaa2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7ff64" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.976367 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/23992ff8-b006-40d1-9048-1950f10fff7d-available-featuregates\") pod \"openshift-config-operator-7777fb866f-x2nd8\" (UID: \"23992ff8-b006-40d1-9048-1950f10fff7d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-x2nd8" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.976398 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3cef2105-ae00-4224-bf5d-246393f3caa6-client-ca\") pod \"controller-manager-879f6c89f-qjcjv\" (UID: \"3cef2105-ae00-4224-bf5d-246393f3caa6\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qjcjv" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.976438 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rhn6\" (UniqueName: \"kubernetes.io/projected/3cef2105-ae00-4224-bf5d-246393f3caa6-kube-api-access-8rhn6\") pod \"controller-manager-879f6c89f-qjcjv\" (UID: \"3cef2105-ae00-4224-bf5d-246393f3caa6\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qjcjv" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.976472 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ffff2831-338e-423e-81f2-c8d3b5de7785-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-z587m\" (UID: \"ffff2831-338e-423e-81f2-c8d3b5de7785\") " pod="openshift-authentication/oauth-openshift-558db77b4-z587m" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.978279 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ffff2831-338e-423e-81f2-c8d3b5de7785-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-z587m\" (UID: \"ffff2831-338e-423e-81f2-c8d3b5de7785\") " pod="openshift-authentication/oauth-openshift-558db77b4-z587m" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.978381 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4df2ba74-d18e-4e94-bee1-703645b1a463-config\") pod \"machine-api-operator-5694c8668f-j7zhr\" (UID: \"4df2ba74-d18e-4e94-bee1-703645b1a463\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j7zhr" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.978488 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/6f7305ef-d76f-4be3-9151-7247f7590fe0-audit\") pod \"apiserver-76f77b778f-gc4cp\" (UID: \"6f7305ef-d76f-4be3-9151-7247f7590fe0\") " pod="openshift-apiserver/apiserver-76f77b778f-gc4cp" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.978525 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f086c3a7-c2a4-4066-aed6-270b5dd6eef8-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-6xp6f\" (UID: \"f086c3a7-c2a4-4066-aed6-270b5dd6eef8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6xp6f" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.978552 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9sqq9\" (UniqueName: \"kubernetes.io/projected/4df2ba74-d18e-4e94-bee1-703645b1a463-kube-api-access-9sqq9\") pod \"machine-api-operator-5694c8668f-j7zhr\" (UID: \"4df2ba74-d18e-4e94-bee1-703645b1a463\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j7zhr" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.978574 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/6f7305ef-d76f-4be3-9151-7247f7590fe0-etcd-serving-ca\") pod \"apiserver-76f77b778f-gc4cp\" (UID: \"6f7305ef-d76f-4be3-9151-7247f7590fe0\") " pod="openshift-apiserver/apiserver-76f77b778f-gc4cp" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.978596 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e5a490b-b9b3-4f19-aebf-14f6c1efbaa2-config\") pod \"machine-approver-56656f9798-7ff64\" (UID: \"6e5a490b-b9b3-4f19-aebf-14f6c1efbaa2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7ff64" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.978619 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/6f7305ef-d76f-4be3-9151-7247f7590fe0-node-pullsecrets\") pod \"apiserver-76f77b778f-gc4cp\" (UID: \"6f7305ef-d76f-4be3-9151-7247f7590fe0\") " pod="openshift-apiserver/apiserver-76f77b778f-gc4cp" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.978641 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97d9c075-80c3-4dcb-aebc-649eec930413-config\") pod \"route-controller-manager-6576b87f9c-4d2qh\" (UID: \"97d9c075-80c3-4dcb-aebc-649eec930413\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4d2qh" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.978660 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/97d9c075-80c3-4dcb-aebc-649eec930413-client-ca\") pod \"route-controller-manager-6576b87f9c-4d2qh\" (UID: \"97d9c075-80c3-4dcb-aebc-649eec930413\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4d2qh" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.977141 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.978697 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ffff2831-338e-423e-81f2-c8d3b5de7785-audit-dir\") pod \"oauth-openshift-558db77b4-z587m\" (UID: \"ffff2831-338e-423e-81f2-c8d3b5de7785\") " pod="openshift-authentication/oauth-openshift-558db77b4-z587m" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.978723 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ffff2831-338e-423e-81f2-c8d3b5de7785-audit-policies\") pod \"oauth-openshift-558db77b4-z587m\" (UID: \"ffff2831-338e-423e-81f2-c8d3b5de7785\") " pod="openshift-authentication/oauth-openshift-558db77b4-z587m" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.978747 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6f7305ef-d76f-4be3-9151-7247f7590fe0-trusted-ca-bundle\") pod \"apiserver-76f77b778f-gc4cp\" (UID: \"6f7305ef-d76f-4be3-9151-7247f7590fe0\") " pod="openshift-apiserver/apiserver-76f77b778f-gc4cp" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.978765 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3cef2105-ae00-4224-bf5d-246393f3caa6-config\") pod \"controller-manager-879f6c89f-qjcjv\" (UID: \"3cef2105-ae00-4224-bf5d-246393f3caa6\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qjcjv" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.978785 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ffff2831-338e-423e-81f2-c8d3b5de7785-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-z587m\" (UID: \"ffff2831-338e-423e-81f2-c8d3b5de7785\") " pod="openshift-authentication/oauth-openshift-558db77b4-z587m" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.977581 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bcfw4"] Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.978857 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2tnlr\" (UniqueName: \"kubernetes.io/projected/297ee903-c479-460e-80df-a0c508c45fbe-kube-api-access-2tnlr\") pod \"apiserver-7bbb656c7d-j26br\" (UID: \"297ee903-c479-460e-80df-a0c508c45fbe\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j26br" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.979121 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ffff2831-338e-423e-81f2-c8d3b5de7785-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-z587m\" (UID: \"ffff2831-338e-423e-81f2-c8d3b5de7785\") " pod="openshift-authentication/oauth-openshift-558db77b4-z587m" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.979156 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f086c3a7-c2a4-4066-aed6-270b5dd6eef8-service-ca-bundle\") pod \"authentication-operator-69f744f599-6xp6f\" (UID: \"f086c3a7-c2a4-4066-aed6-270b5dd6eef8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6xp6f" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.979183 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f086c3a7-c2a4-4066-aed6-270b5dd6eef8-config\") pod \"authentication-operator-69f744f599-6xp6f\" (UID: \"f086c3a7-c2a4-4066-aed6-270b5dd6eef8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6xp6f" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.979206 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6f7305ef-d76f-4be3-9151-7247f7590fe0-audit-dir\") pod \"apiserver-76f77b778f-gc4cp\" (UID: \"6f7305ef-d76f-4be3-9151-7247f7590fe0\") " pod="openshift-apiserver/apiserver-76f77b778f-gc4cp" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.979225 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/297ee903-c479-460e-80df-a0c508c45fbe-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-j26br\" (UID: \"297ee903-c479-460e-80df-a0c508c45fbe\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j26br" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.979245 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/97d9c075-80c3-4dcb-aebc-649eec930413-serving-cert\") pod \"route-controller-manager-6576b87f9c-4d2qh\" (UID: \"97d9c075-80c3-4dcb-aebc-649eec930413\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4d2qh" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.979267 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ffff2831-338e-423e-81f2-c8d3b5de7785-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-z587m\" (UID: \"ffff2831-338e-423e-81f2-c8d3b5de7785\") " pod="openshift-authentication/oauth-openshift-558db77b4-z587m" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.979286 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ffff2831-338e-423e-81f2-c8d3b5de7785-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-z587m\" (UID: \"ffff2831-338e-423e-81f2-c8d3b5de7785\") " pod="openshift-authentication/oauth-openshift-558db77b4-z587m" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.979304 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nw5sf\" (UniqueName: \"kubernetes.io/projected/ffff2831-338e-423e-81f2-c8d3b5de7785-kube-api-access-nw5sf\") pod \"oauth-openshift-558db77b4-z587m\" (UID: \"ffff2831-338e-423e-81f2-c8d3b5de7785\") " pod="openshift-authentication/oauth-openshift-558db77b4-z587m" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.979326 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ffff2831-338e-423e-81f2-c8d3b5de7785-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-z587m\" (UID: \"ffff2831-338e-423e-81f2-c8d3b5de7785\") " pod="openshift-authentication/oauth-openshift-558db77b4-z587m" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.979347 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f086c3a7-c2a4-4066-aed6-270b5dd6eef8-serving-cert\") pod \"authentication-operator-69f744f599-6xp6f\" (UID: \"f086c3a7-c2a4-4066-aed6-270b5dd6eef8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6xp6f" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.979369 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6f7305ef-d76f-4be3-9151-7247f7590fe0-serving-cert\") pod \"apiserver-76f77b778f-gc4cp\" (UID: \"6f7305ef-d76f-4be3-9151-7247f7590fe0\") " pod="openshift-apiserver/apiserver-76f77b778f-gc4cp" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.979392 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/6f7305ef-d76f-4be3-9151-7247f7590fe0-encryption-config\") pod \"apiserver-76f77b778f-gc4cp\" (UID: \"6f7305ef-d76f-4be3-9151-7247f7590fe0\") " pod="openshift-apiserver/apiserver-76f77b778f-gc4cp" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.979413 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/23992ff8-b006-40d1-9048-1950f10fff7d-serving-cert\") pod \"openshift-config-operator-7777fb866f-x2nd8\" (UID: \"23992ff8-b006-40d1-9048-1950f10fff7d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-x2nd8" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.979432 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/297ee903-c479-460e-80df-a0c508c45fbe-audit-policies\") pod \"apiserver-7bbb656c7d-j26br\" (UID: \"297ee903-c479-460e-80df-a0c508c45fbe\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j26br" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.977540 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.979545 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/297ee903-c479-460e-80df-a0c508c45fbe-encryption-config\") pod \"apiserver-7bbb656c7d-j26br\" (UID: \"297ee903-c479-460e-80df-a0c508c45fbe\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j26br" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.984451 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.985734 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-d7c2c"] Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.985934 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bcfw4" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.988506 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.988811 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-zp6l4"] Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.988977 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-d7c2c" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.990108 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-zp6l4" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.993680 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.993992 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-h5l29"] Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.994257 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.994285 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.994308 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.994404 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.994464 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.994487 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.994564 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.994664 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.994701 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.994717 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.994729 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.994808 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-h5l29" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.994827 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.995253 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.995477 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.995782 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-4d257"] Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.996335 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-4d257" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.996366 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.996613 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-wrw67"] Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.996744 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.997184 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wrw67" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.998126 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-d49gn"] Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.998507 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-d49gn" Dec 07 16:04:28 crc kubenswrapper[4716]: I1207 16:04:28.999370 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mmfjq"] Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.002862 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rb966"] Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.002976 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mmfjq" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.003287 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-w7xhs"] Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.003357 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rb966" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.004207 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-w7xhs" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.004422 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-h6gnn"] Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.004997 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-h6gnn" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.005545 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-sq5lw"] Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.005994 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.006186 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-sq5lw" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.006287 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-64rzd"] Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.006968 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-64rzd" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.007132 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6j7q2"] Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.007550 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6j7q2" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.008120 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-kkvsk"] Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.008839 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-kkvsk" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.009095 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-j7zhr"] Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.010050 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bwsl4"] Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.010484 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bwsl4" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.011310 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-l9dgz"] Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.011925 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-l9dgz" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.012256 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-89v7m"] Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.012796 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-89v7m" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.013309 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-v6899"] Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.014016 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-v6899" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.014264 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-tqctv"] Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.014841 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-tqctv" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.016461 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-8sv58"] Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.017052 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-8sv58" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.017769 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29418720-x5rqz"] Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.018553 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29418720-x5rqz" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.018885 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-gc4cp"] Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.020266 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-z587m"] Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.021761 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-qjcjv"] Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.022782 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-n4492"] Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.023484 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-n4492" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.024384 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-64rzd"] Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.026307 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.027033 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zjwcl"] Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.028273 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rb966"] Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.028362 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-9bs98"] Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.029345 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-4d257"] Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.030257 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mmfjq"] Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.031314 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-j26br"] Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.032512 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8sxfn"] Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.033656 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6j7q2"] Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.035017 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-wrw67"] Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.037593 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-ndp5v"] Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.038634 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-b257c"] Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.040988 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-4d2qh"] Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.041063 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-sq5lw"] Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.046797 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.047032 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-x2nd8"] Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.049188 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-zp6l4"] Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.053178 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-h6gnn"] Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.056290 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-v9d58"] Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.057364 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-v9d58" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.061639 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-6xp6f"] Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.062859 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-w7xhs"] Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.064089 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-d7c2c"] Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.065407 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-kkvsk"] Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.067121 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-82bvt"] Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.067220 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.068438 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-82bvt" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.068783 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-d49gn"] Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.069898 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bcfw4"] Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.070963 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-jjn8w"] Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.072109 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-v6899"] Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.073672 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-89v7m"] Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.074795 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-pszqv"] Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.075848 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-v9d58"] Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.076977 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bwsl4"] Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.078108 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29418720-x5rqz"] Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.079161 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-l9dgz"] Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.080422 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-82bvt"] Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.081539 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8439dba6-4f32-4c02-9aad-27b9c40f7575-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-d49gn\" (UID: \"8439dba6-4f32-4c02-9aad-27b9c40f7575\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-d49gn" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.081584 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/51fb93ab-ea9f-4fce-9fd1-a57958bc9c10-trusted-ca\") pod \"console-operator-58897d9998-pszqv\" (UID: \"51fb93ab-ea9f-4fce-9fd1-a57958bc9c10\") " pod="openshift-console-operator/console-operator-58897d9998-pszqv" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.081624 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ffff2831-338e-423e-81f2-c8d3b5de7785-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-z587m\" (UID: \"ffff2831-338e-423e-81f2-c8d3b5de7785\") " pod="openshift-authentication/oauth-openshift-558db77b4-z587m" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.081652 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-8sv58"] Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.081652 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/9a0a9cf5-14e9-41e1-8c6c-27b7b3b8c7b4-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-8sxfn\" (UID: \"9a0a9cf5-14e9-41e1-8c6c-27b7b3b8c7b4\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8sxfn" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.081722 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rvqph\" (UniqueName: \"kubernetes.io/projected/6e5a490b-b9b3-4f19-aebf-14f6c1efbaa2-kube-api-access-rvqph\") pod \"machine-approver-56656f9798-7ff64\" (UID: \"6e5a490b-b9b3-4f19-aebf-14f6c1efbaa2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7ff64" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.081752 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a51dadb3-2af5-4c6c-a973-cdceb34cd79a-config\") pod \"openshift-apiserver-operator-796bbdcf4f-bcfw4\" (UID: \"a51dadb3-2af5-4c6c-a973-cdceb34cd79a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bcfw4" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.081785 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/cbf3f592-bbb1-4596-8ae1-0e62aa366f0b-oauth-serving-cert\") pod \"console-f9d7485db-ndp5v\" (UID: \"cbf3f592-bbb1-4596-8ae1-0e62aa366f0b\") " pod="openshift-console/console-f9d7485db-ndp5v" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.081814 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/23992ff8-b006-40d1-9048-1950f10fff7d-available-featuregates\") pod \"openshift-config-operator-7777fb866f-x2nd8\" (UID: \"23992ff8-b006-40d1-9048-1950f10fff7d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-x2nd8" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.081837 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3cef2105-ae00-4224-bf5d-246393f3caa6-client-ca\") pod \"controller-manager-879f6c89f-qjcjv\" (UID: \"3cef2105-ae00-4224-bf5d-246393f3caa6\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qjcjv" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.081859 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rhn6\" (UniqueName: \"kubernetes.io/projected/3cef2105-ae00-4224-bf5d-246393f3caa6-kube-api-access-8rhn6\") pod \"controller-manager-879f6c89f-qjcjv\" (UID: \"3cef2105-ae00-4224-bf5d-246393f3caa6\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qjcjv" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.081890 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ffff2831-338e-423e-81f2-c8d3b5de7785-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-z587m\" (UID: \"ffff2831-338e-423e-81f2-c8d3b5de7785\") " pod="openshift-authentication/oauth-openshift-558db77b4-z587m" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.081913 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ffff2831-338e-423e-81f2-c8d3b5de7785-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-z587m\" (UID: \"ffff2831-338e-423e-81f2-c8d3b5de7785\") " pod="openshift-authentication/oauth-openshift-558db77b4-z587m" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.081937 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4df2ba74-d18e-4e94-bee1-703645b1a463-config\") pod \"machine-api-operator-5694c8668f-j7zhr\" (UID: \"4df2ba74-d18e-4e94-bee1-703645b1a463\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j7zhr" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.081959 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/6f7305ef-d76f-4be3-9151-7247f7590fe0-audit\") pod \"apiserver-76f77b778f-gc4cp\" (UID: \"6f7305ef-d76f-4be3-9151-7247f7590fe0\") " pod="openshift-apiserver/apiserver-76f77b778f-gc4cp" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.081979 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cbf3f592-bbb1-4596-8ae1-0e62aa366f0b-trusted-ca-bundle\") pod \"console-f9d7485db-ndp5v\" (UID: \"cbf3f592-bbb1-4596-8ae1-0e62aa366f0b\") " pod="openshift-console/console-f9d7485db-ndp5v" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.081998 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97b4v\" (UniqueName: \"kubernetes.io/projected/fc892d74-c6d2-431f-a52c-95a8f4e2d03c-kube-api-access-97b4v\") pod \"dns-operator-744455d44c-zp6l4\" (UID: \"fc892d74-c6d2-431f-a52c-95a8f4e2d03c\") " pod="openshift-dns-operator/dns-operator-744455d44c-zp6l4" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.082019 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fw2t8\" (UniqueName: \"kubernetes.io/projected/cf1baeba-a0c2-4d91-ae76-eb4f0ceccd60-kube-api-access-fw2t8\") pod \"cluster-samples-operator-665b6dd947-jjn8w\" (UID: \"cf1baeba-a0c2-4d91-ae76-eb4f0ceccd60\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-jjn8w" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.082043 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f086c3a7-c2a4-4066-aed6-270b5dd6eef8-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-6xp6f\" (UID: \"f086c3a7-c2a4-4066-aed6-270b5dd6eef8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6xp6f" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.082064 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/c94992de-685b-42c3-8100-93996239d0b0-etcd-service-ca\") pod \"etcd-operator-b45778765-9bs98\" (UID: \"c94992de-685b-42c3-8100-93996239d0b0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9bs98" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.082139 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/85bc0929-de17-4d2d-8b03-1a05d207fb55-stats-auth\") pod \"router-default-5444994796-h5l29\" (UID: \"85bc0929-de17-4d2d-8b03-1a05d207fb55\") " pod="openshift-ingress/router-default-5444994796-h5l29" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.082187 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9sqq9\" (UniqueName: \"kubernetes.io/projected/4df2ba74-d18e-4e94-bee1-703645b1a463-kube-api-access-9sqq9\") pod \"machine-api-operator-5694c8668f-j7zhr\" (UID: \"4df2ba74-d18e-4e94-bee1-703645b1a463\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j7zhr" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.082206 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/6f7305ef-d76f-4be3-9151-7247f7590fe0-etcd-serving-ca\") pod \"apiserver-76f77b778f-gc4cp\" (UID: \"6f7305ef-d76f-4be3-9151-7247f7590fe0\") " pod="openshift-apiserver/apiserver-76f77b778f-gc4cp" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.082226 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e5a490b-b9b3-4f19-aebf-14f6c1efbaa2-config\") pod \"machine-approver-56656f9798-7ff64\" (UID: \"6e5a490b-b9b3-4f19-aebf-14f6c1efbaa2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7ff64" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.082249 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/6f7305ef-d76f-4be3-9151-7247f7590fe0-node-pullsecrets\") pod \"apiserver-76f77b778f-gc4cp\" (UID: \"6f7305ef-d76f-4be3-9151-7247f7590fe0\") " pod="openshift-apiserver/apiserver-76f77b778f-gc4cp" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.082271 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97d9c075-80c3-4dcb-aebc-649eec930413-config\") pod \"route-controller-manager-6576b87f9c-4d2qh\" (UID: \"97d9c075-80c3-4dcb-aebc-649eec930413\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4d2qh" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.082293 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/97d9c075-80c3-4dcb-aebc-649eec930413-client-ca\") pod \"route-controller-manager-6576b87f9c-4d2qh\" (UID: \"97d9c075-80c3-4dcb-aebc-649eec930413\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4d2qh" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.082315 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9d47p\" (UniqueName: \"kubernetes.io/projected/1306f3b2-2982-494e-9db0-86c2e05b7a82-kube-api-access-9d47p\") pod \"downloads-7954f5f757-b257c\" (UID: \"1306f3b2-2982-494e-9db0-86c2e05b7a82\") " pod="openshift-console/downloads-7954f5f757-b257c" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.082324 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/23992ff8-b006-40d1-9048-1950f10fff7d-available-featuregates\") pod \"openshift-config-operator-7777fb866f-x2nd8\" (UID: \"23992ff8-b006-40d1-9048-1950f10fff7d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-x2nd8" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.082342 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8439dba6-4f32-4c02-9aad-27b9c40f7575-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-d49gn\" (UID: \"8439dba6-4f32-4c02-9aad-27b9c40f7575\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-d49gn" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.082368 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/915af7e7-19b0-4b1c-9e55-3471210ccf40-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zjwcl\" (UID: \"915af7e7-19b0-4b1c-9e55-3471210ccf40\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zjwcl" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.082391 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/cbf3f592-bbb1-4596-8ae1-0e62aa366f0b-console-oauth-config\") pod \"console-f9d7485db-ndp5v\" (UID: \"cbf3f592-bbb1-4596-8ae1-0e62aa366f0b\") " pod="openshift-console/console-f9d7485db-ndp5v" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.082413 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a51dadb3-2af5-4c6c-a973-cdceb34cd79a-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-bcfw4\" (UID: \"a51dadb3-2af5-4c6c-a973-cdceb34cd79a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bcfw4" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.082781 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ffff2831-338e-423e-81f2-c8d3b5de7785-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-z587m\" (UID: \"ffff2831-338e-423e-81f2-c8d3b5de7785\") " pod="openshift-authentication/oauth-openshift-558db77b4-z587m" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.083126 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e5a490b-b9b3-4f19-aebf-14f6c1efbaa2-config\") pod \"machine-approver-56656f9798-7ff64\" (UID: \"6e5a490b-b9b3-4f19-aebf-14f6c1efbaa2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7ff64" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.083176 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3cef2105-ae00-4224-bf5d-246393f3caa6-client-ca\") pod \"controller-manager-879f6c89f-qjcjv\" (UID: \"3cef2105-ae00-4224-bf5d-246393f3caa6\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qjcjv" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.083188 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/6f7305ef-d76f-4be3-9151-7247f7590fe0-node-pullsecrets\") pod \"apiserver-76f77b778f-gc4cp\" (UID: \"6f7305ef-d76f-4be3-9151-7247f7590fe0\") " pod="openshift-apiserver/apiserver-76f77b778f-gc4cp" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.083499 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ffff2831-338e-423e-81f2-c8d3b5de7785-audit-dir\") pod \"oauth-openshift-558db77b4-z587m\" (UID: \"ffff2831-338e-423e-81f2-c8d3b5de7785\") " pod="openshift-authentication/oauth-openshift-558db77b4-z587m" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.083567 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-tqctv"] Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.083705 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ffff2831-338e-423e-81f2-c8d3b5de7785-audit-dir\") pod \"oauth-openshift-558db77b4-z587m\" (UID: \"ffff2831-338e-423e-81f2-c8d3b5de7785\") " pod="openshift-authentication/oauth-openshift-558db77b4-z587m" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.083739 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/6f7305ef-d76f-4be3-9151-7247f7590fe0-etcd-serving-ca\") pod \"apiserver-76f77b778f-gc4cp\" (UID: \"6f7305ef-d76f-4be3-9151-7247f7590fe0\") " pod="openshift-apiserver/apiserver-76f77b778f-gc4cp" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.083756 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ffff2831-338e-423e-81f2-c8d3b5de7785-audit-policies\") pod \"oauth-openshift-558db77b4-z587m\" (UID: \"ffff2831-338e-423e-81f2-c8d3b5de7785\") " pod="openshift-authentication/oauth-openshift-558db77b4-z587m" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.083791 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hk885\" (UniqueName: \"kubernetes.io/projected/85bc0929-de17-4d2d-8b03-1a05d207fb55-kube-api-access-hk885\") pod \"router-default-5444994796-h5l29\" (UID: \"85bc0929-de17-4d2d-8b03-1a05d207fb55\") " pod="openshift-ingress/router-default-5444994796-h5l29" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.083882 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-n4492"] Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.084035 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/915af7e7-19b0-4b1c-9e55-3471210ccf40-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zjwcl\" (UID: \"915af7e7-19b0-4b1c-9e55-3471210ccf40\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zjwcl" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.084095 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6f7305ef-d76f-4be3-9151-7247f7590fe0-trusted-ca-bundle\") pod \"apiserver-76f77b778f-gc4cp\" (UID: \"6f7305ef-d76f-4be3-9151-7247f7590fe0\") " pod="openshift-apiserver/apiserver-76f77b778f-gc4cp" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.084144 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3cef2105-ae00-4224-bf5d-246393f3caa6-config\") pod \"controller-manager-879f6c89f-qjcjv\" (UID: \"3cef2105-ae00-4224-bf5d-246393f3caa6\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qjcjv" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.084218 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ffff2831-338e-423e-81f2-c8d3b5de7785-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-z587m\" (UID: \"ffff2831-338e-423e-81f2-c8d3b5de7785\") " pod="openshift-authentication/oauth-openshift-558db77b4-z587m" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.084352 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2tnlr\" (UniqueName: \"kubernetes.io/projected/297ee903-c479-460e-80df-a0c508c45fbe-kube-api-access-2tnlr\") pod \"apiserver-7bbb656c7d-j26br\" (UID: \"297ee903-c479-460e-80df-a0c508c45fbe\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j26br" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.084384 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ffff2831-338e-423e-81f2-c8d3b5de7785-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-z587m\" (UID: \"ffff2831-338e-423e-81f2-c8d3b5de7785\") " pod="openshift-authentication/oauth-openshift-558db77b4-z587m" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.084437 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f086c3a7-c2a4-4066-aed6-270b5dd6eef8-service-ca-bundle\") pod \"authentication-operator-69f744f599-6xp6f\" (UID: \"f086c3a7-c2a4-4066-aed6-270b5dd6eef8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6xp6f" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.084460 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f086c3a7-c2a4-4066-aed6-270b5dd6eef8-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-6xp6f\" (UID: \"f086c3a7-c2a4-4066-aed6-270b5dd6eef8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6xp6f" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.084462 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ffff2831-338e-423e-81f2-c8d3b5de7785-audit-policies\") pod \"oauth-openshift-558db77b4-z587m\" (UID: \"ffff2831-338e-423e-81f2-c8d3b5de7785\") " pod="openshift-authentication/oauth-openshift-558db77b4-z587m" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.084572 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/6f7305ef-d76f-4be3-9151-7247f7590fe0-audit\") pod \"apiserver-76f77b778f-gc4cp\" (UID: \"6f7305ef-d76f-4be3-9151-7247f7590fe0\") " pod="openshift-apiserver/apiserver-76f77b778f-gc4cp" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.084630 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f086c3a7-c2a4-4066-aed6-270b5dd6eef8-config\") pod \"authentication-operator-69f744f599-6xp6f\" (UID: \"f086c3a7-c2a4-4066-aed6-270b5dd6eef8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6xp6f" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.084661 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6f7305ef-d76f-4be3-9151-7247f7590fe0-audit-dir\") pod \"apiserver-76f77b778f-gc4cp\" (UID: \"6f7305ef-d76f-4be3-9151-7247f7590fe0\") " pod="openshift-apiserver/apiserver-76f77b778f-gc4cp" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.084684 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/297ee903-c479-460e-80df-a0c508c45fbe-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-j26br\" (UID: \"297ee903-c479-460e-80df-a0c508c45fbe\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j26br" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.084768 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/97d9c075-80c3-4dcb-aebc-649eec930413-serving-cert\") pod \"route-controller-manager-6576b87f9c-4d2qh\" (UID: \"97d9c075-80c3-4dcb-aebc-649eec930413\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4d2qh" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.084798 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/85bc0929-de17-4d2d-8b03-1a05d207fb55-default-certificate\") pod \"router-default-5444994796-h5l29\" (UID: \"85bc0929-de17-4d2d-8b03-1a05d207fb55\") " pod="openshift-ingress/router-default-5444994796-h5l29" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.084823 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ffff2831-338e-423e-81f2-c8d3b5de7785-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-z587m\" (UID: \"ffff2831-338e-423e-81f2-c8d3b5de7785\") " pod="openshift-authentication/oauth-openshift-558db77b4-z587m" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.084869 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ffff2831-338e-423e-81f2-c8d3b5de7785-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-z587m\" (UID: \"ffff2831-338e-423e-81f2-c8d3b5de7785\") " pod="openshift-authentication/oauth-openshift-558db77b4-z587m" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.084889 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nw5sf\" (UniqueName: \"kubernetes.io/projected/ffff2831-338e-423e-81f2-c8d3b5de7785-kube-api-access-nw5sf\") pod \"oauth-openshift-558db77b4-z587m\" (UID: \"ffff2831-338e-423e-81f2-c8d3b5de7785\") " pod="openshift-authentication/oauth-openshift-558db77b4-z587m" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.084905 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f086c3a7-c2a4-4066-aed6-270b5dd6eef8-service-ca-bundle\") pod \"authentication-operator-69f744f599-6xp6f\" (UID: \"f086c3a7-c2a4-4066-aed6-270b5dd6eef8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6xp6f" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.084912 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bp2f8\" (UniqueName: \"kubernetes.io/projected/92724c2e-d7fd-4673-ae84-8929541f2a13-kube-api-access-bp2f8\") pod \"ingress-operator-5b745b69d9-wrw67\" (UID: \"92724c2e-d7fd-4673-ae84-8929541f2a13\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wrw67" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.084935 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ffff2831-338e-423e-81f2-c8d3b5de7785-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-z587m\" (UID: \"ffff2831-338e-423e-81f2-c8d3b5de7785\") " pod="openshift-authentication/oauth-openshift-558db77b4-z587m" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.084956 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/85bc0929-de17-4d2d-8b03-1a05d207fb55-service-ca-bundle\") pod \"router-default-5444994796-h5l29\" (UID: \"85bc0929-de17-4d2d-8b03-1a05d207fb55\") " pod="openshift-ingress/router-default-5444994796-h5l29" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.084979 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f086c3a7-c2a4-4066-aed6-270b5dd6eef8-serving-cert\") pod \"authentication-operator-69f744f599-6xp6f\" (UID: \"f086c3a7-c2a4-4066-aed6-270b5dd6eef8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6xp6f" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.085343 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6f7305ef-d76f-4be3-9151-7247f7590fe0-serving-cert\") pod \"apiserver-76f77b778f-gc4cp\" (UID: \"6f7305ef-d76f-4be3-9151-7247f7590fe0\") " pod="openshift-apiserver/apiserver-76f77b778f-gc4cp" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.085388 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/6f7305ef-d76f-4be3-9151-7247f7590fe0-encryption-config\") pod \"apiserver-76f77b778f-gc4cp\" (UID: \"6f7305ef-d76f-4be3-9151-7247f7590fe0\") " pod="openshift-apiserver/apiserver-76f77b778f-gc4cp" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.085446 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/23992ff8-b006-40d1-9048-1950f10fff7d-serving-cert\") pod \"openshift-config-operator-7777fb866f-x2nd8\" (UID: \"23992ff8-b006-40d1-9048-1950f10fff7d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-x2nd8" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.085483 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/297ee903-c479-460e-80df-a0c508c45fbe-audit-policies\") pod \"apiserver-7bbb656c7d-j26br\" (UID: \"297ee903-c479-460e-80df-a0c508c45fbe\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j26br" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.085545 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/297ee903-c479-460e-80df-a0c508c45fbe-encryption-config\") pod \"apiserver-7bbb656c7d-j26br\" (UID: \"297ee903-c479-460e-80df-a0c508c45fbe\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j26br" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.085624 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/92724c2e-d7fd-4673-ae84-8929541f2a13-trusted-ca\") pod \"ingress-operator-5b745b69d9-wrw67\" (UID: \"92724c2e-d7fd-4673-ae84-8929541f2a13\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wrw67" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.085689 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8439dba6-4f32-4c02-9aad-27b9c40f7575-config\") pod \"kube-controller-manager-operator-78b949d7b-d49gn\" (UID: \"8439dba6-4f32-4c02-9aad-27b9c40f7575\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-d49gn" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.085730 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vcjxn\" (UniqueName: \"kubernetes.io/projected/c94992de-685b-42c3-8100-93996239d0b0-kube-api-access-vcjxn\") pod \"etcd-operator-b45778765-9bs98\" (UID: \"c94992de-685b-42c3-8100-93996239d0b0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9bs98" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.085810 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-gcr9f"] Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.085953 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97d9c075-80c3-4dcb-aebc-649eec930413-config\") pod \"route-controller-manager-6576b87f9c-4d2qh\" (UID: \"97d9c075-80c3-4dcb-aebc-649eec930413\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4d2qh" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.086194 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/97d9c075-80c3-4dcb-aebc-649eec930413-client-ca\") pod \"route-controller-manager-6576b87f9c-4d2qh\" (UID: \"97d9c075-80c3-4dcb-aebc-649eec930413\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4d2qh" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.086389 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4df2ba74-d18e-4e94-bee1-703645b1a463-config\") pod \"machine-api-operator-5694c8668f-j7zhr\" (UID: \"4df2ba74-d18e-4e94-bee1-703645b1a463\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j7zhr" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.086512 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-gcr9f" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.086710 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3cef2105-ae00-4224-bf5d-246393f3caa6-config\") pod \"controller-manager-879f6c89f-qjcjv\" (UID: \"3cef2105-ae00-4224-bf5d-246393f3caa6\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qjcjv" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.086772 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.086793 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmljs\" (UniqueName: \"kubernetes.io/projected/cbf3f592-bbb1-4596-8ae1-0e62aa366f0b-kube-api-access-tmljs\") pod \"console-f9d7485db-ndp5v\" (UID: \"cbf3f592-bbb1-4596-8ae1-0e62aa366f0b\") " pod="openshift-console/console-f9d7485db-ndp5v" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.086853 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3cef2105-ae00-4224-bf5d-246393f3caa6-serving-cert\") pod \"controller-manager-879f6c89f-qjcjv\" (UID: \"3cef2105-ae00-4224-bf5d-246393f3caa6\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qjcjv" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.086877 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3cef2105-ae00-4224-bf5d-246393f3caa6-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-qjcjv\" (UID: \"3cef2105-ae00-4224-bf5d-246393f3caa6\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qjcjv" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.086894 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6f7305ef-d76f-4be3-9151-7247f7590fe0-trusted-ca-bundle\") pod \"apiserver-76f77b778f-gc4cp\" (UID: \"6f7305ef-d76f-4be3-9151-7247f7590fe0\") " pod="openshift-apiserver/apiserver-76f77b778f-gc4cp" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.086907 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/297ee903-c479-460e-80df-a0c508c45fbe-audit-dir\") pod \"apiserver-7bbb656c7d-j26br\" (UID: \"297ee903-c479-460e-80df-a0c508c45fbe\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j26br" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.086929 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c94992de-685b-42c3-8100-93996239d0b0-config\") pod \"etcd-operator-b45778765-9bs98\" (UID: \"c94992de-685b-42c3-8100-93996239d0b0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9bs98" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.086953 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9a0a9cf5-14e9-41e1-8c6c-27b7b3b8c7b4-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-8sxfn\" (UID: \"9a0a9cf5-14e9-41e1-8c6c-27b7b3b8c7b4\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8sxfn" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.086986 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/297ee903-c479-460e-80df-a0c508c45fbe-etcd-client\") pod \"apiserver-7bbb656c7d-j26br\" (UID: \"297ee903-c479-460e-80df-a0c508c45fbe\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j26br" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.087014 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/85bc0929-de17-4d2d-8b03-1a05d207fb55-metrics-certs\") pod \"router-default-5444994796-h5l29\" (UID: \"85bc0929-de17-4d2d-8b03-1a05d207fb55\") " pod="openshift-ingress/router-default-5444994796-h5l29" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.089564 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ffff2831-338e-423e-81f2-c8d3b5de7785-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-z587m\" (UID: \"ffff2831-338e-423e-81f2-c8d3b5de7785\") " pod="openshift-authentication/oauth-openshift-558db77b4-z587m" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.089591 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ffff2831-338e-423e-81f2-c8d3b5de7785-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-z587m\" (UID: \"ffff2831-338e-423e-81f2-c8d3b5de7785\") " pod="openshift-authentication/oauth-openshift-558db77b4-z587m" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.089827 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f086c3a7-c2a4-4066-aed6-270b5dd6eef8-serving-cert\") pod \"authentication-operator-69f744f599-6xp6f\" (UID: \"f086c3a7-c2a4-4066-aed6-270b5dd6eef8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6xp6f" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.089885 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3cef2105-ae00-4224-bf5d-246393f3caa6-serving-cert\") pod \"controller-manager-879f6c89f-qjcjv\" (UID: \"3cef2105-ae00-4224-bf5d-246393f3caa6\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qjcjv" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.090114 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/9a0a9cf5-14e9-41e1-8c6c-27b7b3b8c7b4-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-8sxfn\" (UID: \"9a0a9cf5-14e9-41e1-8c6c-27b7b3b8c7b4\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8sxfn" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.090323 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/297ee903-c479-460e-80df-a0c508c45fbe-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-j26br\" (UID: \"297ee903-c479-460e-80df-a0c508c45fbe\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j26br" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.090611 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ffff2831-338e-423e-81f2-c8d3b5de7785-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-z587m\" (UID: \"ffff2831-338e-423e-81f2-c8d3b5de7785\") " pod="openshift-authentication/oauth-openshift-558db77b4-z587m" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.090899 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3cef2105-ae00-4224-bf5d-246393f3caa6-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-qjcjv\" (UID: \"3cef2105-ae00-4224-bf5d-246393f3caa6\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qjcjv" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.091066 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f086c3a7-c2a4-4066-aed6-270b5dd6eef8-config\") pod \"authentication-operator-69f744f599-6xp6f\" (UID: \"f086c3a7-c2a4-4066-aed6-270b5dd6eef8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6xp6f" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.091155 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/fc892d74-c6d2-431f-a52c-95a8f4e2d03c-metrics-tls\") pod \"dns-operator-744455d44c-zp6l4\" (UID: \"fc892d74-c6d2-431f-a52c-95a8f4e2d03c\") " pod="openshift-dns-operator/dns-operator-744455d44c-zp6l4" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.091183 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/915af7e7-19b0-4b1c-9e55-3471210ccf40-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zjwcl\" (UID: \"915af7e7-19b0-4b1c-9e55-3471210ccf40\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zjwcl" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.091230 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f7305ef-d76f-4be3-9151-7247f7590fe0-config\") pod \"apiserver-76f77b778f-gc4cp\" (UID: \"6f7305ef-d76f-4be3-9151-7247f7590fe0\") " pod="openshift-apiserver/apiserver-76f77b778f-gc4cp" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.091284 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/92724c2e-d7fd-4673-ae84-8929541f2a13-bound-sa-token\") pod \"ingress-operator-5b745b69d9-wrw67\" (UID: \"92724c2e-d7fd-4673-ae84-8929541f2a13\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wrw67" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.091294 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/297ee903-c479-460e-80df-a0c508c45fbe-audit-policies\") pod \"apiserver-7bbb656c7d-j26br\" (UID: \"297ee903-c479-460e-80df-a0c508c45fbe\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j26br" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.091308 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/cbf3f592-bbb1-4596-8ae1-0e62aa366f0b-console-serving-cert\") pod \"console-f9d7485db-ndp5v\" (UID: \"cbf3f592-bbb1-4596-8ae1-0e62aa366f0b\") " pod="openshift-console/console-f9d7485db-ndp5v" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.091335 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-clrp8\" (UniqueName: \"kubernetes.io/projected/9a0a9cf5-14e9-41e1-8c6c-27b7b3b8c7b4-kube-api-access-clrp8\") pod \"cluster-image-registry-operator-dc59b4c8b-8sxfn\" (UID: \"9a0a9cf5-14e9-41e1-8c6c-27b7b3b8c7b4\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8sxfn" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.091356 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/297ee903-c479-460e-80df-a0c508c45fbe-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-j26br\" (UID: \"297ee903-c479-460e-80df-a0c508c45fbe\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j26br" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.091376 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k9ksj\" (UniqueName: \"kubernetes.io/projected/51fb93ab-ea9f-4fce-9fd1-a57958bc9c10-kube-api-access-k9ksj\") pod \"console-operator-58897d9998-pszqv\" (UID: \"51fb93ab-ea9f-4fce-9fd1-a57958bc9c10\") " pod="openshift-console-operator/console-operator-58897d9998-pszqv" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.091402 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/297ee903-c479-460e-80df-a0c508c45fbe-audit-dir\") pod \"apiserver-7bbb656c7d-j26br\" (UID: \"297ee903-c479-460e-80df-a0c508c45fbe\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j26br" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.091688 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ffff2831-338e-423e-81f2-c8d3b5de7785-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-z587m\" (UID: \"ffff2831-338e-423e-81f2-c8d3b5de7785\") " pod="openshift-authentication/oauth-openshift-558db77b4-z587m" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.091965 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/297ee903-c479-460e-80df-a0c508c45fbe-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-j26br\" (UID: \"297ee903-c479-460e-80df-a0c508c45fbe\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j26br" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.092004 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f7305ef-d76f-4be3-9151-7247f7590fe0-config\") pod \"apiserver-76f77b778f-gc4cp\" (UID: \"6f7305ef-d76f-4be3-9151-7247f7590fe0\") " pod="openshift-apiserver/apiserver-76f77b778f-gc4cp" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.092047 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4nhkp\" (UniqueName: \"kubernetes.io/projected/6f7305ef-d76f-4be3-9151-7247f7590fe0-kube-api-access-4nhkp\") pod \"apiserver-76f77b778f-gc4cp\" (UID: \"6f7305ef-d76f-4be3-9151-7247f7590fe0\") " pod="openshift-apiserver/apiserver-76f77b778f-gc4cp" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.092090 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-knwnm\" (UniqueName: \"kubernetes.io/projected/23992ff8-b006-40d1-9048-1950f10fff7d-kube-api-access-knwnm\") pod \"openshift-config-operator-7777fb866f-x2nd8\" (UID: \"23992ff8-b006-40d1-9048-1950f10fff7d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-x2nd8" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.092116 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/4df2ba74-d18e-4e94-bee1-703645b1a463-images\") pod \"machine-api-operator-5694c8668f-j7zhr\" (UID: \"4df2ba74-d18e-4e94-bee1-703645b1a463\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j7zhr" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.092183 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/6e5a490b-b9b3-4f19-aebf-14f6c1efbaa2-machine-approver-tls\") pod \"machine-approver-56656f9798-7ff64\" (UID: \"6e5a490b-b9b3-4f19-aebf-14f6c1efbaa2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7ff64" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.092209 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ffff2831-338e-423e-81f2-c8d3b5de7785-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-z587m\" (UID: \"ffff2831-338e-423e-81f2-c8d3b5de7785\") " pod="openshift-authentication/oauth-openshift-558db77b4-z587m" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.092240 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/cf1baeba-a0c2-4d91-ae76-eb4f0ceccd60-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-jjn8w\" (UID: \"cf1baeba-a0c2-4d91-ae76-eb4f0ceccd60\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-jjn8w" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.092283 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/cbf3f592-bbb1-4596-8ae1-0e62aa366f0b-console-config\") pod \"console-f9d7485db-ndp5v\" (UID: \"cbf3f592-bbb1-4596-8ae1-0e62aa366f0b\") " pod="openshift-console/console-f9d7485db-ndp5v" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.092302 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/cbf3f592-bbb1-4596-8ae1-0e62aa366f0b-service-ca\") pod \"console-f9d7485db-ndp5v\" (UID: \"cbf3f592-bbb1-4596-8ae1-0e62aa366f0b\") " pod="openshift-console/console-f9d7485db-ndp5v" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.092328 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xlqp9\" (UniqueName: \"kubernetes.io/projected/f086c3a7-c2a4-4066-aed6-270b5dd6eef8-kube-api-access-xlqp9\") pod \"authentication-operator-69f744f599-6xp6f\" (UID: \"f086c3a7-c2a4-4066-aed6-270b5dd6eef8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6xp6f" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.092350 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/4df2ba74-d18e-4e94-bee1-703645b1a463-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-j7zhr\" (UID: \"4df2ba74-d18e-4e94-bee1-703645b1a463\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j7zhr" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.092371 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51fb93ab-ea9f-4fce-9fd1-a57958bc9c10-config\") pod \"console-operator-58897d9998-pszqv\" (UID: \"51fb93ab-ea9f-4fce-9fd1-a57958bc9c10\") " pod="openshift-console-operator/console-operator-58897d9998-pszqv" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.092392 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9a0a9cf5-14e9-41e1-8c6c-27b7b3b8c7b4-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-8sxfn\" (UID: \"9a0a9cf5-14e9-41e1-8c6c-27b7b3b8c7b4\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8sxfn" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.092415 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/6f7305ef-d76f-4be3-9151-7247f7590fe0-etcd-client\") pod \"apiserver-76f77b778f-gc4cp\" (UID: \"6f7305ef-d76f-4be3-9151-7247f7590fe0\") " pod="openshift-apiserver/apiserver-76f77b778f-gc4cp" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.092496 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qlvv6\" (UniqueName: \"kubernetes.io/projected/97d9c075-80c3-4dcb-aebc-649eec930413-kube-api-access-qlvv6\") pod \"route-controller-manager-6576b87f9c-4d2qh\" (UID: \"97d9c075-80c3-4dcb-aebc-649eec930413\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4d2qh" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.092517 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6e5a490b-b9b3-4f19-aebf-14f6c1efbaa2-auth-proxy-config\") pod \"machine-approver-56656f9798-7ff64\" (UID: \"6e5a490b-b9b3-4f19-aebf-14f6c1efbaa2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7ff64" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.092545 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ffff2831-338e-423e-81f2-c8d3b5de7785-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-z587m\" (UID: \"ffff2831-338e-423e-81f2-c8d3b5de7785\") " pod="openshift-authentication/oauth-openshift-558db77b4-z587m" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.092565 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/297ee903-c479-460e-80df-a0c508c45fbe-serving-cert\") pod \"apiserver-7bbb656c7d-j26br\" (UID: \"297ee903-c479-460e-80df-a0c508c45fbe\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j26br" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.092587 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6rqws\" (UniqueName: \"kubernetes.io/projected/a51dadb3-2af5-4c6c-a973-cdceb34cd79a-kube-api-access-6rqws\") pod \"openshift-apiserver-operator-796bbdcf4f-bcfw4\" (UID: \"a51dadb3-2af5-4c6c-a973-cdceb34cd79a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bcfw4" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.092607 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c94992de-685b-42c3-8100-93996239d0b0-serving-cert\") pod \"etcd-operator-b45778765-9bs98\" (UID: \"c94992de-685b-42c3-8100-93996239d0b0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9bs98" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.092690 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ffff2831-338e-423e-81f2-c8d3b5de7785-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-z587m\" (UID: \"ffff2831-338e-423e-81f2-c8d3b5de7785\") " pod="openshift-authentication/oauth-openshift-558db77b4-z587m" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.092736 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6f7305ef-d76f-4be3-9151-7247f7590fe0-audit-dir\") pod \"apiserver-76f77b778f-gc4cp\" (UID: \"6f7305ef-d76f-4be3-9151-7247f7590fe0\") " pod="openshift-apiserver/apiserver-76f77b778f-gc4cp" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.093351 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ffff2831-338e-423e-81f2-c8d3b5de7785-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-z587m\" (UID: \"ffff2831-338e-423e-81f2-c8d3b5de7785\") " pod="openshift-authentication/oauth-openshift-558db77b4-z587m" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.093498 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/4df2ba74-d18e-4e94-bee1-703645b1a463-images\") pod \"machine-api-operator-5694c8668f-j7zhr\" (UID: \"4df2ba74-d18e-4e94-bee1-703645b1a463\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j7zhr" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.093687 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6f7305ef-d76f-4be3-9151-7247f7590fe0-serving-cert\") pod \"apiserver-76f77b778f-gc4cp\" (UID: \"6f7305ef-d76f-4be3-9151-7247f7590fe0\") " pod="openshift-apiserver/apiserver-76f77b778f-gc4cp" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.093917 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/23992ff8-b006-40d1-9048-1950f10fff7d-serving-cert\") pod \"openshift-config-operator-7777fb866f-x2nd8\" (UID: \"23992ff8-b006-40d1-9048-1950f10fff7d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-x2nd8" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.094130 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ffff2831-338e-423e-81f2-c8d3b5de7785-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-z587m\" (UID: \"ffff2831-338e-423e-81f2-c8d3b5de7785\") " pod="openshift-authentication/oauth-openshift-558db77b4-z587m" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.094235 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/6f7305ef-d76f-4be3-9151-7247f7590fe0-image-import-ca\") pod \"apiserver-76f77b778f-gc4cp\" (UID: \"6f7305ef-d76f-4be3-9151-7247f7590fe0\") " pod="openshift-apiserver/apiserver-76f77b778f-gc4cp" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.094303 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c94992de-685b-42c3-8100-93996239d0b0-etcd-client\") pod \"etcd-operator-b45778765-9bs98\" (UID: \"c94992de-685b-42c3-8100-93996239d0b0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9bs98" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.094354 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/c94992de-685b-42c3-8100-93996239d0b0-etcd-ca\") pod \"etcd-operator-b45778765-9bs98\" (UID: \"c94992de-685b-42c3-8100-93996239d0b0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9bs98" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.094387 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/51fb93ab-ea9f-4fce-9fd1-a57958bc9c10-serving-cert\") pod \"console-operator-58897d9998-pszqv\" (UID: \"51fb93ab-ea9f-4fce-9fd1-a57958bc9c10\") " pod="openshift-console-operator/console-operator-58897d9998-pszqv" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.094452 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ffff2831-338e-423e-81f2-c8d3b5de7785-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-z587m\" (UID: \"ffff2831-338e-423e-81f2-c8d3b5de7785\") " pod="openshift-authentication/oauth-openshift-558db77b4-z587m" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.094476 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ffff2831-338e-423e-81f2-c8d3b5de7785-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-z587m\" (UID: \"ffff2831-338e-423e-81f2-c8d3b5de7785\") " pod="openshift-authentication/oauth-openshift-558db77b4-z587m" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.094491 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/92724c2e-d7fd-4673-ae84-8929541f2a13-metrics-tls\") pod \"ingress-operator-5b745b69d9-wrw67\" (UID: \"92724c2e-d7fd-4673-ae84-8929541f2a13\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wrw67" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.094604 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ffff2831-338e-423e-81f2-c8d3b5de7785-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-z587m\" (UID: \"ffff2831-338e-423e-81f2-c8d3b5de7785\") " pod="openshift-authentication/oauth-openshift-558db77b4-z587m" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.095151 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6e5a490b-b9b3-4f19-aebf-14f6c1efbaa2-auth-proxy-config\") pod \"machine-approver-56656f9798-7ff64\" (UID: \"6e5a490b-b9b3-4f19-aebf-14f6c1efbaa2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7ff64" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.095564 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/6f7305ef-d76f-4be3-9151-7247f7590fe0-image-import-ca\") pod \"apiserver-76f77b778f-gc4cp\" (UID: \"6f7305ef-d76f-4be3-9151-7247f7590fe0\") " pod="openshift-apiserver/apiserver-76f77b778f-gc4cp" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.095690 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9a0a9cf5-14e9-41e1-8c6c-27b7b3b8c7b4-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-8sxfn\" (UID: \"9a0a9cf5-14e9-41e1-8c6c-27b7b3b8c7b4\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8sxfn" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.096251 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/6f7305ef-d76f-4be3-9151-7247f7590fe0-encryption-config\") pod \"apiserver-76f77b778f-gc4cp\" (UID: \"6f7305ef-d76f-4be3-9151-7247f7590fe0\") " pod="openshift-apiserver/apiserver-76f77b778f-gc4cp" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.096355 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/6e5a490b-b9b3-4f19-aebf-14f6c1efbaa2-machine-approver-tls\") pod \"machine-approver-56656f9798-7ff64\" (UID: \"6e5a490b-b9b3-4f19-aebf-14f6c1efbaa2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7ff64" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.096669 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/297ee903-c479-460e-80df-a0c508c45fbe-encryption-config\") pod \"apiserver-7bbb656c7d-j26br\" (UID: \"297ee903-c479-460e-80df-a0c508c45fbe\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j26br" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.096836 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/4df2ba74-d18e-4e94-bee1-703645b1a463-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-j7zhr\" (UID: \"4df2ba74-d18e-4e94-bee1-703645b1a463\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j7zhr" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.097439 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/297ee903-c479-460e-80df-a0c508c45fbe-serving-cert\") pod \"apiserver-7bbb656c7d-j26br\" (UID: \"297ee903-c479-460e-80df-a0c508c45fbe\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j26br" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.097472 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/6f7305ef-d76f-4be3-9151-7247f7590fe0-etcd-client\") pod \"apiserver-76f77b778f-gc4cp\" (UID: \"6f7305ef-d76f-4be3-9151-7247f7590fe0\") " pod="openshift-apiserver/apiserver-76f77b778f-gc4cp" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.098124 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ffff2831-338e-423e-81f2-c8d3b5de7785-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-z587m\" (UID: \"ffff2831-338e-423e-81f2-c8d3b5de7785\") " pod="openshift-authentication/oauth-openshift-558db77b4-z587m" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.098832 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/297ee903-c479-460e-80df-a0c508c45fbe-etcd-client\") pod \"apiserver-7bbb656c7d-j26br\" (UID: \"297ee903-c479-460e-80df-a0c508c45fbe\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j26br" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.101603 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/97d9c075-80c3-4dcb-aebc-649eec930413-serving-cert\") pod \"route-controller-manager-6576b87f9c-4d2qh\" (UID: \"97d9c075-80c3-4dcb-aebc-649eec930413\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4d2qh" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.107568 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.127032 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.145954 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.167208 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.187438 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.195262 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bp2f8\" (UniqueName: \"kubernetes.io/projected/92724c2e-d7fd-4673-ae84-8929541f2a13-kube-api-access-bp2f8\") pod \"ingress-operator-5b745b69d9-wrw67\" (UID: \"92724c2e-d7fd-4673-ae84-8929541f2a13\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wrw67" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.195305 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/85bc0929-de17-4d2d-8b03-1a05d207fb55-default-certificate\") pod \"router-default-5444994796-h5l29\" (UID: \"85bc0929-de17-4d2d-8b03-1a05d207fb55\") " pod="openshift-ingress/router-default-5444994796-h5l29" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.195326 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/85bc0929-de17-4d2d-8b03-1a05d207fb55-service-ca-bundle\") pod \"router-default-5444994796-h5l29\" (UID: \"85bc0929-de17-4d2d-8b03-1a05d207fb55\") " pod="openshift-ingress/router-default-5444994796-h5l29" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.195358 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/92724c2e-d7fd-4673-ae84-8929541f2a13-trusted-ca\") pod \"ingress-operator-5b745b69d9-wrw67\" (UID: \"92724c2e-d7fd-4673-ae84-8929541f2a13\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wrw67" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.195378 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8439dba6-4f32-4c02-9aad-27b9c40f7575-config\") pod \"kube-controller-manager-operator-78b949d7b-d49gn\" (UID: \"8439dba6-4f32-4c02-9aad-27b9c40f7575\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-d49gn" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.195398 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vcjxn\" (UniqueName: \"kubernetes.io/projected/c94992de-685b-42c3-8100-93996239d0b0-kube-api-access-vcjxn\") pod \"etcd-operator-b45778765-9bs98\" (UID: \"c94992de-685b-42c3-8100-93996239d0b0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9bs98" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.195422 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmljs\" (UniqueName: \"kubernetes.io/projected/cbf3f592-bbb1-4596-8ae1-0e62aa366f0b-kube-api-access-tmljs\") pod \"console-f9d7485db-ndp5v\" (UID: \"cbf3f592-bbb1-4596-8ae1-0e62aa366f0b\") " pod="openshift-console/console-f9d7485db-ndp5v" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.195456 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c94992de-685b-42c3-8100-93996239d0b0-config\") pod \"etcd-operator-b45778765-9bs98\" (UID: \"c94992de-685b-42c3-8100-93996239d0b0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9bs98" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.195479 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/85bc0929-de17-4d2d-8b03-1a05d207fb55-metrics-certs\") pod \"router-default-5444994796-h5l29\" (UID: \"85bc0929-de17-4d2d-8b03-1a05d207fb55\") " pod="openshift-ingress/router-default-5444994796-h5l29" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.195499 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/fc892d74-c6d2-431f-a52c-95a8f4e2d03c-metrics-tls\") pod \"dns-operator-744455d44c-zp6l4\" (UID: \"fc892d74-c6d2-431f-a52c-95a8f4e2d03c\") " pod="openshift-dns-operator/dns-operator-744455d44c-zp6l4" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.195520 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/915af7e7-19b0-4b1c-9e55-3471210ccf40-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zjwcl\" (UID: \"915af7e7-19b0-4b1c-9e55-3471210ccf40\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zjwcl" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.196605 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/92724c2e-d7fd-4673-ae84-8929541f2a13-bound-sa-token\") pod \"ingress-operator-5b745b69d9-wrw67\" (UID: \"92724c2e-d7fd-4673-ae84-8929541f2a13\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wrw67" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.196641 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/cbf3f592-bbb1-4596-8ae1-0e62aa366f0b-console-serving-cert\") pod \"console-f9d7485db-ndp5v\" (UID: \"cbf3f592-bbb1-4596-8ae1-0e62aa366f0b\") " pod="openshift-console/console-f9d7485db-ndp5v" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.196691 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k9ksj\" (UniqueName: \"kubernetes.io/projected/51fb93ab-ea9f-4fce-9fd1-a57958bc9c10-kube-api-access-k9ksj\") pod \"console-operator-58897d9998-pszqv\" (UID: \"51fb93ab-ea9f-4fce-9fd1-a57958bc9c10\") " pod="openshift-console-operator/console-operator-58897d9998-pszqv" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.196721 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/cf1baeba-a0c2-4d91-ae76-eb4f0ceccd60-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-jjn8w\" (UID: \"cf1baeba-a0c2-4d91-ae76-eb4f0ceccd60\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-jjn8w" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.196758 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51fb93ab-ea9f-4fce-9fd1-a57958bc9c10-config\") pod \"console-operator-58897d9998-pszqv\" (UID: \"51fb93ab-ea9f-4fce-9fd1-a57958bc9c10\") " pod="openshift-console-operator/console-operator-58897d9998-pszqv" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.196785 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/cbf3f592-bbb1-4596-8ae1-0e62aa366f0b-console-config\") pod \"console-f9d7485db-ndp5v\" (UID: \"cbf3f592-bbb1-4596-8ae1-0e62aa366f0b\") " pod="openshift-console/console-f9d7485db-ndp5v" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.196812 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/cbf3f592-bbb1-4596-8ae1-0e62aa366f0b-service-ca\") pod \"console-f9d7485db-ndp5v\" (UID: \"cbf3f592-bbb1-4596-8ae1-0e62aa366f0b\") " pod="openshift-console/console-f9d7485db-ndp5v" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.196845 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c94992de-685b-42c3-8100-93996239d0b0-serving-cert\") pod \"etcd-operator-b45778765-9bs98\" (UID: \"c94992de-685b-42c3-8100-93996239d0b0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9bs98" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.196872 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6rqws\" (UniqueName: \"kubernetes.io/projected/a51dadb3-2af5-4c6c-a973-cdceb34cd79a-kube-api-access-6rqws\") pod \"openshift-apiserver-operator-796bbdcf4f-bcfw4\" (UID: \"a51dadb3-2af5-4c6c-a973-cdceb34cd79a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bcfw4" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.196896 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c94992de-685b-42c3-8100-93996239d0b0-etcd-client\") pod \"etcd-operator-b45778765-9bs98\" (UID: \"c94992de-685b-42c3-8100-93996239d0b0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9bs98" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.196918 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/92724c2e-d7fd-4673-ae84-8929541f2a13-metrics-tls\") pod \"ingress-operator-5b745b69d9-wrw67\" (UID: \"92724c2e-d7fd-4673-ae84-8929541f2a13\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wrw67" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.196941 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/c94992de-685b-42c3-8100-93996239d0b0-etcd-ca\") pod \"etcd-operator-b45778765-9bs98\" (UID: \"c94992de-685b-42c3-8100-93996239d0b0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9bs98" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.196965 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/51fb93ab-ea9f-4fce-9fd1-a57958bc9c10-serving-cert\") pod \"console-operator-58897d9998-pszqv\" (UID: \"51fb93ab-ea9f-4fce-9fd1-a57958bc9c10\") " pod="openshift-console-operator/console-operator-58897d9998-pszqv" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.196990 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8439dba6-4f32-4c02-9aad-27b9c40f7575-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-d49gn\" (UID: \"8439dba6-4f32-4c02-9aad-27b9c40f7575\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-d49gn" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.197248 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/51fb93ab-ea9f-4fce-9fd1-a57958bc9c10-trusted-ca\") pod \"console-operator-58897d9998-pszqv\" (UID: \"51fb93ab-ea9f-4fce-9fd1-a57958bc9c10\") " pod="openshift-console-operator/console-operator-58897d9998-pszqv" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.197282 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a51dadb3-2af5-4c6c-a973-cdceb34cd79a-config\") pod \"openshift-apiserver-operator-796bbdcf4f-bcfw4\" (UID: \"a51dadb3-2af5-4c6c-a973-cdceb34cd79a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bcfw4" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.197345 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/cbf3f592-bbb1-4596-8ae1-0e62aa366f0b-oauth-serving-cert\") pod \"console-f9d7485db-ndp5v\" (UID: \"cbf3f592-bbb1-4596-8ae1-0e62aa366f0b\") " pod="openshift-console/console-f9d7485db-ndp5v" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.197393 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cbf3f592-bbb1-4596-8ae1-0e62aa366f0b-trusted-ca-bundle\") pod \"console-f9d7485db-ndp5v\" (UID: \"cbf3f592-bbb1-4596-8ae1-0e62aa366f0b\") " pod="openshift-console/console-f9d7485db-ndp5v" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.197417 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/c94992de-685b-42c3-8100-93996239d0b0-etcd-service-ca\") pod \"etcd-operator-b45778765-9bs98\" (UID: \"c94992de-685b-42c3-8100-93996239d0b0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9bs98" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.197445 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97b4v\" (UniqueName: \"kubernetes.io/projected/fc892d74-c6d2-431f-a52c-95a8f4e2d03c-kube-api-access-97b4v\") pod \"dns-operator-744455d44c-zp6l4\" (UID: \"fc892d74-c6d2-431f-a52c-95a8f4e2d03c\") " pod="openshift-dns-operator/dns-operator-744455d44c-zp6l4" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.197469 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fw2t8\" (UniqueName: \"kubernetes.io/projected/cf1baeba-a0c2-4d91-ae76-eb4f0ceccd60-kube-api-access-fw2t8\") pod \"cluster-samples-operator-665b6dd947-jjn8w\" (UID: \"cf1baeba-a0c2-4d91-ae76-eb4f0ceccd60\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-jjn8w" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.197506 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/85bc0929-de17-4d2d-8b03-1a05d207fb55-stats-auth\") pod \"router-default-5444994796-h5l29\" (UID: \"85bc0929-de17-4d2d-8b03-1a05d207fb55\") " pod="openshift-ingress/router-default-5444994796-h5l29" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.197537 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9d47p\" (UniqueName: \"kubernetes.io/projected/1306f3b2-2982-494e-9db0-86c2e05b7a82-kube-api-access-9d47p\") pod \"downloads-7954f5f757-b257c\" (UID: \"1306f3b2-2982-494e-9db0-86c2e05b7a82\") " pod="openshift-console/downloads-7954f5f757-b257c" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.197564 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8439dba6-4f32-4c02-9aad-27b9c40f7575-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-d49gn\" (UID: \"8439dba6-4f32-4c02-9aad-27b9c40f7575\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-d49gn" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.197588 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/915af7e7-19b0-4b1c-9e55-3471210ccf40-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zjwcl\" (UID: \"915af7e7-19b0-4b1c-9e55-3471210ccf40\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zjwcl" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.197610 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/cbf3f592-bbb1-4596-8ae1-0e62aa366f0b-console-oauth-config\") pod \"console-f9d7485db-ndp5v\" (UID: \"cbf3f592-bbb1-4596-8ae1-0e62aa366f0b\") " pod="openshift-console/console-f9d7485db-ndp5v" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.197639 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a51dadb3-2af5-4c6c-a973-cdceb34cd79a-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-bcfw4\" (UID: \"a51dadb3-2af5-4c6c-a973-cdceb34cd79a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bcfw4" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.197649 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c94992de-685b-42c3-8100-93996239d0b0-config\") pod \"etcd-operator-b45778765-9bs98\" (UID: \"c94992de-685b-42c3-8100-93996239d0b0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9bs98" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.197667 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hk885\" (UniqueName: \"kubernetes.io/projected/85bc0929-de17-4d2d-8b03-1a05d207fb55-kube-api-access-hk885\") pod \"router-default-5444994796-h5l29\" (UID: \"85bc0929-de17-4d2d-8b03-1a05d207fb55\") " pod="openshift-ingress/router-default-5444994796-h5l29" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.197872 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/cbf3f592-bbb1-4596-8ae1-0e62aa366f0b-console-config\") pod \"console-f9d7485db-ndp5v\" (UID: \"cbf3f592-bbb1-4596-8ae1-0e62aa366f0b\") " pod="openshift-console/console-f9d7485db-ndp5v" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.197888 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/915af7e7-19b0-4b1c-9e55-3471210ccf40-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zjwcl\" (UID: \"915af7e7-19b0-4b1c-9e55-3471210ccf40\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zjwcl" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.198497 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/915af7e7-19b0-4b1c-9e55-3471210ccf40-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zjwcl\" (UID: \"915af7e7-19b0-4b1c-9e55-3471210ccf40\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zjwcl" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.198956 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a51dadb3-2af5-4c6c-a973-cdceb34cd79a-config\") pod \"openshift-apiserver-operator-796bbdcf4f-bcfw4\" (UID: \"a51dadb3-2af5-4c6c-a973-cdceb34cd79a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bcfw4" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.199098 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/cbf3f592-bbb1-4596-8ae1-0e62aa366f0b-service-ca\") pod \"console-f9d7485db-ndp5v\" (UID: \"cbf3f592-bbb1-4596-8ae1-0e62aa366f0b\") " pod="openshift-console/console-f9d7485db-ndp5v" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.199235 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51fb93ab-ea9f-4fce-9fd1-a57958bc9c10-config\") pod \"console-operator-58897d9998-pszqv\" (UID: \"51fb93ab-ea9f-4fce-9fd1-a57958bc9c10\") " pod="openshift-console-operator/console-operator-58897d9998-pszqv" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.199370 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/cbf3f592-bbb1-4596-8ae1-0e62aa366f0b-oauth-serving-cert\") pod \"console-f9d7485db-ndp5v\" (UID: \"cbf3f592-bbb1-4596-8ae1-0e62aa366f0b\") " pod="openshift-console/console-f9d7485db-ndp5v" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.199377 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/c94992de-685b-42c3-8100-93996239d0b0-etcd-ca\") pod \"etcd-operator-b45778765-9bs98\" (UID: \"c94992de-685b-42c3-8100-93996239d0b0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9bs98" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.200623 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/c94992de-685b-42c3-8100-93996239d0b0-etcd-service-ca\") pod \"etcd-operator-b45778765-9bs98\" (UID: \"c94992de-685b-42c3-8100-93996239d0b0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9bs98" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.200930 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/51fb93ab-ea9f-4fce-9fd1-a57958bc9c10-trusted-ca\") pod \"console-operator-58897d9998-pszqv\" (UID: \"51fb93ab-ea9f-4fce-9fd1-a57958bc9c10\") " pod="openshift-console-operator/console-operator-58897d9998-pszqv" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.201462 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cbf3f592-bbb1-4596-8ae1-0e62aa366f0b-trusted-ca-bundle\") pod \"console-f9d7485db-ndp5v\" (UID: \"cbf3f592-bbb1-4596-8ae1-0e62aa366f0b\") " pod="openshift-console/console-f9d7485db-ndp5v" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.203234 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/51fb93ab-ea9f-4fce-9fd1-a57958bc9c10-serving-cert\") pod \"console-operator-58897d9998-pszqv\" (UID: \"51fb93ab-ea9f-4fce-9fd1-a57958bc9c10\") " pod="openshift-console-operator/console-operator-58897d9998-pszqv" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.203514 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/cbf3f592-bbb1-4596-8ae1-0e62aa366f0b-console-oauth-config\") pod \"console-f9d7485db-ndp5v\" (UID: \"cbf3f592-bbb1-4596-8ae1-0e62aa366f0b\") " pod="openshift-console/console-f9d7485db-ndp5v" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.208700 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.211196 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c94992de-685b-42c3-8100-93996239d0b0-serving-cert\") pod \"etcd-operator-b45778765-9bs98\" (UID: \"c94992de-685b-42c3-8100-93996239d0b0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9bs98" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.211484 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c94992de-685b-42c3-8100-93996239d0b0-etcd-client\") pod \"etcd-operator-b45778765-9bs98\" (UID: \"c94992de-685b-42c3-8100-93996239d0b0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9bs98" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.212576 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/cf1baeba-a0c2-4d91-ae76-eb4f0ceccd60-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-jjn8w\" (UID: \"cf1baeba-a0c2-4d91-ae76-eb4f0ceccd60\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-jjn8w" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.212578 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/cbf3f592-bbb1-4596-8ae1-0e62aa366f0b-console-serving-cert\") pod \"console-f9d7485db-ndp5v\" (UID: \"cbf3f592-bbb1-4596-8ae1-0e62aa366f0b\") " pod="openshift-console/console-f9d7485db-ndp5v" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.213135 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/915af7e7-19b0-4b1c-9e55-3471210ccf40-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zjwcl\" (UID: \"915af7e7-19b0-4b1c-9e55-3471210ccf40\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zjwcl" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.227200 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.241570 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/fc892d74-c6d2-431f-a52c-95a8f4e2d03c-metrics-tls\") pod \"dns-operator-744455d44c-zp6l4\" (UID: \"fc892d74-c6d2-431f-a52c-95a8f4e2d03c\") " pod="openshift-dns-operator/dns-operator-744455d44c-zp6l4" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.246714 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.266529 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.274215 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a51dadb3-2af5-4c6c-a973-cdceb34cd79a-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-bcfw4\" (UID: \"a51dadb3-2af5-4c6c-a973-cdceb34cd79a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bcfw4" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.286493 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.306168 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.319666 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/85bc0929-de17-4d2d-8b03-1a05d207fb55-default-certificate\") pod \"router-default-5444994796-h5l29\" (UID: \"85bc0929-de17-4d2d-8b03-1a05d207fb55\") " pod="openshift-ingress/router-default-5444994796-h5l29" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.327561 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.333295 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/85bc0929-de17-4d2d-8b03-1a05d207fb55-stats-auth\") pod \"router-default-5444994796-h5l29\" (UID: \"85bc0929-de17-4d2d-8b03-1a05d207fb55\") " pod="openshift-ingress/router-default-5444994796-h5l29" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.346587 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.347515 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/85bc0929-de17-4d2d-8b03-1a05d207fb55-service-ca-bundle\") pod \"router-default-5444994796-h5l29\" (UID: \"85bc0929-de17-4d2d-8b03-1a05d207fb55\") " pod="openshift-ingress/router-default-5444994796-h5l29" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.365999 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.388182 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.407884 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.421604 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/85bc0929-de17-4d2d-8b03-1a05d207fb55-metrics-certs\") pod \"router-default-5444994796-h5l29\" (UID: \"85bc0929-de17-4d2d-8b03-1a05d207fb55\") " pod="openshift-ingress/router-default-5444994796-h5l29" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.428668 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.447120 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.467582 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.506952 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.529461 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.547936 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.553764 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/92724c2e-d7fd-4673-ae84-8929541f2a13-metrics-tls\") pod \"ingress-operator-5b745b69d9-wrw67\" (UID: \"92724c2e-d7fd-4673-ae84-8929541f2a13\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wrw67" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.573875 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.578483 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/92724c2e-d7fd-4673-ae84-8929541f2a13-trusted-ca\") pod \"ingress-operator-5b745b69d9-wrw67\" (UID: \"92724c2e-d7fd-4673-ae84-8929541f2a13\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wrw67" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.588112 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.606704 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.628242 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.647642 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.657324 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.657337 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nx2lh" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.657701 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.662798 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8439dba6-4f32-4c02-9aad-27b9c40f7575-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-d49gn\" (UID: \"8439dba6-4f32-4c02-9aad-27b9c40f7575\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-d49gn" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.667173 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.676683 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8439dba6-4f32-4c02-9aad-27b9c40f7575-config\") pod \"kube-controller-manager-operator-78b949d7b-d49gn\" (UID: \"8439dba6-4f32-4c02-9aad-27b9c40f7575\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-d49gn" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.707302 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.728140 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.746670 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.767204 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.786984 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.806014 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.827939 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.847348 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.866558 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.886851 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.906473 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.926049 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.946056 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.967252 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 07 16:04:29 crc kubenswrapper[4716]: I1207 16:04:29.987300 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 07 16:04:30 crc kubenswrapper[4716]: I1207 16:04:30.007408 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 07 16:04:30 crc kubenswrapper[4716]: I1207 16:04:30.025213 4716 request.go:700] Waited for 1.017847595s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver-operator/configmaps?fieldSelector=metadata.name%3Dkube-apiserver-operator-config&limit=500&resourceVersion=0 Dec 07 16:04:30 crc kubenswrapper[4716]: I1207 16:04:30.026967 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 07 16:04:30 crc kubenswrapper[4716]: I1207 16:04:30.047013 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 07 16:04:30 crc kubenswrapper[4716]: I1207 16:04:30.066398 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 07 16:04:30 crc kubenswrapper[4716]: I1207 16:04:30.086743 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 07 16:04:30 crc kubenswrapper[4716]: I1207 16:04:30.106524 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 07 16:04:30 crc kubenswrapper[4716]: I1207 16:04:30.127643 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 07 16:04:30 crc kubenswrapper[4716]: I1207 16:04:30.147294 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 07 16:04:30 crc kubenswrapper[4716]: I1207 16:04:30.167983 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 07 16:04:30 crc kubenswrapper[4716]: I1207 16:04:30.187777 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 07 16:04:30 crc kubenswrapper[4716]: I1207 16:04:30.206274 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 07 16:04:30 crc kubenswrapper[4716]: I1207 16:04:30.227586 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 07 16:04:30 crc kubenswrapper[4716]: I1207 16:04:30.248156 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 07 16:04:30 crc kubenswrapper[4716]: I1207 16:04:30.266952 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 07 16:04:30 crc kubenswrapper[4716]: I1207 16:04:30.286935 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 07 16:04:30 crc kubenswrapper[4716]: I1207 16:04:30.307120 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 07 16:04:30 crc kubenswrapper[4716]: I1207 16:04:30.327414 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 07 16:04:30 crc kubenswrapper[4716]: I1207 16:04:30.346645 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 07 16:04:30 crc kubenswrapper[4716]: I1207 16:04:30.376933 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 07 16:04:30 crc kubenswrapper[4716]: I1207 16:04:30.387456 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 07 16:04:30 crc kubenswrapper[4716]: I1207 16:04:30.407215 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 07 16:04:30 crc kubenswrapper[4716]: I1207 16:04:30.426861 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 07 16:04:30 crc kubenswrapper[4716]: I1207 16:04:30.445834 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 07 16:04:30 crc kubenswrapper[4716]: I1207 16:04:30.467936 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 07 16:04:30 crc kubenswrapper[4716]: I1207 16:04:30.485978 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 07 16:04:30 crc kubenswrapper[4716]: I1207 16:04:30.505556 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 07 16:04:30 crc kubenswrapper[4716]: I1207 16:04:30.527917 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 07 16:04:30 crc kubenswrapper[4716]: I1207 16:04:30.546651 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 07 16:04:30 crc kubenswrapper[4716]: I1207 16:04:30.568176 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 07 16:04:30 crc kubenswrapper[4716]: I1207 16:04:30.586725 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 07 16:04:30 crc kubenswrapper[4716]: I1207 16:04:30.606768 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 07 16:04:30 crc kubenswrapper[4716]: I1207 16:04:30.628107 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 07 16:04:30 crc kubenswrapper[4716]: I1207 16:04:30.647422 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 07 16:04:30 crc kubenswrapper[4716]: I1207 16:04:30.667665 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 07 16:04:30 crc kubenswrapper[4716]: I1207 16:04:30.686960 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 07 16:04:30 crc kubenswrapper[4716]: I1207 16:04:30.707203 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 07 16:04:30 crc kubenswrapper[4716]: I1207 16:04:30.726321 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 07 16:04:30 crc kubenswrapper[4716]: I1207 16:04:30.747504 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 07 16:04:30 crc kubenswrapper[4716]: I1207 16:04:30.767071 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 07 16:04:30 crc kubenswrapper[4716]: I1207 16:04:30.786854 4716 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 07 16:04:30 crc kubenswrapper[4716]: I1207 16:04:30.806954 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 07 16:04:30 crc kubenswrapper[4716]: I1207 16:04:30.854865 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rvqph\" (UniqueName: \"kubernetes.io/projected/6e5a490b-b9b3-4f19-aebf-14f6c1efbaa2-kube-api-access-rvqph\") pod \"machine-approver-56656f9798-7ff64\" (UID: \"6e5a490b-b9b3-4f19-aebf-14f6c1efbaa2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7ff64" Dec 07 16:04:30 crc kubenswrapper[4716]: I1207 16:04:30.876000 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rhn6\" (UniqueName: \"kubernetes.io/projected/3cef2105-ae00-4224-bf5d-246393f3caa6-kube-api-access-8rhn6\") pod \"controller-manager-879f6c89f-qjcjv\" (UID: \"3cef2105-ae00-4224-bf5d-246393f3caa6\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qjcjv" Dec 07 16:04:30 crc kubenswrapper[4716]: I1207 16:04:30.887128 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9sqq9\" (UniqueName: \"kubernetes.io/projected/4df2ba74-d18e-4e94-bee1-703645b1a463-kube-api-access-9sqq9\") pod \"machine-api-operator-5694c8668f-j7zhr\" (UID: \"4df2ba74-d18e-4e94-bee1-703645b1a463\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j7zhr" Dec 07 16:04:30 crc kubenswrapper[4716]: I1207 16:04:30.907384 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 07 16:04:30 crc kubenswrapper[4716]: I1207 16:04:30.914685 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2tnlr\" (UniqueName: \"kubernetes.io/projected/297ee903-c479-460e-80df-a0c508c45fbe-kube-api-access-2tnlr\") pod \"apiserver-7bbb656c7d-j26br\" (UID: \"297ee903-c479-460e-80df-a0c508c45fbe\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j26br" Dec 07 16:04:30 crc kubenswrapper[4716]: I1207 16:04:30.927367 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 07 16:04:30 crc kubenswrapper[4716]: I1207 16:04:30.947397 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 07 16:04:30 crc kubenswrapper[4716]: I1207 16:04:30.980380 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-qjcjv" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.000640 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9a0a9cf5-14e9-41e1-8c6c-27b7b3b8c7b4-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-8sxfn\" (UID: \"9a0a9cf5-14e9-41e1-8c6c-27b7b3b8c7b4\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8sxfn" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.001743 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-clrp8\" (UniqueName: \"kubernetes.io/projected/9a0a9cf5-14e9-41e1-8c6c-27b7b3b8c7b4-kube-api-access-clrp8\") pod \"cluster-image-registry-operator-dc59b4c8b-8sxfn\" (UID: \"9a0a9cf5-14e9-41e1-8c6c-27b7b3b8c7b4\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8sxfn" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.017948 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j26br" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.034005 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-j7zhr" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.034480 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nw5sf\" (UniqueName: \"kubernetes.io/projected/ffff2831-338e-423e-81f2-c8d3b5de7785-kube-api-access-nw5sf\") pod \"oauth-openshift-558db77b4-z587m\" (UID: \"ffff2831-338e-423e-81f2-c8d3b5de7785\") " pod="openshift-authentication/oauth-openshift-558db77b4-z587m" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.034578 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7ff64" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.042601 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4nhkp\" (UniqueName: \"kubernetes.io/projected/6f7305ef-d76f-4be3-9151-7247f7590fe0-kube-api-access-4nhkp\") pod \"apiserver-76f77b778f-gc4cp\" (UID: \"6f7305ef-d76f-4be3-9151-7247f7590fe0\") " pod="openshift-apiserver/apiserver-76f77b778f-gc4cp" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.044825 4716 request.go:700] Waited for 1.951221985s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication-operator/serviceaccounts/authentication-operator/token Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.061415 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xlqp9\" (UniqueName: \"kubernetes.io/projected/f086c3a7-c2a4-4066-aed6-270b5dd6eef8-kube-api-access-xlqp9\") pod \"authentication-operator-69f744f599-6xp6f\" (UID: \"f086c3a7-c2a4-4066-aed6-270b5dd6eef8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6xp6f" Dec 07 16:04:31 crc kubenswrapper[4716]: W1207 16:04:31.072182 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6e5a490b_b9b3_4f19_aebf_14f6c1efbaa2.slice/crio-06dd01263babd531d4c9a62ea9245f915f9049e39f2cbaf72c231e2c8b3a6e59 WatchSource:0}: Error finding container 06dd01263babd531d4c9a62ea9245f915f9049e39f2cbaf72c231e2c8b3a6e59: Status 404 returned error can't find the container with id 06dd01263babd531d4c9a62ea9245f915f9049e39f2cbaf72c231e2c8b3a6e59 Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.081680 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-z587m" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.085129 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-knwnm\" (UniqueName: \"kubernetes.io/projected/23992ff8-b006-40d1-9048-1950f10fff7d-kube-api-access-knwnm\") pod \"openshift-config-operator-7777fb866f-x2nd8\" (UID: \"23992ff8-b006-40d1-9048-1950f10fff7d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-x2nd8" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.105381 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8sxfn" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.111705 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-x2nd8" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.119323 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qlvv6\" (UniqueName: \"kubernetes.io/projected/97d9c075-80c3-4dcb-aebc-649eec930413-kube-api-access-qlvv6\") pod \"route-controller-manager-6576b87f9c-4d2qh\" (UID: \"97d9c075-80c3-4dcb-aebc-649eec930413\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4d2qh" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.144331 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bp2f8\" (UniqueName: \"kubernetes.io/projected/92724c2e-d7fd-4673-ae84-8929541f2a13-kube-api-access-bp2f8\") pod \"ingress-operator-5b745b69d9-wrw67\" (UID: \"92724c2e-d7fd-4673-ae84-8929541f2a13\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wrw67" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.161238 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmljs\" (UniqueName: \"kubernetes.io/projected/cbf3f592-bbb1-4596-8ae1-0e62aa366f0b-kube-api-access-tmljs\") pod \"console-f9d7485db-ndp5v\" (UID: \"cbf3f592-bbb1-4596-8ae1-0e62aa366f0b\") " pod="openshift-console/console-f9d7485db-ndp5v" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.161719 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vcjxn\" (UniqueName: \"kubernetes.io/projected/c94992de-685b-42c3-8100-93996239d0b0-kube-api-access-vcjxn\") pod \"etcd-operator-b45778765-9bs98\" (UID: \"c94992de-685b-42c3-8100-93996239d0b0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9bs98" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.181302 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/92724c2e-d7fd-4673-ae84-8929541f2a13-bound-sa-token\") pod \"ingress-operator-5b745b69d9-wrw67\" (UID: \"92724c2e-d7fd-4673-ae84-8929541f2a13\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wrw67" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.201623 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6rqws\" (UniqueName: \"kubernetes.io/projected/a51dadb3-2af5-4c6c-a973-cdceb34cd79a-kube-api-access-6rqws\") pod \"openshift-apiserver-operator-796bbdcf4f-bcfw4\" (UID: \"a51dadb3-2af5-4c6c-a973-cdceb34cd79a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bcfw4" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.233904 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wrw67" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.234503 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k9ksj\" (UniqueName: \"kubernetes.io/projected/51fb93ab-ea9f-4fce-9fd1-a57958bc9c10-kube-api-access-k9ksj\") pod \"console-operator-58897d9998-pszqv\" (UID: \"51fb93ab-ea9f-4fce-9fd1-a57958bc9c10\") " pod="openshift-console-operator/console-operator-58897d9998-pszqv" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.252001 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9d47p\" (UniqueName: \"kubernetes.io/projected/1306f3b2-2982-494e-9db0-86c2e05b7a82-kube-api-access-9d47p\") pod \"downloads-7954f5f757-b257c\" (UID: \"1306f3b2-2982-494e-9db0-86c2e05b7a82\") " pod="openshift-console/downloads-7954f5f757-b257c" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.266322 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fw2t8\" (UniqueName: \"kubernetes.io/projected/cf1baeba-a0c2-4d91-ae76-eb4f0ceccd60-kube-api-access-fw2t8\") pod \"cluster-samples-operator-665b6dd947-jjn8w\" (UID: \"cf1baeba-a0c2-4d91-ae76-eb4f0ceccd60\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-jjn8w" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.271060 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-qjcjv"] Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.283072 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97b4v\" (UniqueName: \"kubernetes.io/projected/fc892d74-c6d2-431f-a52c-95a8f4e2d03c-kube-api-access-97b4v\") pod \"dns-operator-744455d44c-zp6l4\" (UID: \"fc892d74-c6d2-431f-a52c-95a8f4e2d03c\") " pod="openshift-dns-operator/dns-operator-744455d44c-zp6l4" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.302072 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8439dba6-4f32-4c02-9aad-27b9c40f7575-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-d49gn\" (UID: \"8439dba6-4f32-4c02-9aad-27b9c40f7575\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-d49gn" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.302784 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-gc4cp" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.315582 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4d2qh" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.323055 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/915af7e7-19b0-4b1c-9e55-3471210ccf40-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zjwcl\" (UID: \"915af7e7-19b0-4b1c-9e55-3471210ccf40\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zjwcl" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.341433 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-6xp6f" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.347088 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-j26br"] Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.349274 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hk885\" (UniqueName: \"kubernetes.io/projected/85bc0929-de17-4d2d-8b03-1a05d207fb55-kube-api-access-hk885\") pod \"router-default-5444994796-h5l29\" (UID: \"85bc0929-de17-4d2d-8b03-1a05d207fb55\") " pod="openshift-ingress/router-default-5444994796-h5l29" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.372689 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.377736 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7ff64" event={"ID":"6e5a490b-b9b3-4f19-aebf-14f6c1efbaa2","Type":"ContainerStarted","Data":"06dd01263babd531d4c9a62ea9245f915f9049e39f2cbaf72c231e2c8b3a6e59"} Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.379709 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-qjcjv" event={"ID":"3cef2105-ae00-4224-bf5d-246393f3caa6","Type":"ContainerStarted","Data":"829cc2178638ad9f3b726631e977079615f192b96abda8d78a8fa8cccf2b9f5b"} Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.386486 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.407643 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.410779 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-ndp5v" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.417995 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zjwcl" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.420433 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-z587m"] Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.424798 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-pszqv" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.431137 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-jjn8w" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.432134 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.440560 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-9bs98" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.456353 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-b257c" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.462617 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bcfw4" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.464060 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8sxfn"] Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.476266 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-zp6l4" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.482906 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-h5l29" Dec 07 16:04:31 crc kubenswrapper[4716]: W1207 16:04:31.499293 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9a0a9cf5_14e9_41e1_8c6c_27b7b3b8c7b4.slice/crio-ea2aaf5d2d2ba52d98b48ddacfa69aa8d75f65d779fffe754046a153f3a0c04d WatchSource:0}: Error finding container ea2aaf5d2d2ba52d98b48ddacfa69aa8d75f65d779fffe754046a153f3a0c04d: Status 404 returned error can't find the container with id ea2aaf5d2d2ba52d98b48ddacfa69aa8d75f65d779fffe754046a153f3a0c04d Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.526854 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-j7zhr"] Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.539189 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18deffad-cf79-40dd-95f3-ed8410002584-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-d7c2c\" (UID: \"18deffad-cf79-40dd-95f3-ed8410002584\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-d7c2c" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.539240 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/75569b15-b1c1-46bf-83f2-cd6eb0a38381-installation-pull-secrets\") pod \"image-registry-697d97f7c8-4d257\" (UID: \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d257" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.539266 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/75569b15-b1c1-46bf-83f2-cd6eb0a38381-registry-certificates\") pod \"image-registry-697d97f7c8-4d257\" (UID: \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d257" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.539293 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/75569b15-b1c1-46bf-83f2-cd6eb0a38381-bound-sa-token\") pod \"image-registry-697d97f7c8-4d257\" (UID: \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d257" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.539320 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wgns8\" (UniqueName: \"kubernetes.io/projected/18deffad-cf79-40dd-95f3-ed8410002584-kube-api-access-wgns8\") pod \"openshift-controller-manager-operator-756b6f6bc6-d7c2c\" (UID: \"18deffad-cf79-40dd-95f3-ed8410002584\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-d7c2c" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.539345 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d257\" (UID: \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d257" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.539419 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/75569b15-b1c1-46bf-83f2-cd6eb0a38381-registry-tls\") pod \"image-registry-697d97f7c8-4d257\" (UID: \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d257" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.539443 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hgpp9\" (UniqueName: \"kubernetes.io/projected/75569b15-b1c1-46bf-83f2-cd6eb0a38381-kube-api-access-hgpp9\") pod \"image-registry-697d97f7c8-4d257\" (UID: \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d257" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.539459 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/18deffad-cf79-40dd-95f3-ed8410002584-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-d7c2c\" (UID: \"18deffad-cf79-40dd-95f3-ed8410002584\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-d7c2c" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.539488 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/75569b15-b1c1-46bf-83f2-cd6eb0a38381-ca-trust-extracted\") pod \"image-registry-697d97f7c8-4d257\" (UID: \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d257" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.539507 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/75569b15-b1c1-46bf-83f2-cd6eb0a38381-trusted-ca\") pod \"image-registry-697d97f7c8-4d257\" (UID: \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d257" Dec 07 16:04:31 crc kubenswrapper[4716]: E1207 16:04:31.540685 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 16:04:32.04067346 +0000 UTC m=+134.730958372 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d257" (UID: "75569b15-b1c1-46bf-83f2-cd6eb0a38381") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.541432 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-d49gn" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.548532 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-wrw67"] Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.640946 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 16:04:31 crc kubenswrapper[4716]: E1207 16:04:31.641020 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 16:04:32.14099815 +0000 UTC m=+134.831283062 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.641214 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/fe74013f-e94b-4600-8d02-6f73efbd4212-srv-cert\") pod \"catalog-operator-68c6474976-bwsl4\" (UID: \"fe74013f-e94b-4600-8d02-6f73efbd4212\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bwsl4" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.641281 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/d27e02ff-86da-4078-934b-fd472b471f69-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-sq5lw\" (UID: \"d27e02ff-86da-4078-934b-fd472b471f69\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-sq5lw" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.641343 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/98efd33b-de1f-4414-b9cc-55fb0d423784-config-volume\") pod \"collect-profiles-29418720-x5rqz\" (UID: \"98efd33b-de1f-4414-b9cc-55fb0d423784\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418720-x5rqz" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.641376 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/372517a0-98df-460c-a7ed-a01624cb747a-certs\") pod \"machine-config-server-gcr9f\" (UID: \"372517a0-98df-460c-a7ed-a01624cb747a\") " pod="openshift-machine-config-operator/machine-config-server-gcr9f" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.641426 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/75569b15-b1c1-46bf-83f2-cd6eb0a38381-registry-certificates\") pod \"image-registry-697d97f7c8-4d257\" (UID: \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d257" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.641472 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wgns8\" (UniqueName: \"kubernetes.io/projected/18deffad-cf79-40dd-95f3-ed8410002584-kube-api-access-wgns8\") pod \"openshift-controller-manager-operator-756b6f6bc6-d7c2c\" (UID: \"18deffad-cf79-40dd-95f3-ed8410002584\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-d7c2c" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.641517 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/fe74013f-e94b-4600-8d02-6f73efbd4212-profile-collector-cert\") pod \"catalog-operator-68c6474976-bwsl4\" (UID: \"fe74013f-e94b-4600-8d02-6f73efbd4212\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bwsl4" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.641546 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/38ef2332-5fee-44a2-b56c-bc6db5fad2e0-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-w7xhs\" (UID: \"38ef2332-5fee-44a2-b56c-bc6db5fad2e0\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-w7xhs" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.641668 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/55935fe0-005a-4f9f-80cd-9fa3348cc90a-webhook-cert\") pod \"packageserver-d55dfcdfc-l9dgz\" (UID: \"55935fe0-005a-4f9f-80cd-9fa3348cc90a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-l9dgz" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.641751 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/af51b694-cd0f-4c33-ba6e-1f4666178820-registration-dir\") pod \"csi-hostpathplugin-82bvt\" (UID: \"af51b694-cd0f-4c33-ba6e-1f4666178820\") " pod="hostpath-provisioner/csi-hostpathplugin-82bvt" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.641843 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jmt77\" (UniqueName: \"kubernetes.io/projected/98efd33b-de1f-4414-b9cc-55fb0d423784-kube-api-access-jmt77\") pod \"collect-profiles-29418720-x5rqz\" (UID: \"98efd33b-de1f-4414-b9cc-55fb0d423784\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418720-x5rqz" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.642619 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/6b4e49ef-27ab-417c-9c6b-1ad7f215604c-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-89v7m\" (UID: \"6b4e49ef-27ab-417c-9c6b-1ad7f215604c\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-89v7m" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.642671 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwnxj\" (UniqueName: \"kubernetes.io/projected/6b4e49ef-27ab-417c-9c6b-1ad7f215604c-kube-api-access-kwnxj\") pod \"package-server-manager-789f6589d5-89v7m\" (UID: \"6b4e49ef-27ab-417c-9c6b-1ad7f215604c\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-89v7m" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.642714 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1e623490-3813-4a3a-b7bf-beb8b870ed9e-serving-cert\") pod \"service-ca-operator-777779d784-tqctv\" (UID: \"1e623490-3813-4a3a-b7bf-beb8b870ed9e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-tqctv" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.642743 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/af51b694-cd0f-4c33-ba6e-1f4666178820-csi-data-dir\") pod \"csi-hostpathplugin-82bvt\" (UID: \"af51b694-cd0f-4c33-ba6e-1f4666178820\") " pod="hostpath-provisioner/csi-hostpathplugin-82bvt" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.642785 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/75569b15-b1c1-46bf-83f2-cd6eb0a38381-registry-certificates\") pod \"image-registry-697d97f7c8-4d257\" (UID: \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d257" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.642923 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/55935fe0-005a-4f9f-80cd-9fa3348cc90a-tmpfs\") pod \"packageserver-d55dfcdfc-l9dgz\" (UID: \"55935fe0-005a-4f9f-80cd-9fa3348cc90a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-l9dgz" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.643473 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/75569b15-b1c1-46bf-83f2-cd6eb0a38381-registry-tls\") pod \"image-registry-697d97f7c8-4d257\" (UID: \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d257" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.643689 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h4x4s\" (UniqueName: \"kubernetes.io/projected/fe74013f-e94b-4600-8d02-6f73efbd4212-kube-api-access-h4x4s\") pod \"catalog-operator-68c6474976-bwsl4\" (UID: \"fe74013f-e94b-4600-8d02-6f73efbd4212\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bwsl4" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.643833 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/98efd33b-de1f-4414-b9cc-55fb0d423784-secret-volume\") pod \"collect-profiles-29418720-x5rqz\" (UID: \"98efd33b-de1f-4414-b9cc-55fb0d423784\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418720-x5rqz" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.643881 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/db753ef2-2939-42dd-886d-17a223c47360-config-volume\") pod \"dns-default-n4492\" (UID: \"db753ef2-2939-42dd-886d-17a223c47360\") " pod="openshift-dns/dns-default-n4492" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.643991 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/5b5665d7-a97a-4d47-b37e-bd3f1643c7cf-signing-key\") pod \"service-ca-9c57cc56f-8sv58\" (UID: \"5b5665d7-a97a-4d47-b37e-bd3f1643c7cf\") " pod="openshift-service-ca/service-ca-9c57cc56f-8sv58" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.644184 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/75569b15-b1c1-46bf-83f2-cd6eb0a38381-trusted-ca\") pod \"image-registry-697d97f7c8-4d257\" (UID: \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d257" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.644477 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/060bb436-de36-408a-b742-f3768a575801-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-rb966\" (UID: \"060bb436-de36-408a-b742-f3768a575801\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rb966" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.644634 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/db753ef2-2939-42dd-886d-17a223c47360-metrics-tls\") pod \"dns-default-n4492\" (UID: \"db753ef2-2939-42dd-886d-17a223c47360\") " pod="openshift-dns/dns-default-n4492" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.644671 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffn72\" (UniqueName: \"kubernetes.io/projected/1e623490-3813-4a3a-b7bf-beb8b870ed9e-kube-api-access-ffn72\") pod \"service-ca-operator-777779d784-tqctv\" (UID: \"1e623490-3813-4a3a-b7bf-beb8b870ed9e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-tqctv" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.644698 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/38c4a34b-864b-42d0-a3cb-90ede4226ccd-profile-collector-cert\") pod \"olm-operator-6b444d44fb-6j7q2\" (UID: \"38c4a34b-864b-42d0-a3cb-90ede4226ccd\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6j7q2" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.644744 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/372517a0-98df-460c-a7ed-a01624cb747a-node-bootstrap-token\") pod \"machine-config-server-gcr9f\" (UID: \"372517a0-98df-460c-a7ed-a01624cb747a\") " pod="openshift-machine-config-operator/machine-config-server-gcr9f" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.644762 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2g952\" (UniqueName: \"kubernetes.io/projected/5d2c595e-c4a6-446d-836a-766d1ec44516-kube-api-access-2g952\") pod \"ingress-canary-v9d58\" (UID: \"5d2c595e-c4a6-446d-836a-766d1ec44516\") " pod="openshift-ingress-canary/ingress-canary-v9d58" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.644794 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18deffad-cf79-40dd-95f3-ed8410002584-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-d7c2c\" (UID: \"18deffad-cf79-40dd-95f3-ed8410002584\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-d7c2c" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.644810 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9kgw4\" (UniqueName: \"kubernetes.io/projected/060bb436-de36-408a-b742-f3768a575801-kube-api-access-9kgw4\") pod \"kube-storage-version-migrator-operator-b67b599dd-rb966\" (UID: \"060bb436-de36-408a-b742-f3768a575801\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rb966" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.644827 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/38ef2332-5fee-44a2-b56c-bc6db5fad2e0-proxy-tls\") pod \"machine-config-controller-84d6567774-w7xhs\" (UID: \"38ef2332-5fee-44a2-b56c-bc6db5fad2e0\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-w7xhs" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.644855 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sptdt\" (UniqueName: \"kubernetes.io/projected/74c41933-56ed-4f49-a808-1d7f1f1829b2-kube-api-access-sptdt\") pod \"machine-config-operator-74547568cd-h6gnn\" (UID: \"74c41933-56ed-4f49-a808-1d7f1f1829b2\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-h6gnn" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.644873 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/75569b15-b1c1-46bf-83f2-cd6eb0a38381-installation-pull-secrets\") pod \"image-registry-697d97f7c8-4d257\" (UID: \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d257" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.644898 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/38c4a34b-864b-42d0-a3cb-90ede4226ccd-srv-cert\") pod \"olm-operator-6b444d44fb-6j7q2\" (UID: \"38c4a34b-864b-42d0-a3cb-90ede4226ccd\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6j7q2" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.644913 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxvt2\" (UniqueName: \"kubernetes.io/projected/af51b694-cd0f-4c33-ba6e-1f4666178820-kube-api-access-zxvt2\") pod \"csi-hostpathplugin-82bvt\" (UID: \"af51b694-cd0f-4c33-ba6e-1f4666178820\") " pod="hostpath-provisioner/csi-hostpathplugin-82bvt" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.644940 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/75569b15-b1c1-46bf-83f2-cd6eb0a38381-bound-sa-token\") pod \"image-registry-697d97f7c8-4d257\" (UID: \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d257" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.644979 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/55935fe0-005a-4f9f-80cd-9fa3348cc90a-apiservice-cert\") pod \"packageserver-d55dfcdfc-l9dgz\" (UID: \"55935fe0-005a-4f9f-80cd-9fa3348cc90a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-l9dgz" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.644994 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhwhw\" (UniqueName: \"kubernetes.io/projected/9345bd64-bb11-4d7b-96ca-9ae4a32b4e60-kube-api-access-qhwhw\") pod \"migrator-59844c95c7-kkvsk\" (UID: \"9345bd64-bb11-4d7b-96ca-9ae4a32b4e60\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-kkvsk" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.645027 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d257\" (UID: \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d257" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.645045 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mx47d\" (UniqueName: \"kubernetes.io/projected/38c4a34b-864b-42d0-a3cb-90ede4226ccd-kube-api-access-mx47d\") pod \"olm-operator-6b444d44fb-6j7q2\" (UID: \"38c4a34b-864b-42d0-a3cb-90ede4226ccd\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6j7q2" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.645062 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/af51b694-cd0f-4c33-ba6e-1f4666178820-mountpoint-dir\") pod \"csi-hostpathplugin-82bvt\" (UID: \"af51b694-cd0f-4c33-ba6e-1f4666178820\") " pod="hostpath-provisioner/csi-hostpathplugin-82bvt" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.645124 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/580ae9d7-004b-4d66-ba7b-1566de502f66-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-64rzd\" (UID: \"580ae9d7-004b-4d66-ba7b-1566de502f66\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-64rzd" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.645141 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/af51b694-cd0f-4c33-ba6e-1f4666178820-socket-dir\") pod \"csi-hostpathplugin-82bvt\" (UID: \"af51b694-cd0f-4c33-ba6e-1f4666178820\") " pod="hostpath-provisioner/csi-hostpathplugin-82bvt" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.645166 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2nq9\" (UniqueName: \"kubernetes.io/projected/372517a0-98df-460c-a7ed-a01624cb747a-kube-api-access-m2nq9\") pod \"machine-config-server-gcr9f\" (UID: \"372517a0-98df-460c-a7ed-a01624cb747a\") " pod="openshift-machine-config-operator/machine-config-server-gcr9f" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.645180 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7mfvh\" (UniqueName: \"kubernetes.io/projected/db753ef2-2939-42dd-886d-17a223c47360-kube-api-access-7mfvh\") pod \"dns-default-n4492\" (UID: \"db753ef2-2939-42dd-886d-17a223c47360\") " pod="openshift-dns/dns-default-n4492" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.645201 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gj7mx\" (UniqueName: \"kubernetes.io/projected/d27e02ff-86da-4078-934b-fd472b471f69-kube-api-access-gj7mx\") pod \"multus-admission-controller-857f4d67dd-sq5lw\" (UID: \"d27e02ff-86da-4078-934b-fd472b471f69\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-sq5lw" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.645222 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vljq6\" (UniqueName: \"kubernetes.io/projected/5b5665d7-a97a-4d47-b37e-bd3f1643c7cf-kube-api-access-vljq6\") pod \"service-ca-9c57cc56f-8sv58\" (UID: \"5b5665d7-a97a-4d47-b37e-bd3f1643c7cf\") " pod="openshift-service-ca/service-ca-9c57cc56f-8sv58" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.645239 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmqh5\" (UniqueName: \"kubernetes.io/projected/7a9cc84d-c02b-466f-9c1c-c1ae60e697ef-kube-api-access-dmqh5\") pod \"marketplace-operator-79b997595-v6899\" (UID: \"7a9cc84d-c02b-466f-9c1c-c1ae60e697ef\") " pod="openshift-marketplace/marketplace-operator-79b997595-v6899" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.645256 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7a9cc84d-c02b-466f-9c1c-c1ae60e697ef-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-v6899\" (UID: \"7a9cc84d-c02b-466f-9c1c-c1ae60e697ef\") " pod="openshift-marketplace/marketplace-operator-79b997595-v6899" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.645271 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e623490-3813-4a3a-b7bf-beb8b870ed9e-config\") pod \"service-ca-operator-777779d784-tqctv\" (UID: \"1e623490-3813-4a3a-b7bf-beb8b870ed9e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-tqctv" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.645298 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/18deffad-cf79-40dd-95f3-ed8410002584-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-d7c2c\" (UID: \"18deffad-cf79-40dd-95f3-ed8410002584\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-d7c2c" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.645316 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hgpp9\" (UniqueName: \"kubernetes.io/projected/75569b15-b1c1-46bf-83f2-cd6eb0a38381-kube-api-access-hgpp9\") pod \"image-registry-697d97f7c8-4d257\" (UID: \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d257" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.645333 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/74c41933-56ed-4f49-a808-1d7f1f1829b2-proxy-tls\") pod \"machine-config-operator-74547568cd-h6gnn\" (UID: \"74c41933-56ed-4f49-a808-1d7f1f1829b2\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-h6gnn" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.645348 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/5b5665d7-a97a-4d47-b37e-bd3f1643c7cf-signing-cabundle\") pod \"service-ca-9c57cc56f-8sv58\" (UID: \"5b5665d7-a97a-4d47-b37e-bd3f1643c7cf\") " pod="openshift-service-ca/service-ca-9c57cc56f-8sv58" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.645362 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/af51b694-cd0f-4c33-ba6e-1f4666178820-plugins-dir\") pod \"csi-hostpathplugin-82bvt\" (UID: \"af51b694-cd0f-4c33-ba6e-1f4666178820\") " pod="hostpath-provisioner/csi-hostpathplugin-82bvt" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.645380 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/74c41933-56ed-4f49-a808-1d7f1f1829b2-images\") pod \"machine-config-operator-74547568cd-h6gnn\" (UID: \"74c41933-56ed-4f49-a808-1d7f1f1829b2\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-h6gnn" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.645398 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/060bb436-de36-408a-b742-f3768a575801-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-rb966\" (UID: \"060bb436-de36-408a-b742-f3768a575801\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rb966" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.645424 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/db812416-33e4-462c-a28f-ec40b629dc37-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-mmfjq\" (UID: \"db812416-33e4-462c-a28f-ec40b629dc37\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mmfjq" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.645463 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7a9cc84d-c02b-466f-9c1c-c1ae60e697ef-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-v6899\" (UID: \"7a9cc84d-c02b-466f-9c1c-c1ae60e697ef\") " pod="openshift-marketplace/marketplace-operator-79b997595-v6899" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.645478 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5d2c595e-c4a6-446d-836a-766d1ec44516-cert\") pod \"ingress-canary-v9d58\" (UID: \"5d2c595e-c4a6-446d-836a-766d1ec44516\") " pod="openshift-ingress-canary/ingress-canary-v9d58" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.645495 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/75569b15-b1c1-46bf-83f2-cd6eb0a38381-ca-trust-extracted\") pod \"image-registry-697d97f7c8-4d257\" (UID: \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d257" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.645510 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nf2wf\" (UniqueName: \"kubernetes.io/projected/db812416-33e4-462c-a28f-ec40b629dc37-kube-api-access-nf2wf\") pod \"control-plane-machine-set-operator-78cbb6b69f-mmfjq\" (UID: \"db812416-33e4-462c-a28f-ec40b629dc37\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mmfjq" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.645526 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jvmkm\" (UniqueName: \"kubernetes.io/projected/55935fe0-005a-4f9f-80cd-9fa3348cc90a-kube-api-access-jvmkm\") pod \"packageserver-d55dfcdfc-l9dgz\" (UID: \"55935fe0-005a-4f9f-80cd-9fa3348cc90a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-l9dgz" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.645542 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/580ae9d7-004b-4d66-ba7b-1566de502f66-config\") pod \"kube-apiserver-operator-766d6c64bb-64rzd\" (UID: \"580ae9d7-004b-4d66-ba7b-1566de502f66\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-64rzd" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.645595 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/74c41933-56ed-4f49-a808-1d7f1f1829b2-auth-proxy-config\") pod \"machine-config-operator-74547568cd-h6gnn\" (UID: \"74c41933-56ed-4f49-a808-1d7f1f1829b2\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-h6gnn" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.646352 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/75569b15-b1c1-46bf-83f2-cd6eb0a38381-ca-trust-extracted\") pod \"image-registry-697d97f7c8-4d257\" (UID: \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d257" Dec 07 16:04:31 crc kubenswrapper[4716]: E1207 16:04:31.646555 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 16:04:32.14654671 +0000 UTC m=+134.836831622 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d257" (UID: "75569b15-b1c1-46bf-83f2-cd6eb0a38381") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.646809 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4kp64\" (UniqueName: \"kubernetes.io/projected/38ef2332-5fee-44a2-b56c-bc6db5fad2e0-kube-api-access-4kp64\") pod \"machine-config-controller-84d6567774-w7xhs\" (UID: \"38ef2332-5fee-44a2-b56c-bc6db5fad2e0\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-w7xhs" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.646841 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/580ae9d7-004b-4d66-ba7b-1566de502f66-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-64rzd\" (UID: \"580ae9d7-004b-4d66-ba7b-1566de502f66\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-64rzd" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.648978 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18deffad-cf79-40dd-95f3-ed8410002584-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-d7c2c\" (UID: \"18deffad-cf79-40dd-95f3-ed8410002584\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-d7c2c" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.649200 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/75569b15-b1c1-46bf-83f2-cd6eb0a38381-registry-tls\") pod \"image-registry-697d97f7c8-4d257\" (UID: \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d257" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.652634 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-4d2qh"] Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.652692 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/75569b15-b1c1-46bf-83f2-cd6eb0a38381-installation-pull-secrets\") pod \"image-registry-697d97f7c8-4d257\" (UID: \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d257" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.654039 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/18deffad-cf79-40dd-95f3-ed8410002584-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-d7c2c\" (UID: \"18deffad-cf79-40dd-95f3-ed8410002584\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-d7c2c" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.681120 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wgns8\" (UniqueName: \"kubernetes.io/projected/18deffad-cf79-40dd-95f3-ed8410002584-kube-api-access-wgns8\") pod \"openshift-controller-manager-operator-756b6f6bc6-d7c2c\" (UID: \"18deffad-cf79-40dd-95f3-ed8410002584\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-d7c2c" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.700430 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/75569b15-b1c1-46bf-83f2-cd6eb0a38381-bound-sa-token\") pod \"image-registry-697d97f7c8-4d257\" (UID: \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d257" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.715612 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/75569b15-b1c1-46bf-83f2-cd6eb0a38381-trusted-ca\") pod \"image-registry-697d97f7c8-4d257\" (UID: \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d257" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.723552 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hgpp9\" (UniqueName: \"kubernetes.io/projected/75569b15-b1c1-46bf-83f2-cd6eb0a38381-kube-api-access-hgpp9\") pod \"image-registry-697d97f7c8-4d257\" (UID: \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d257" Dec 07 16:04:31 crc kubenswrapper[4716]: W1207 16:04:31.740062 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod97d9c075_80c3_4dcb_aebc_649eec930413.slice/crio-9db3e06592b07e08f8a1d6f4cbb7ad969e1f1bce6cbc0de81cf9f32d702d66c6 WatchSource:0}: Error finding container 9db3e06592b07e08f8a1d6f4cbb7ad969e1f1bce6cbc0de81cf9f32d702d66c6: Status 404 returned error can't find the container with id 9db3e06592b07e08f8a1d6f4cbb7ad969e1f1bce6cbc0de81cf9f32d702d66c6 Dec 07 16:04:31 crc kubenswrapper[4716]: W1207 16:04:31.742582 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod92724c2e_d7fd_4673_ae84_8929541f2a13.slice/crio-4d7c79e2b44bfedd3982b6b48910cc1ae0fffdb1285a256836c9109abd0b79d6 WatchSource:0}: Error finding container 4d7c79e2b44bfedd3982b6b48910cc1ae0fffdb1285a256836c9109abd0b79d6: Status 404 returned error can't find the container with id 4d7c79e2b44bfedd3982b6b48910cc1ae0fffdb1285a256836c9109abd0b79d6 Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.747180 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 16:04:31 crc kubenswrapper[4716]: E1207 16:04:31.747289 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 16:04:32.247273413 +0000 UTC m=+134.937558325 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.747381 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/38c4a34b-864b-42d0-a3cb-90ede4226ccd-srv-cert\") pod \"olm-operator-6b444d44fb-6j7q2\" (UID: \"38c4a34b-864b-42d0-a3cb-90ede4226ccd\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6j7q2" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.747405 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zxvt2\" (UniqueName: \"kubernetes.io/projected/af51b694-cd0f-4c33-ba6e-1f4666178820-kube-api-access-zxvt2\") pod \"csi-hostpathplugin-82bvt\" (UID: \"af51b694-cd0f-4c33-ba6e-1f4666178820\") " pod="hostpath-provisioner/csi-hostpathplugin-82bvt" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.747424 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/55935fe0-005a-4f9f-80cd-9fa3348cc90a-apiservice-cert\") pod \"packageserver-d55dfcdfc-l9dgz\" (UID: \"55935fe0-005a-4f9f-80cd-9fa3348cc90a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-l9dgz" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.747442 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhwhw\" (UniqueName: \"kubernetes.io/projected/9345bd64-bb11-4d7b-96ca-9ae4a32b4e60-kube-api-access-qhwhw\") pod \"migrator-59844c95c7-kkvsk\" (UID: \"9345bd64-bb11-4d7b-96ca-9ae4a32b4e60\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-kkvsk" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.747582 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d257\" (UID: \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d257" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.747604 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mx47d\" (UniqueName: \"kubernetes.io/projected/38c4a34b-864b-42d0-a3cb-90ede4226ccd-kube-api-access-mx47d\") pod \"olm-operator-6b444d44fb-6j7q2\" (UID: \"38c4a34b-864b-42d0-a3cb-90ede4226ccd\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6j7q2" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.747636 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/af51b694-cd0f-4c33-ba6e-1f4666178820-mountpoint-dir\") pod \"csi-hostpathplugin-82bvt\" (UID: \"af51b694-cd0f-4c33-ba6e-1f4666178820\") " pod="hostpath-provisioner/csi-hostpathplugin-82bvt" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.747657 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/580ae9d7-004b-4d66-ba7b-1566de502f66-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-64rzd\" (UID: \"580ae9d7-004b-4d66-ba7b-1566de502f66\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-64rzd" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.747671 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/af51b694-cd0f-4c33-ba6e-1f4666178820-socket-dir\") pod \"csi-hostpathplugin-82bvt\" (UID: \"af51b694-cd0f-4c33-ba6e-1f4666178820\") " pod="hostpath-provisioner/csi-hostpathplugin-82bvt" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.747725 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/af51b694-cd0f-4c33-ba6e-1f4666178820-mountpoint-dir\") pod \"csi-hostpathplugin-82bvt\" (UID: \"af51b694-cd0f-4c33-ba6e-1f4666178820\") " pod="hostpath-provisioner/csi-hostpathplugin-82bvt" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.747766 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7mfvh\" (UniqueName: \"kubernetes.io/projected/db753ef2-2939-42dd-886d-17a223c47360-kube-api-access-7mfvh\") pod \"dns-default-n4492\" (UID: \"db753ef2-2939-42dd-886d-17a223c47360\") " pod="openshift-dns/dns-default-n4492" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.747785 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m2nq9\" (UniqueName: \"kubernetes.io/projected/372517a0-98df-460c-a7ed-a01624cb747a-kube-api-access-m2nq9\") pod \"machine-config-server-gcr9f\" (UID: \"372517a0-98df-460c-a7ed-a01624cb747a\") " pod="openshift-machine-config-operator/machine-config-server-gcr9f" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.747872 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gj7mx\" (UniqueName: \"kubernetes.io/projected/d27e02ff-86da-4078-934b-fd472b471f69-kube-api-access-gj7mx\") pod \"multus-admission-controller-857f4d67dd-sq5lw\" (UID: \"d27e02ff-86da-4078-934b-fd472b471f69\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-sq5lw" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.747898 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmqh5\" (UniqueName: \"kubernetes.io/projected/7a9cc84d-c02b-466f-9c1c-c1ae60e697ef-kube-api-access-dmqh5\") pod \"marketplace-operator-79b997595-v6899\" (UID: \"7a9cc84d-c02b-466f-9c1c-c1ae60e697ef\") " pod="openshift-marketplace/marketplace-operator-79b997595-v6899" Dec 07 16:04:31 crc kubenswrapper[4716]: E1207 16:04:31.747964 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 16:04:32.247933133 +0000 UTC m=+134.938218045 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d257" (UID: "75569b15-b1c1-46bf-83f2-cd6eb0a38381") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.748029 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vljq6\" (UniqueName: \"kubernetes.io/projected/5b5665d7-a97a-4d47-b37e-bd3f1643c7cf-kube-api-access-vljq6\") pod \"service-ca-9c57cc56f-8sv58\" (UID: \"5b5665d7-a97a-4d47-b37e-bd3f1643c7cf\") " pod="openshift-service-ca/service-ca-9c57cc56f-8sv58" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.748041 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/af51b694-cd0f-4c33-ba6e-1f4666178820-socket-dir\") pod \"csi-hostpathplugin-82bvt\" (UID: \"af51b694-cd0f-4c33-ba6e-1f4666178820\") " pod="hostpath-provisioner/csi-hostpathplugin-82bvt" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.748054 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7a9cc84d-c02b-466f-9c1c-c1ae60e697ef-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-v6899\" (UID: \"7a9cc84d-c02b-466f-9c1c-c1ae60e697ef\") " pod="openshift-marketplace/marketplace-operator-79b997595-v6899" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.748094 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e623490-3813-4a3a-b7bf-beb8b870ed9e-config\") pod \"service-ca-operator-777779d784-tqctv\" (UID: \"1e623490-3813-4a3a-b7bf-beb8b870ed9e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-tqctv" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.748115 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/74c41933-56ed-4f49-a808-1d7f1f1829b2-images\") pod \"machine-config-operator-74547568cd-h6gnn\" (UID: \"74c41933-56ed-4f49-a808-1d7f1f1829b2\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-h6gnn" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.748134 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/74c41933-56ed-4f49-a808-1d7f1f1829b2-proxy-tls\") pod \"machine-config-operator-74547568cd-h6gnn\" (UID: \"74c41933-56ed-4f49-a808-1d7f1f1829b2\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-h6gnn" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.748149 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/5b5665d7-a97a-4d47-b37e-bd3f1643c7cf-signing-cabundle\") pod \"service-ca-9c57cc56f-8sv58\" (UID: \"5b5665d7-a97a-4d47-b37e-bd3f1643c7cf\") " pod="openshift-service-ca/service-ca-9c57cc56f-8sv58" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.748164 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/af51b694-cd0f-4c33-ba6e-1f4666178820-plugins-dir\") pod \"csi-hostpathplugin-82bvt\" (UID: \"af51b694-cd0f-4c33-ba6e-1f4666178820\") " pod="hostpath-provisioner/csi-hostpathplugin-82bvt" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.748180 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/db812416-33e4-462c-a28f-ec40b629dc37-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-mmfjq\" (UID: \"db812416-33e4-462c-a28f-ec40b629dc37\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mmfjq" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.748197 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/060bb436-de36-408a-b742-f3768a575801-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-rb966\" (UID: \"060bb436-de36-408a-b742-f3768a575801\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rb966" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.748214 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7a9cc84d-c02b-466f-9c1c-c1ae60e697ef-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-v6899\" (UID: \"7a9cc84d-c02b-466f-9c1c-c1ae60e697ef\") " pod="openshift-marketplace/marketplace-operator-79b997595-v6899" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.748229 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5d2c595e-c4a6-446d-836a-766d1ec44516-cert\") pod \"ingress-canary-v9d58\" (UID: \"5d2c595e-c4a6-446d-836a-766d1ec44516\") " pod="openshift-ingress-canary/ingress-canary-v9d58" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.748246 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nf2wf\" (UniqueName: \"kubernetes.io/projected/db812416-33e4-462c-a28f-ec40b629dc37-kube-api-access-nf2wf\") pod \"control-plane-machine-set-operator-78cbb6b69f-mmfjq\" (UID: \"db812416-33e4-462c-a28f-ec40b629dc37\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mmfjq" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.748260 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jvmkm\" (UniqueName: \"kubernetes.io/projected/55935fe0-005a-4f9f-80cd-9fa3348cc90a-kube-api-access-jvmkm\") pod \"packageserver-d55dfcdfc-l9dgz\" (UID: \"55935fe0-005a-4f9f-80cd-9fa3348cc90a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-l9dgz" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.748277 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/580ae9d7-004b-4d66-ba7b-1566de502f66-config\") pod \"kube-apiserver-operator-766d6c64bb-64rzd\" (UID: \"580ae9d7-004b-4d66-ba7b-1566de502f66\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-64rzd" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.748295 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/74c41933-56ed-4f49-a808-1d7f1f1829b2-auth-proxy-config\") pod \"machine-config-operator-74547568cd-h6gnn\" (UID: \"74c41933-56ed-4f49-a808-1d7f1f1829b2\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-h6gnn" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.748325 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4kp64\" (UniqueName: \"kubernetes.io/projected/38ef2332-5fee-44a2-b56c-bc6db5fad2e0-kube-api-access-4kp64\") pod \"machine-config-controller-84d6567774-w7xhs\" (UID: \"38ef2332-5fee-44a2-b56c-bc6db5fad2e0\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-w7xhs" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.748341 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/580ae9d7-004b-4d66-ba7b-1566de502f66-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-64rzd\" (UID: \"580ae9d7-004b-4d66-ba7b-1566de502f66\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-64rzd" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.748356 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/fe74013f-e94b-4600-8d02-6f73efbd4212-srv-cert\") pod \"catalog-operator-68c6474976-bwsl4\" (UID: \"fe74013f-e94b-4600-8d02-6f73efbd4212\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bwsl4" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.748372 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/d27e02ff-86da-4078-934b-fd472b471f69-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-sq5lw\" (UID: \"d27e02ff-86da-4078-934b-fd472b471f69\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-sq5lw" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.748388 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/372517a0-98df-460c-a7ed-a01624cb747a-certs\") pod \"machine-config-server-gcr9f\" (UID: \"372517a0-98df-460c-a7ed-a01624cb747a\") " pod="openshift-machine-config-operator/machine-config-server-gcr9f" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.748403 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/98efd33b-de1f-4414-b9cc-55fb0d423784-config-volume\") pod \"collect-profiles-29418720-x5rqz\" (UID: \"98efd33b-de1f-4414-b9cc-55fb0d423784\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418720-x5rqz" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.748422 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/fe74013f-e94b-4600-8d02-6f73efbd4212-profile-collector-cert\") pod \"catalog-operator-68c6474976-bwsl4\" (UID: \"fe74013f-e94b-4600-8d02-6f73efbd4212\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bwsl4" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.748438 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/38ef2332-5fee-44a2-b56c-bc6db5fad2e0-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-w7xhs\" (UID: \"38ef2332-5fee-44a2-b56c-bc6db5fad2e0\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-w7xhs" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.748456 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/55935fe0-005a-4f9f-80cd-9fa3348cc90a-webhook-cert\") pod \"packageserver-d55dfcdfc-l9dgz\" (UID: \"55935fe0-005a-4f9f-80cd-9fa3348cc90a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-l9dgz" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.748472 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/af51b694-cd0f-4c33-ba6e-1f4666178820-registration-dir\") pod \"csi-hostpathplugin-82bvt\" (UID: \"af51b694-cd0f-4c33-ba6e-1f4666178820\") " pod="hostpath-provisioner/csi-hostpathplugin-82bvt" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.748490 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jmt77\" (UniqueName: \"kubernetes.io/projected/98efd33b-de1f-4414-b9cc-55fb0d423784-kube-api-access-jmt77\") pod \"collect-profiles-29418720-x5rqz\" (UID: \"98efd33b-de1f-4414-b9cc-55fb0d423784\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418720-x5rqz" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.748517 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/6b4e49ef-27ab-417c-9c6b-1ad7f215604c-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-89v7m\" (UID: \"6b4e49ef-27ab-417c-9c6b-1ad7f215604c\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-89v7m" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.748536 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwnxj\" (UniqueName: \"kubernetes.io/projected/6b4e49ef-27ab-417c-9c6b-1ad7f215604c-kube-api-access-kwnxj\") pod \"package-server-manager-789f6589d5-89v7m\" (UID: \"6b4e49ef-27ab-417c-9c6b-1ad7f215604c\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-89v7m" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.748551 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1e623490-3813-4a3a-b7bf-beb8b870ed9e-serving-cert\") pod \"service-ca-operator-777779d784-tqctv\" (UID: \"1e623490-3813-4a3a-b7bf-beb8b870ed9e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-tqctv" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.748567 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/af51b694-cd0f-4c33-ba6e-1f4666178820-csi-data-dir\") pod \"csi-hostpathplugin-82bvt\" (UID: \"af51b694-cd0f-4c33-ba6e-1f4666178820\") " pod="hostpath-provisioner/csi-hostpathplugin-82bvt" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.748595 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/55935fe0-005a-4f9f-80cd-9fa3348cc90a-tmpfs\") pod \"packageserver-d55dfcdfc-l9dgz\" (UID: \"55935fe0-005a-4f9f-80cd-9fa3348cc90a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-l9dgz" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.748621 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h4x4s\" (UniqueName: \"kubernetes.io/projected/fe74013f-e94b-4600-8d02-6f73efbd4212-kube-api-access-h4x4s\") pod \"catalog-operator-68c6474976-bwsl4\" (UID: \"fe74013f-e94b-4600-8d02-6f73efbd4212\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bwsl4" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.748645 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/98efd33b-de1f-4414-b9cc-55fb0d423784-secret-volume\") pod \"collect-profiles-29418720-x5rqz\" (UID: \"98efd33b-de1f-4414-b9cc-55fb0d423784\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418720-x5rqz" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.748666 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/db753ef2-2939-42dd-886d-17a223c47360-config-volume\") pod \"dns-default-n4492\" (UID: \"db753ef2-2939-42dd-886d-17a223c47360\") " pod="openshift-dns/dns-default-n4492" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.748642 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e623490-3813-4a3a-b7bf-beb8b870ed9e-config\") pod \"service-ca-operator-777779d784-tqctv\" (UID: \"1e623490-3813-4a3a-b7bf-beb8b870ed9e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-tqctv" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.748683 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/5b5665d7-a97a-4d47-b37e-bd3f1643c7cf-signing-key\") pod \"service-ca-9c57cc56f-8sv58\" (UID: \"5b5665d7-a97a-4d47-b37e-bd3f1643c7cf\") " pod="openshift-service-ca/service-ca-9c57cc56f-8sv58" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.748709 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/060bb436-de36-408a-b742-f3768a575801-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-rb966\" (UID: \"060bb436-de36-408a-b742-f3768a575801\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rb966" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.748723 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/db753ef2-2939-42dd-886d-17a223c47360-metrics-tls\") pod \"dns-default-n4492\" (UID: \"db753ef2-2939-42dd-886d-17a223c47360\") " pod="openshift-dns/dns-default-n4492" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.748739 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffn72\" (UniqueName: \"kubernetes.io/projected/1e623490-3813-4a3a-b7bf-beb8b870ed9e-kube-api-access-ffn72\") pod \"service-ca-operator-777779d784-tqctv\" (UID: \"1e623490-3813-4a3a-b7bf-beb8b870ed9e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-tqctv" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.748757 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/38c4a34b-864b-42d0-a3cb-90ede4226ccd-profile-collector-cert\") pod \"olm-operator-6b444d44fb-6j7q2\" (UID: \"38c4a34b-864b-42d0-a3cb-90ede4226ccd\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6j7q2" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.748776 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/372517a0-98df-460c-a7ed-a01624cb747a-node-bootstrap-token\") pod \"machine-config-server-gcr9f\" (UID: \"372517a0-98df-460c-a7ed-a01624cb747a\") " pod="openshift-machine-config-operator/machine-config-server-gcr9f" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.748790 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2g952\" (UniqueName: \"kubernetes.io/projected/5d2c595e-c4a6-446d-836a-766d1ec44516-kube-api-access-2g952\") pod \"ingress-canary-v9d58\" (UID: \"5d2c595e-c4a6-446d-836a-766d1ec44516\") " pod="openshift-ingress-canary/ingress-canary-v9d58" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.748806 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9kgw4\" (UniqueName: \"kubernetes.io/projected/060bb436-de36-408a-b742-f3768a575801-kube-api-access-9kgw4\") pod \"kube-storage-version-migrator-operator-b67b599dd-rb966\" (UID: \"060bb436-de36-408a-b742-f3768a575801\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rb966" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.748821 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/38ef2332-5fee-44a2-b56c-bc6db5fad2e0-proxy-tls\") pod \"machine-config-controller-84d6567774-w7xhs\" (UID: \"38ef2332-5fee-44a2-b56c-bc6db5fad2e0\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-w7xhs" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.748836 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sptdt\" (UniqueName: \"kubernetes.io/projected/74c41933-56ed-4f49-a808-1d7f1f1829b2-kube-api-access-sptdt\") pod \"machine-config-operator-74547568cd-h6gnn\" (UID: \"74c41933-56ed-4f49-a808-1d7f1f1829b2\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-h6gnn" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.749167 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7a9cc84d-c02b-466f-9c1c-c1ae60e697ef-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-v6899\" (UID: \"7a9cc84d-c02b-466f-9c1c-c1ae60e697ef\") " pod="openshift-marketplace/marketplace-operator-79b997595-v6899" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.749847 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/af51b694-cd0f-4c33-ba6e-1f4666178820-plugins-dir\") pod \"csi-hostpathplugin-82bvt\" (UID: \"af51b694-cd0f-4c33-ba6e-1f4666178820\") " pod="hostpath-provisioner/csi-hostpathplugin-82bvt" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.749965 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/38ef2332-5fee-44a2-b56c-bc6db5fad2e0-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-w7xhs\" (UID: \"38ef2332-5fee-44a2-b56c-bc6db5fad2e0\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-w7xhs" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.750561 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/98efd33b-de1f-4414-b9cc-55fb0d423784-config-volume\") pod \"collect-profiles-29418720-x5rqz\" (UID: \"98efd33b-de1f-4414-b9cc-55fb0d423784\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418720-x5rqz" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.751064 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/af51b694-cd0f-4c33-ba6e-1f4666178820-registration-dir\") pod \"csi-hostpathplugin-82bvt\" (UID: \"af51b694-cd0f-4c33-ba6e-1f4666178820\") " pod="hostpath-provisioner/csi-hostpathplugin-82bvt" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.751150 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/74c41933-56ed-4f49-a808-1d7f1f1829b2-images\") pod \"machine-config-operator-74547568cd-h6gnn\" (UID: \"74c41933-56ed-4f49-a808-1d7f1f1829b2\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-h6gnn" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.751803 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/5b5665d7-a97a-4d47-b37e-bd3f1643c7cf-signing-cabundle\") pod \"service-ca-9c57cc56f-8sv58\" (UID: \"5b5665d7-a97a-4d47-b37e-bd3f1643c7cf\") " pod="openshift-service-ca/service-ca-9c57cc56f-8sv58" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.753875 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/74c41933-56ed-4f49-a808-1d7f1f1829b2-auth-proxy-config\") pod \"machine-config-operator-74547568cd-h6gnn\" (UID: \"74c41933-56ed-4f49-a808-1d7f1f1829b2\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-h6gnn" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.754257 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/af51b694-cd0f-4c33-ba6e-1f4666178820-csi-data-dir\") pod \"csi-hostpathplugin-82bvt\" (UID: \"af51b694-cd0f-4c33-ba6e-1f4666178820\") " pod="hostpath-provisioner/csi-hostpathplugin-82bvt" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.754379 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/db753ef2-2939-42dd-886d-17a223c47360-config-volume\") pod \"dns-default-n4492\" (UID: \"db753ef2-2939-42dd-886d-17a223c47360\") " pod="openshift-dns/dns-default-n4492" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.754648 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/55935fe0-005a-4f9f-80cd-9fa3348cc90a-tmpfs\") pod \"packageserver-d55dfcdfc-l9dgz\" (UID: \"55935fe0-005a-4f9f-80cd-9fa3348cc90a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-l9dgz" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.755735 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/d27e02ff-86da-4078-934b-fd472b471f69-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-sq5lw\" (UID: \"d27e02ff-86da-4078-934b-fd472b471f69\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-sq5lw" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.755925 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/060bb436-de36-408a-b742-f3768a575801-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-rb966\" (UID: \"060bb436-de36-408a-b742-f3768a575801\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rb966" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.756914 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/98efd33b-de1f-4414-b9cc-55fb0d423784-secret-volume\") pod \"collect-profiles-29418720-x5rqz\" (UID: \"98efd33b-de1f-4414-b9cc-55fb0d423784\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418720-x5rqz" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.758336 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7a9cc84d-c02b-466f-9c1c-c1ae60e697ef-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-v6899\" (UID: \"7a9cc84d-c02b-466f-9c1c-c1ae60e697ef\") " pod="openshift-marketplace/marketplace-operator-79b997595-v6899" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.758517 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/55935fe0-005a-4f9f-80cd-9fa3348cc90a-webhook-cert\") pod \"packageserver-d55dfcdfc-l9dgz\" (UID: \"55935fe0-005a-4f9f-80cd-9fa3348cc90a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-l9dgz" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.759092 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/060bb436-de36-408a-b742-f3768a575801-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-rb966\" (UID: \"060bb436-de36-408a-b742-f3768a575801\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rb966" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.772484 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/38c4a34b-864b-42d0-a3cb-90ede4226ccd-srv-cert\") pod \"olm-operator-6b444d44fb-6j7q2\" (UID: \"38c4a34b-864b-42d0-a3cb-90ede4226ccd\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6j7q2" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.772772 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/38c4a34b-864b-42d0-a3cb-90ede4226ccd-profile-collector-cert\") pod \"olm-operator-6b444d44fb-6j7q2\" (UID: \"38c4a34b-864b-42d0-a3cb-90ede4226ccd\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6j7q2" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.773095 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/580ae9d7-004b-4d66-ba7b-1566de502f66-config\") pod \"kube-apiserver-operator-766d6c64bb-64rzd\" (UID: \"580ae9d7-004b-4d66-ba7b-1566de502f66\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-64rzd" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.773126 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/74c41933-56ed-4f49-a808-1d7f1f1829b2-proxy-tls\") pod \"machine-config-operator-74547568cd-h6gnn\" (UID: \"74c41933-56ed-4f49-a808-1d7f1f1829b2\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-h6gnn" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.773460 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/db753ef2-2939-42dd-886d-17a223c47360-metrics-tls\") pod \"dns-default-n4492\" (UID: \"db753ef2-2939-42dd-886d-17a223c47360\") " pod="openshift-dns/dns-default-n4492" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.773734 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/372517a0-98df-460c-a7ed-a01624cb747a-certs\") pod \"machine-config-server-gcr9f\" (UID: \"372517a0-98df-460c-a7ed-a01624cb747a\") " pod="openshift-machine-config-operator/machine-config-server-gcr9f" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.773846 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/38ef2332-5fee-44a2-b56c-bc6db5fad2e0-proxy-tls\") pod \"machine-config-controller-84d6567774-w7xhs\" (UID: \"38ef2332-5fee-44a2-b56c-bc6db5fad2e0\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-w7xhs" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.774124 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/6b4e49ef-27ab-417c-9c6b-1ad7f215604c-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-89v7m\" (UID: \"6b4e49ef-27ab-417c-9c6b-1ad7f215604c\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-89v7m" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.774381 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5d2c595e-c4a6-446d-836a-766d1ec44516-cert\") pod \"ingress-canary-v9d58\" (UID: \"5d2c595e-c4a6-446d-836a-766d1ec44516\") " pod="openshift-ingress-canary/ingress-canary-v9d58" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.774428 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/5b5665d7-a97a-4d47-b37e-bd3f1643c7cf-signing-key\") pod \"service-ca-9c57cc56f-8sv58\" (UID: \"5b5665d7-a97a-4d47-b37e-bd3f1643c7cf\") " pod="openshift-service-ca/service-ca-9c57cc56f-8sv58" Dec 07 16:04:31 crc kubenswrapper[4716]: W1207 16:04:31.774699 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4df2ba74_d18e_4e94_bee1_703645b1a463.slice/crio-621aeb63ecbbc2e32e0453b6e478e07fe7bc847b4aa5612ee20e2f2d22be3bb3 WatchSource:0}: Error finding container 621aeb63ecbbc2e32e0453b6e478e07fe7bc847b4aa5612ee20e2f2d22be3bb3: Status 404 returned error can't find the container with id 621aeb63ecbbc2e32e0453b6e478e07fe7bc847b4aa5612ee20e2f2d22be3bb3 Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.774903 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/fe74013f-e94b-4600-8d02-6f73efbd4212-srv-cert\") pod \"catalog-operator-68c6474976-bwsl4\" (UID: \"fe74013f-e94b-4600-8d02-6f73efbd4212\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bwsl4" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.775694 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/55935fe0-005a-4f9f-80cd-9fa3348cc90a-apiservice-cert\") pod \"packageserver-d55dfcdfc-l9dgz\" (UID: \"55935fe0-005a-4f9f-80cd-9fa3348cc90a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-l9dgz" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.776160 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1e623490-3813-4a3a-b7bf-beb8b870ed9e-serving-cert\") pod \"service-ca-operator-777779d784-tqctv\" (UID: \"1e623490-3813-4a3a-b7bf-beb8b870ed9e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-tqctv" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.776163 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/580ae9d7-004b-4d66-ba7b-1566de502f66-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-64rzd\" (UID: \"580ae9d7-004b-4d66-ba7b-1566de502f66\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-64rzd" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.777003 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-d7c2c" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.783664 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/db812416-33e4-462c-a28f-ec40b629dc37-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-mmfjq\" (UID: \"db812416-33e4-462c-a28f-ec40b629dc37\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mmfjq" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.784113 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/372517a0-98df-460c-a7ed-a01624cb747a-node-bootstrap-token\") pod \"machine-config-server-gcr9f\" (UID: \"372517a0-98df-460c-a7ed-a01624cb747a\") " pod="openshift-machine-config-operator/machine-config-server-gcr9f" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.784694 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/fe74013f-e94b-4600-8d02-6f73efbd4212-profile-collector-cert\") pod \"catalog-operator-68c6474976-bwsl4\" (UID: \"fe74013f-e94b-4600-8d02-6f73efbd4212\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bwsl4" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.797328 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhwhw\" (UniqueName: \"kubernetes.io/projected/9345bd64-bb11-4d7b-96ca-9ae4a32b4e60-kube-api-access-qhwhw\") pod \"migrator-59844c95c7-kkvsk\" (UID: \"9345bd64-bb11-4d7b-96ca-9ae4a32b4e60\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-kkvsk" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.829407 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxvt2\" (UniqueName: \"kubernetes.io/projected/af51b694-cd0f-4c33-ba6e-1f4666178820-kube-api-access-zxvt2\") pod \"csi-hostpathplugin-82bvt\" (UID: \"af51b694-cd0f-4c33-ba6e-1f4666178820\") " pod="hostpath-provisioner/csi-hostpathplugin-82bvt" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.836494 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mx47d\" (UniqueName: \"kubernetes.io/projected/38c4a34b-864b-42d0-a3cb-90ede4226ccd-kube-api-access-mx47d\") pod \"olm-operator-6b444d44fb-6j7q2\" (UID: \"38c4a34b-864b-42d0-a3cb-90ede4226ccd\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6j7q2" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.849410 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 16:04:31 crc kubenswrapper[4716]: E1207 16:04:31.849885 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 16:04:32.34986865 +0000 UTC m=+135.040153562 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.851230 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-x2nd8"] Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.858977 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-gc4cp"] Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.862747 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/580ae9d7-004b-4d66-ba7b-1566de502f66-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-64rzd\" (UID: \"580ae9d7-004b-4d66-ba7b-1566de502f66\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-64rzd" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.884416 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-64rzd" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.884587 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m2nq9\" (UniqueName: \"kubernetes.io/projected/372517a0-98df-460c-a7ed-a01624cb747a-kube-api-access-m2nq9\") pod \"machine-config-server-gcr9f\" (UID: \"372517a0-98df-460c-a7ed-a01624cb747a\") " pod="openshift-machine-config-operator/machine-config-server-gcr9f" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.895760 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6j7q2" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.903844 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gj7mx\" (UniqueName: \"kubernetes.io/projected/d27e02ff-86da-4078-934b-fd472b471f69-kube-api-access-gj7mx\") pod \"multus-admission-controller-857f4d67dd-sq5lw\" (UID: \"d27e02ff-86da-4078-934b-fd472b471f69\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-sq5lw" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.917999 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-ndp5v"] Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.924659 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmqh5\" (UniqueName: \"kubernetes.io/projected/7a9cc84d-c02b-466f-9c1c-c1ae60e697ef-kube-api-access-dmqh5\") pod \"marketplace-operator-79b997595-v6899\" (UID: \"7a9cc84d-c02b-466f-9c1c-c1ae60e697ef\") " pod="openshift-marketplace/marketplace-operator-79b997595-v6899" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.926783 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-kkvsk" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.938656 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-6xp6f"] Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.938718 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-pszqv"] Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.939979 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zjwcl"] Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.944812 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vljq6\" (UniqueName: \"kubernetes.io/projected/5b5665d7-a97a-4d47-b37e-bd3f1643c7cf-kube-api-access-vljq6\") pod \"service-ca-9c57cc56f-8sv58\" (UID: \"5b5665d7-a97a-4d47-b37e-bd3f1643c7cf\") " pod="openshift-service-ca/service-ca-9c57cc56f-8sv58" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.951376 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d257\" (UID: \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d257" Dec 07 16:04:31 crc kubenswrapper[4716]: E1207 16:04:31.951690 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 16:04:32.451678114 +0000 UTC m=+135.141963026 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d257" (UID: "75569b15-b1c1-46bf-83f2-cd6eb0a38381") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.956845 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-v6899" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.961353 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sptdt\" (UniqueName: \"kubernetes.io/projected/74c41933-56ed-4f49-a808-1d7f1f1829b2-kube-api-access-sptdt\") pod \"machine-config-operator-74547568cd-h6gnn\" (UID: \"74c41933-56ed-4f49-a808-1d7f1f1829b2\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-h6gnn" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.971426 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-8sv58" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.981179 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h4x4s\" (UniqueName: \"kubernetes.io/projected/fe74013f-e94b-4600-8d02-6f73efbd4212-kube-api-access-h4x4s\") pod \"catalog-operator-68c6474976-bwsl4\" (UID: \"fe74013f-e94b-4600-8d02-6f73efbd4212\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bwsl4" Dec 07 16:04:31 crc kubenswrapper[4716]: I1207 16:04:31.993796 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7mfvh\" (UniqueName: \"kubernetes.io/projected/db753ef2-2939-42dd-886d-17a223c47360-kube-api-access-7mfvh\") pod \"dns-default-n4492\" (UID: \"db753ef2-2939-42dd-886d-17a223c47360\") " pod="openshift-dns/dns-default-n4492" Dec 07 16:04:32 crc kubenswrapper[4716]: I1207 16:04:32.002054 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jvmkm\" (UniqueName: \"kubernetes.io/projected/55935fe0-005a-4f9f-80cd-9fa3348cc90a-kube-api-access-jvmkm\") pod \"packageserver-d55dfcdfc-l9dgz\" (UID: \"55935fe0-005a-4f9f-80cd-9fa3348cc90a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-l9dgz" Dec 07 16:04:32 crc kubenswrapper[4716]: I1207 16:04:32.022990 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-82bvt" Dec 07 16:04:32 crc kubenswrapper[4716]: I1207 16:04:32.026611 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nf2wf\" (UniqueName: \"kubernetes.io/projected/db812416-33e4-462c-a28f-ec40b629dc37-kube-api-access-nf2wf\") pod \"control-plane-machine-set-operator-78cbb6b69f-mmfjq\" (UID: \"db812416-33e4-462c-a28f-ec40b629dc37\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mmfjq" Dec 07 16:04:32 crc kubenswrapper[4716]: I1207 16:04:32.030232 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-gcr9f" Dec 07 16:04:32 crc kubenswrapper[4716]: I1207 16:04:32.039260 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-b257c"] Dec 07 16:04:32 crc kubenswrapper[4716]: W1207 16:04:32.040432 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6f7305ef_d76f_4be3_9151_7247f7590fe0.slice/crio-6ce2b42ebdd69f6cb3efaf4524bbf9cea198af8f5a80c808f8764030f00b83d4 WatchSource:0}: Error finding container 6ce2b42ebdd69f6cb3efaf4524bbf9cea198af8f5a80c808f8764030f00b83d4: Status 404 returned error can't find the container with id 6ce2b42ebdd69f6cb3efaf4524bbf9cea198af8f5a80c808f8764030f00b83d4 Dec 07 16:04:32 crc kubenswrapper[4716]: I1207 16:04:32.044825 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jmt77\" (UniqueName: \"kubernetes.io/projected/98efd33b-de1f-4414-b9cc-55fb0d423784-kube-api-access-jmt77\") pod \"collect-profiles-29418720-x5rqz\" (UID: \"98efd33b-de1f-4414-b9cc-55fb0d423784\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418720-x5rqz" Dec 07 16:04:32 crc kubenswrapper[4716]: I1207 16:04:32.052837 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 16:04:32 crc kubenswrapper[4716]: E1207 16:04:32.053209 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 16:04:32.55318832 +0000 UTC m=+135.243473232 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:04:32 crc kubenswrapper[4716]: I1207 16:04:32.053341 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d257\" (UID: \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d257" Dec 07 16:04:32 crc kubenswrapper[4716]: E1207 16:04:32.053653 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 16:04:32.553624553 +0000 UTC m=+135.243909465 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d257" (UID: "75569b15-b1c1-46bf-83f2-cd6eb0a38381") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:04:32 crc kubenswrapper[4716]: I1207 16:04:32.060412 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4kp64\" (UniqueName: \"kubernetes.io/projected/38ef2332-5fee-44a2-b56c-bc6db5fad2e0-kube-api-access-4kp64\") pod \"machine-config-controller-84d6567774-w7xhs\" (UID: \"38ef2332-5fee-44a2-b56c-bc6db5fad2e0\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-w7xhs" Dec 07 16:04:32 crc kubenswrapper[4716]: W1207 16:04:32.077851 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod915af7e7_19b0_4b1c_9e55_3471210ccf40.slice/crio-877f75b5570f4b8f7906d4b12c7c61cc74bb16d75030e18974c202abbbc221fa WatchSource:0}: Error finding container 877f75b5570f4b8f7906d4b12c7c61cc74bb16d75030e18974c202abbbc221fa: Status 404 returned error can't find the container with id 877f75b5570f4b8f7906d4b12c7c61cc74bb16d75030e18974c202abbbc221fa Dec 07 16:04:32 crc kubenswrapper[4716]: I1207 16:04:32.086994 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffn72\" (UniqueName: \"kubernetes.io/projected/1e623490-3813-4a3a-b7bf-beb8b870ed9e-kube-api-access-ffn72\") pod \"service-ca-operator-777779d784-tqctv\" (UID: \"1e623490-3813-4a3a-b7bf-beb8b870ed9e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-tqctv" Dec 07 16:04:32 crc kubenswrapper[4716]: W1207 16:04:32.092946 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1306f3b2_2982_494e_9db0_86c2e05b7a82.slice/crio-b6b1ba60f2be298356b609d90e800522b74a521656626c35ac7f1d03ba96056a WatchSource:0}: Error finding container b6b1ba60f2be298356b609d90e800522b74a521656626c35ac7f1d03ba96056a: Status 404 returned error can't find the container with id b6b1ba60f2be298356b609d90e800522b74a521656626c35ac7f1d03ba96056a Dec 07 16:04:32 crc kubenswrapper[4716]: I1207 16:04:32.104580 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwnxj\" (UniqueName: \"kubernetes.io/projected/6b4e49ef-27ab-417c-9c6b-1ad7f215604c-kube-api-access-kwnxj\") pod \"package-server-manager-789f6589d5-89v7m\" (UID: \"6b4e49ef-27ab-417c-9c6b-1ad7f215604c\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-89v7m" Dec 07 16:04:32 crc kubenswrapper[4716]: I1207 16:04:32.123663 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2g952\" (UniqueName: \"kubernetes.io/projected/5d2c595e-c4a6-446d-836a-766d1ec44516-kube-api-access-2g952\") pod \"ingress-canary-v9d58\" (UID: \"5d2c595e-c4a6-446d-836a-766d1ec44516\") " pod="openshift-ingress-canary/ingress-canary-v9d58" Dec 07 16:04:32 crc kubenswrapper[4716]: I1207 16:04:32.140638 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9kgw4\" (UniqueName: \"kubernetes.io/projected/060bb436-de36-408a-b742-f3768a575801-kube-api-access-9kgw4\") pod \"kube-storage-version-migrator-operator-b67b599dd-rb966\" (UID: \"060bb436-de36-408a-b742-f3768a575801\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rb966" Dec 07 16:04:32 crc kubenswrapper[4716]: I1207 16:04:32.153476 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mmfjq" Dec 07 16:04:32 crc kubenswrapper[4716]: I1207 16:04:32.154164 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 16:04:32 crc kubenswrapper[4716]: I1207 16:04:32.154660 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rb966" Dec 07 16:04:32 crc kubenswrapper[4716]: E1207 16:04:32.154847 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 16:04:32.654821099 +0000 UTC m=+135.345106011 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:04:32 crc kubenswrapper[4716]: I1207 16:04:32.169391 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-w7xhs" Dec 07 16:04:32 crc kubenswrapper[4716]: I1207 16:04:32.186453 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-sq5lw" Dec 07 16:04:32 crc kubenswrapper[4716]: I1207 16:04:32.187916 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-h6gnn" Dec 07 16:04:32 crc kubenswrapper[4716]: I1207 16:04:32.208502 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-jjn8w"] Dec 07 16:04:32 crc kubenswrapper[4716]: I1207 16:04:32.209665 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-9bs98"] Dec 07 16:04:32 crc kubenswrapper[4716]: I1207 16:04:32.233635 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bwsl4" Dec 07 16:04:32 crc kubenswrapper[4716]: I1207 16:04:32.243040 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-l9dgz" Dec 07 16:04:32 crc kubenswrapper[4716]: I1207 16:04:32.249532 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-89v7m" Dec 07 16:04:32 crc kubenswrapper[4716]: I1207 16:04:32.256767 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d257\" (UID: \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d257" Dec 07 16:04:32 crc kubenswrapper[4716]: E1207 16:04:32.257169 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 16:04:32.757156908 +0000 UTC m=+135.447441820 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d257" (UID: "75569b15-b1c1-46bf-83f2-cd6eb0a38381") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:04:32 crc kubenswrapper[4716]: I1207 16:04:32.264530 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-tqctv" Dec 07 16:04:32 crc kubenswrapper[4716]: I1207 16:04:32.272671 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bcfw4"] Dec 07 16:04:32 crc kubenswrapper[4716]: I1207 16:04:32.276425 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-d49gn"] Dec 07 16:04:32 crc kubenswrapper[4716]: I1207 16:04:32.277263 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29418720-x5rqz" Dec 07 16:04:32 crc kubenswrapper[4716]: I1207 16:04:32.286501 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-n4492" Dec 07 16:04:32 crc kubenswrapper[4716]: I1207 16:04:32.292841 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-v9d58" Dec 07 16:04:32 crc kubenswrapper[4716]: I1207 16:04:32.308976 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-v6899"] Dec 07 16:04:32 crc kubenswrapper[4716]: W1207 16:04:32.315552 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc94992de_685b_42c3_8100_93996239d0b0.slice/crio-f9b0c96d338433c05d2613a932da175fe57d99021624d580327df21a79827084 WatchSource:0}: Error finding container f9b0c96d338433c05d2613a932da175fe57d99021624d580327df21a79827084: Status 404 returned error can't find the container with id f9b0c96d338433c05d2613a932da175fe57d99021624d580327df21a79827084 Dec 07 16:04:32 crc kubenswrapper[4716]: I1207 16:04:32.331834 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-zp6l4"] Dec 07 16:04:32 crc kubenswrapper[4716]: I1207 16:04:32.358112 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 16:04:32 crc kubenswrapper[4716]: E1207 16:04:32.358442 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 16:04:32.858426847 +0000 UTC m=+135.548711759 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:04:32 crc kubenswrapper[4716]: I1207 16:04:32.403729 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zjwcl" event={"ID":"915af7e7-19b0-4b1c-9e55-3471210ccf40","Type":"ContainerStarted","Data":"877f75b5570f4b8f7906d4b12c7c61cc74bb16d75030e18974c202abbbc221fa"} Dec 07 16:04:32 crc kubenswrapper[4716]: I1207 16:04:32.454665 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-qjcjv" event={"ID":"3cef2105-ae00-4224-bf5d-246393f3caa6","Type":"ContainerStarted","Data":"612a6e8ee10acdd7b730f5775b4e6a6ac3fb0b8ebd6af77189069e871b78a6fd"} Dec 07 16:04:32 crc kubenswrapper[4716]: I1207 16:04:32.455550 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-qjcjv" Dec 07 16:04:32 crc kubenswrapper[4716]: I1207 16:04:32.458945 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d257\" (UID: \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d257" Dec 07 16:04:32 crc kubenswrapper[4716]: E1207 16:04:32.459253 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 16:04:32.959241863 +0000 UTC m=+135.649526775 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d257" (UID: "75569b15-b1c1-46bf-83f2-cd6eb0a38381") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:04:32 crc kubenswrapper[4716]: I1207 16:04:32.464991 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-64rzd"] Dec 07 16:04:32 crc kubenswrapper[4716]: I1207 16:04:32.469361 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4d2qh" event={"ID":"97d9c075-80c3-4dcb-aebc-649eec930413","Type":"ContainerStarted","Data":"9db3e06592b07e08f8a1d6f4cbb7ad969e1f1bce6cbc0de81cf9f32d702d66c6"} Dec 07 16:04:32 crc kubenswrapper[4716]: I1207 16:04:32.470773 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-6xp6f" event={"ID":"f086c3a7-c2a4-4066-aed6-270b5dd6eef8","Type":"ContainerStarted","Data":"210ecdd915682476366ef28f1b0c669c7969b1f53717f28309438ae2790998de"} Dec 07 16:04:32 crc kubenswrapper[4716]: I1207 16:04:32.471446 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-z587m" event={"ID":"ffff2831-338e-423e-81f2-c8d3b5de7785","Type":"ContainerStarted","Data":"659375d5d974bfaecb1f1380f95cb82e35d3ceb12360c50b80b664804fcee30e"} Dec 07 16:04:32 crc kubenswrapper[4716]: I1207 16:04:32.483376 4716 generic.go:334] "Generic (PLEG): container finished" podID="297ee903-c479-460e-80df-a0c508c45fbe" containerID="f5f6392ec3eb3f377cc7c1a5388dc121d63b6451f3942ec17f78b40bb1887474" exitCode=0 Dec 07 16:04:32 crc kubenswrapper[4716]: I1207 16:04:32.483450 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j26br" event={"ID":"297ee903-c479-460e-80df-a0c508c45fbe","Type":"ContainerDied","Data":"f5f6392ec3eb3f377cc7c1a5388dc121d63b6451f3942ec17f78b40bb1887474"} Dec 07 16:04:32 crc kubenswrapper[4716]: I1207 16:04:32.483489 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j26br" event={"ID":"297ee903-c479-460e-80df-a0c508c45fbe","Type":"ContainerStarted","Data":"183f81f5b55af77ecb38ef5c432046c8e1c98c3854f1ff4d11d38eba27d3e4f7"} Dec 07 16:04:32 crc kubenswrapper[4716]: I1207 16:04:32.483829 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-qjcjv" Dec 07 16:04:32 crc kubenswrapper[4716]: I1207 16:04:32.487825 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-gc4cp" event={"ID":"6f7305ef-d76f-4be3-9151-7247f7590fe0","Type":"ContainerStarted","Data":"6ce2b42ebdd69f6cb3efaf4524bbf9cea198af8f5a80c808f8764030f00b83d4"} Dec 07 16:04:32 crc kubenswrapper[4716]: I1207 16:04:32.489676 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-h5l29" event={"ID":"85bc0929-de17-4d2d-8b03-1a05d207fb55","Type":"ContainerStarted","Data":"cad74b5ecc4f0a810895a93a7ebc7d7fa02fc87cc8a41ea99b98515a136bfcdd"} Dec 07 16:04:32 crc kubenswrapper[4716]: I1207 16:04:32.490998 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-9bs98" event={"ID":"c94992de-685b-42c3-8100-93996239d0b0","Type":"ContainerStarted","Data":"f9b0c96d338433c05d2613a932da175fe57d99021624d580327df21a79827084"} Dec 07 16:04:32 crc kubenswrapper[4716]: I1207 16:04:32.496283 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8sxfn" event={"ID":"9a0a9cf5-14e9-41e1-8c6c-27b7b3b8c7b4","Type":"ContainerStarted","Data":"ea2aaf5d2d2ba52d98b48ddacfa69aa8d75f65d779fffe754046a153f3a0c04d"} Dec 07 16:04:32 crc kubenswrapper[4716]: I1207 16:04:32.501738 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-d7c2c"] Dec 07 16:04:32 crc kubenswrapper[4716]: I1207 16:04:32.503101 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-ndp5v" event={"ID":"cbf3f592-bbb1-4596-8ae1-0e62aa366f0b","Type":"ContainerStarted","Data":"af061ef8ee4d7573b0865c8a6560467f821a1a39b9e853cda55704fe1185a924"} Dec 07 16:04:32 crc kubenswrapper[4716]: I1207 16:04:32.504248 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-j7zhr" event={"ID":"4df2ba74-d18e-4e94-bee1-703645b1a463","Type":"ContainerStarted","Data":"621aeb63ecbbc2e32e0453b6e478e07fe7bc847b4aa5612ee20e2f2d22be3bb3"} Dec 07 16:04:32 crc kubenswrapper[4716]: I1207 16:04:32.505486 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wrw67" event={"ID":"92724c2e-d7fd-4673-ae84-8929541f2a13","Type":"ContainerStarted","Data":"4d7c79e2b44bfedd3982b6b48910cc1ae0fffdb1285a256836c9109abd0b79d6"} Dec 07 16:04:32 crc kubenswrapper[4716]: I1207 16:04:32.506313 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-pszqv" event={"ID":"51fb93ab-ea9f-4fce-9fd1-a57958bc9c10","Type":"ContainerStarted","Data":"a5e59744558311a933aa81b41d8d2bae8e881a5e332e6366af85982a44afd44d"} Dec 07 16:04:32 crc kubenswrapper[4716]: I1207 16:04:32.507266 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7ff64" event={"ID":"6e5a490b-b9b3-4f19-aebf-14f6c1efbaa2","Type":"ContainerStarted","Data":"8cd7234a1d1d2a57524bec3a79d4b5770a5921afa87888def13c34ea94a3f083"} Dec 07 16:04:32 crc kubenswrapper[4716]: I1207 16:04:32.508004 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-b257c" event={"ID":"1306f3b2-2982-494e-9db0-86c2e05b7a82","Type":"ContainerStarted","Data":"b6b1ba60f2be298356b609d90e800522b74a521656626c35ac7f1d03ba96056a"} Dec 07 16:04:32 crc kubenswrapper[4716]: I1207 16:04:32.509195 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-x2nd8" event={"ID":"23992ff8-b006-40d1-9048-1950f10fff7d","Type":"ContainerStarted","Data":"06a132f581dd374bebf5d41366d48ec2f653d1988421469793cbbb77d765c057"} Dec 07 16:04:32 crc kubenswrapper[4716]: I1207 16:04:32.561941 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 16:04:32 crc kubenswrapper[4716]: E1207 16:04:32.562122 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 16:04:33.062063656 +0000 UTC m=+135.752348568 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:04:32 crc kubenswrapper[4716]: I1207 16:04:32.562490 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d257\" (UID: \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d257" Dec 07 16:04:32 crc kubenswrapper[4716]: E1207 16:04:32.566036 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 16:04:33.0660215 +0000 UTC m=+135.756306412 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d257" (UID: "75569b15-b1c1-46bf-83f2-cd6eb0a38381") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:04:32 crc kubenswrapper[4716]: I1207 16:04:32.588701 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6j7q2"] Dec 07 16:04:32 crc kubenswrapper[4716]: I1207 16:04:32.663336 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 16:04:32 crc kubenswrapper[4716]: E1207 16:04:32.664888 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 16:04:33.164869719 +0000 UTC m=+135.855154641 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:04:32 crc kubenswrapper[4716]: I1207 16:04:32.742347 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-qjcjv" podStartSLOduration=113.742330201 podStartE2EDuration="1m53.742330201s" podCreationTimestamp="2025-12-07 16:02:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:04:32.741058044 +0000 UTC m=+135.431342956" watchObservedRunningTime="2025-12-07 16:04:32.742330201 +0000 UTC m=+135.432615113" Dec 07 16:04:32 crc kubenswrapper[4716]: I1207 16:04:32.765022 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d257\" (UID: \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d257" Dec 07 16:04:32 crc kubenswrapper[4716]: E1207 16:04:32.780112 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 16:04:33.280071429 +0000 UTC m=+135.970356341 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d257" (UID: "75569b15-b1c1-46bf-83f2-cd6eb0a38381") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:04:32 crc kubenswrapper[4716]: I1207 16:04:32.828237 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-8sv58"] Dec 07 16:04:32 crc kubenswrapper[4716]: I1207 16:04:32.867001 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 16:04:32 crc kubenswrapper[4716]: E1207 16:04:32.867304 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 16:04:33.367287632 +0000 UTC m=+136.057572544 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:04:32 crc kubenswrapper[4716]: I1207 16:04:32.908969 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-kkvsk"] Dec 07 16:04:32 crc kubenswrapper[4716]: I1207 16:04:32.970309 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d257\" (UID: \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d257" Dec 07 16:04:32 crc kubenswrapper[4716]: E1207 16:04:32.971282 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 16:04:33.471270209 +0000 UTC m=+136.161555121 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d257" (UID: "75569b15-b1c1-46bf-83f2-cd6eb0a38381") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:04:33 crc kubenswrapper[4716]: I1207 16:04:33.044328 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8sxfn" podStartSLOduration=114.044314514 podStartE2EDuration="1m54.044314514s" podCreationTimestamp="2025-12-07 16:02:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:04:33.043730327 +0000 UTC m=+135.734015239" watchObservedRunningTime="2025-12-07 16:04:33.044314514 +0000 UTC m=+135.734599426" Dec 07 16:04:33 crc kubenswrapper[4716]: I1207 16:04:33.071274 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 16:04:33 crc kubenswrapper[4716]: E1207 16:04:33.071642 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 16:04:33.571626061 +0000 UTC m=+136.261910973 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:04:33 crc kubenswrapper[4716]: I1207 16:04:33.174713 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d257\" (UID: \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d257" Dec 07 16:04:33 crc kubenswrapper[4716]: E1207 16:04:33.174982 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 16:04:33.67497169 +0000 UTC m=+136.365256602 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d257" (UID: "75569b15-b1c1-46bf-83f2-cd6eb0a38381") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:04:33 crc kubenswrapper[4716]: W1207 16:04:33.212683 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9345bd64_bb11_4d7b_96ca_9ae4a32b4e60.slice/crio-e0ea7a1014d996e339ec9ce938a72917dc4ca1c07995d0b75d319bb5b97432d3 WatchSource:0}: Error finding container e0ea7a1014d996e339ec9ce938a72917dc4ca1c07995d0b75d319bb5b97432d3: Status 404 returned error can't find the container with id e0ea7a1014d996e339ec9ce938a72917dc4ca1c07995d0b75d319bb5b97432d3 Dec 07 16:04:33 crc kubenswrapper[4716]: I1207 16:04:33.282757 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 16:04:33 crc kubenswrapper[4716]: E1207 16:04:33.283143 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 16:04:33.783128997 +0000 UTC m=+136.473413909 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:04:33 crc kubenswrapper[4716]: I1207 16:04:33.389659 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d257\" (UID: \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d257" Dec 07 16:04:33 crc kubenswrapper[4716]: E1207 16:04:33.390012 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 16:04:33.890001037 +0000 UTC m=+136.580285949 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d257" (UID: "75569b15-b1c1-46bf-83f2-cd6eb0a38381") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:04:33 crc kubenswrapper[4716]: I1207 16:04:33.490601 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 16:04:33 crc kubenswrapper[4716]: E1207 16:04:33.491240 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 16:04:33.991215264 +0000 UTC m=+136.681500176 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:04:33 crc kubenswrapper[4716]: I1207 16:04:33.591986 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d257\" (UID: \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d257" Dec 07 16:04:33 crc kubenswrapper[4716]: E1207 16:04:33.592352 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 16:04:34.092330708 +0000 UTC m=+136.782615620 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d257" (UID: "75569b15-b1c1-46bf-83f2-cd6eb0a38381") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:04:33 crc kubenswrapper[4716]: I1207 16:04:33.699898 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 16:04:33 crc kubenswrapper[4716]: E1207 16:04:33.700277 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 16:04:34.200262039 +0000 UTC m=+136.890546951 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:04:33 crc kubenswrapper[4716]: I1207 16:04:33.738594 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-82bvt"] Dec 07 16:04:33 crc kubenswrapper[4716]: I1207 16:04:33.740205 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-sq5lw"] Dec 07 16:04:33 crc kubenswrapper[4716]: I1207 16:04:33.743565 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-pszqv" event={"ID":"51fb93ab-ea9f-4fce-9fd1-a57958bc9c10","Type":"ContainerStarted","Data":"8ebd628052b709cb51c39c5e4e449f29cbfeab39691fd318328dff8fa7746b72"} Dec 07 16:04:33 crc kubenswrapper[4716]: I1207 16:04:33.744003 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-pszqv" Dec 07 16:04:33 crc kubenswrapper[4716]: I1207 16:04:33.754117 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-d49gn" event={"ID":"8439dba6-4f32-4c02-9aad-27b9c40f7575","Type":"ContainerStarted","Data":"85d78d140dad6592d416bac8c32c4bc1f62992bf62bc9dfd5e4b7b7ab1d09eaa"} Dec 07 16:04:33 crc kubenswrapper[4716]: I1207 16:04:33.792831 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-pszqv" podStartSLOduration=114.792813416 podStartE2EDuration="1m54.792813416s" podCreationTimestamp="2025-12-07 16:02:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:04:33.792326942 +0000 UTC m=+136.482611854" watchObservedRunningTime="2025-12-07 16:04:33.792813416 +0000 UTC m=+136.483098328" Dec 07 16:04:33 crc kubenswrapper[4716]: I1207 16:04:33.793537 4716 patch_prober.go:28] interesting pod/console-operator-58897d9998-pszqv container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.33:8443/readyz\": dial tcp 10.217.0.33:8443: connect: connection refused" start-of-body= Dec 07 16:04:33 crc kubenswrapper[4716]: I1207 16:04:33.793585 4716 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-pszqv" podUID="51fb93ab-ea9f-4fce-9fd1-a57958bc9c10" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.33:8443/readyz\": dial tcp 10.217.0.33:8443: connect: connection refused" Dec 07 16:04:33 crc kubenswrapper[4716]: I1207 16:04:33.793654 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-d7c2c" event={"ID":"18deffad-cf79-40dd-95f3-ed8410002584","Type":"ContainerStarted","Data":"62c7a0df3bd49b0bb9287ad8a3ead4217defe8fe5e41e777b335050e03dd6fa3"} Dec 07 16:04:33 crc kubenswrapper[4716]: I1207 16:04:33.801509 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d257\" (UID: \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d257" Dec 07 16:04:33 crc kubenswrapper[4716]: E1207 16:04:33.802287 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 16:04:34.302269109 +0000 UTC m=+136.992554021 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d257" (UID: "75569b15-b1c1-46bf-83f2-cd6eb0a38381") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:04:33 crc kubenswrapper[4716]: I1207 16:04:33.803308 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-v6899" event={"ID":"7a9cc84d-c02b-466f-9c1c-c1ae60e697ef","Type":"ContainerStarted","Data":"beab624532a9706a8e95993e9414eb9c16bb7e0e9dd3a1c7c1d54e439d0fb3c4"} Dec 07 16:04:33 crc kubenswrapper[4716]: I1207 16:04:33.818522 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-gcr9f" event={"ID":"372517a0-98df-460c-a7ed-a01624cb747a","Type":"ContainerStarted","Data":"64309beb0f19718a91406907509b388d8c12cabacce24540ffd4279e98e308b5"} Dec 07 16:04:33 crc kubenswrapper[4716]: I1207 16:04:33.904637 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 16:04:33 crc kubenswrapper[4716]: E1207 16:04:33.905740 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 16:04:34.4057236 +0000 UTC m=+137.096008512 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:04:33 crc kubenswrapper[4716]: I1207 16:04:33.908033 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wrw67" event={"ID":"92724c2e-d7fd-4673-ae84-8929541f2a13","Type":"ContainerStarted","Data":"0d55ff952ba7b5582f2a5a6aa7cb1e59224debc894f4abc3a6a54740acdf440e"} Dec 07 16:04:33 crc kubenswrapper[4716]: I1207 16:04:33.920143 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-ndp5v" event={"ID":"cbf3f592-bbb1-4596-8ae1-0e62aa366f0b","Type":"ContainerStarted","Data":"98b5a90edae2b9e0c3df9cc09a9dd99f59934f6e38e7da5499cddff1743f5663"} Dec 07 16:04:33 crc kubenswrapper[4716]: I1207 16:04:33.943597 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-zp6l4" event={"ID":"fc892d74-c6d2-431f-a52c-95a8f4e2d03c","Type":"ContainerStarted","Data":"0d7f1f58884588c39c02fb874998789d2cb9efe0de94f360fbaec41c1b9be2f4"} Dec 07 16:04:33 crc kubenswrapper[4716]: I1207 16:04:33.992362 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-ndp5v" podStartSLOduration=114.992346217 podStartE2EDuration="1m54.992346217s" podCreationTimestamp="2025-12-07 16:02:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:04:33.968554421 +0000 UTC m=+136.658839333" watchObservedRunningTime="2025-12-07 16:04:33.992346217 +0000 UTC m=+136.682631129" Dec 07 16:04:34 crc kubenswrapper[4716]: I1207 16:04:34.021469 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d257\" (UID: \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d257" Dec 07 16:04:34 crc kubenswrapper[4716]: E1207 16:04:34.022698 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 16:04:34.522683681 +0000 UTC m=+137.212968593 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d257" (UID: "75569b15-b1c1-46bf-83f2-cd6eb0a38381") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:04:34 crc kubenswrapper[4716]: I1207 16:04:34.046392 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-jjn8w" event={"ID":"cf1baeba-a0c2-4d91-ae76-eb4f0ceccd60","Type":"ContainerStarted","Data":"25860bff0d306dea4d443b9b1aade6b8767886cf4f0f35af773506fca804ac95"} Dec 07 16:04:34 crc kubenswrapper[4716]: I1207 16:04:34.065599 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6j7q2" event={"ID":"38c4a34b-864b-42d0-a3cb-90ede4226ccd","Type":"ContainerStarted","Data":"d17cb3689425a1d01924aed2774fd54a6b9299d2b5d285c391b889c6a2bf6309"} Dec 07 16:04:34 crc kubenswrapper[4716]: I1207 16:04:34.093673 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bcfw4" event={"ID":"a51dadb3-2af5-4c6c-a973-cdceb34cd79a","Type":"ContainerStarted","Data":"81dadafade81e62ccd9928b1249d69ca132131e3b7f5be3b5cc3b2eb0773f405"} Dec 07 16:04:34 crc kubenswrapper[4716]: I1207 16:04:34.110348 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4d2qh" event={"ID":"97d9c075-80c3-4dcb-aebc-649eec930413","Type":"ContainerStarted","Data":"d69ed019410a6b4f19b787ca716186162802145b814e2f0be53510d9806175c6"} Dec 07 16:04:34 crc kubenswrapper[4716]: I1207 16:04:34.111586 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4d2qh" Dec 07 16:04:34 crc kubenswrapper[4716]: I1207 16:04:34.119519 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-b257c" event={"ID":"1306f3b2-2982-494e-9db0-86c2e05b7a82","Type":"ContainerStarted","Data":"9a2cd34a1d750c2bb295f2debdd0982db8e8549d59cdce49453c509e8cfa91c4"} Dec 07 16:04:34 crc kubenswrapper[4716]: I1207 16:04:34.120570 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-b257c" Dec 07 16:04:34 crc kubenswrapper[4716]: I1207 16:04:34.123272 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 16:04:34 crc kubenswrapper[4716]: I1207 16:04:34.125709 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4d2qh" Dec 07 16:04:34 crc kubenswrapper[4716]: E1207 16:04:34.128642 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 16:04:34.628618874 +0000 UTC m=+137.318903786 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:04:34 crc kubenswrapper[4716]: I1207 16:04:34.129170 4716 patch_prober.go:28] interesting pod/downloads-7954f5f757-b257c container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" start-of-body= Dec 07 16:04:34 crc kubenswrapper[4716]: I1207 16:04:34.129204 4716 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-b257c" podUID="1306f3b2-2982-494e-9db0-86c2e05b7a82" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" Dec 07 16:04:34 crc kubenswrapper[4716]: I1207 16:04:34.141577 4716 generic.go:334] "Generic (PLEG): container finished" podID="23992ff8-b006-40d1-9048-1950f10fff7d" containerID="360d913d8793764aef3cb95ff7aeadecea4b13525c4234b25e1c8c9f93859584" exitCode=0 Dec 07 16:04:34 crc kubenswrapper[4716]: I1207 16:04:34.142322 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-x2nd8" event={"ID":"23992ff8-b006-40d1-9048-1950f10fff7d","Type":"ContainerDied","Data":"360d913d8793764aef3cb95ff7aeadecea4b13525c4234b25e1c8c9f93859584"} Dec 07 16:04:34 crc kubenswrapper[4716]: I1207 16:04:34.146020 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4d2qh" podStartSLOduration=114.145995515 podStartE2EDuration="1m54.145995515s" podCreationTimestamp="2025-12-07 16:02:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:04:34.145036157 +0000 UTC m=+136.835321069" watchObservedRunningTime="2025-12-07 16:04:34.145995515 +0000 UTC m=+136.836280427" Dec 07 16:04:34 crc kubenswrapper[4716]: I1207 16:04:34.146300 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-kkvsk" event={"ID":"9345bd64-bb11-4d7b-96ca-9ae4a32b4e60","Type":"ContainerStarted","Data":"e0ea7a1014d996e339ec9ce938a72917dc4ca1c07995d0b75d319bb5b97432d3"} Dec 07 16:04:34 crc kubenswrapper[4716]: I1207 16:04:34.153336 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-h5l29" event={"ID":"85bc0929-de17-4d2d-8b03-1a05d207fb55","Type":"ContainerStarted","Data":"49b276ad5824c4c2e62b4de651a8c4a26f7885e2a43c03c87f4b076871237804"} Dec 07 16:04:34 crc kubenswrapper[4716]: I1207 16:04:34.171594 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8sxfn" event={"ID":"9a0a9cf5-14e9-41e1-8c6c-27b7b3b8c7b4","Type":"ContainerStarted","Data":"bdc66869701bb86a2541203c8cec21dd39450ec02762fdd63459add90eb18dce"} Dec 07 16:04:34 crc kubenswrapper[4716]: I1207 16:04:34.180508 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-6xp6f" event={"ID":"f086c3a7-c2a4-4066-aed6-270b5dd6eef8","Type":"ContainerStarted","Data":"be45c23a9cb3a5a938387cb1eb14035d706e3826089428e8338306b878e54e1e"} Dec 07 16:04:34 crc kubenswrapper[4716]: I1207 16:04:34.186932 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-z587m" event={"ID":"ffff2831-338e-423e-81f2-c8d3b5de7785","Type":"ContainerStarted","Data":"70e6048ead4ed66e668777500ea8cf6f6fe15c89bca412f93a7a7cc1823076b3"} Dec 07 16:04:34 crc kubenswrapper[4716]: I1207 16:04:34.186975 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-tqctv"] Dec 07 16:04:34 crc kubenswrapper[4716]: I1207 16:04:34.187681 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-z587m" Dec 07 16:04:34 crc kubenswrapper[4716]: I1207 16:04:34.188753 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-j7zhr" event={"ID":"4df2ba74-d18e-4e94-bee1-703645b1a463","Type":"ContainerStarted","Data":"e35da83e70ccca189dc54f35e4ee89ccc822aef7a3b7f1ada2649c1a11043041"} Dec 07 16:04:34 crc kubenswrapper[4716]: I1207 16:04:34.194468 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-8sv58" event={"ID":"5b5665d7-a97a-4d47-b37e-bd3f1643c7cf","Type":"ContainerStarted","Data":"5650e5658485dd4df062a9a67bac02eaeac91210e98ac5eda9bff91e211cb7d6"} Dec 07 16:04:34 crc kubenswrapper[4716]: I1207 16:04:34.200260 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zjwcl" event={"ID":"915af7e7-19b0-4b1c-9e55-3471210ccf40","Type":"ContainerStarted","Data":"6d563ad3675dea9ef97c258906b87f6cb079e891b40e4a8d7b4d81d828705fe6"} Dec 07 16:04:34 crc kubenswrapper[4716]: I1207 16:04:34.200574 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-b257c" podStartSLOduration=115.200553357 podStartE2EDuration="1m55.200553357s" podCreationTimestamp="2025-12-07 16:02:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:04:34.199803476 +0000 UTC m=+136.890088388" watchObservedRunningTime="2025-12-07 16:04:34.200553357 +0000 UTC m=+136.890838269" Dec 07 16:04:34 crc kubenswrapper[4716]: I1207 16:04:34.217669 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-z587m" Dec 07 16:04:34 crc kubenswrapper[4716]: I1207 16:04:34.224003 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-64rzd" event={"ID":"580ae9d7-004b-4d66-ba7b-1566de502f66","Type":"ContainerStarted","Data":"961a1fc7f9ade2938408f6933c4699092d1bc9a5eca6bcf56c766cab398e276e"} Dec 07 16:04:34 crc kubenswrapper[4716]: I1207 16:04:34.226294 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d257\" (UID: \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d257" Dec 07 16:04:34 crc kubenswrapper[4716]: E1207 16:04:34.228117 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 16:04:34.728098991 +0000 UTC m=+137.418383993 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d257" (UID: "75569b15-b1c1-46bf-83f2-cd6eb0a38381") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:04:34 crc kubenswrapper[4716]: I1207 16:04:34.247144 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bwsl4"] Dec 07 16:04:34 crc kubenswrapper[4716]: I1207 16:04:34.278270 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mmfjq"] Dec 07 16:04:34 crc kubenswrapper[4716]: I1207 16:04:34.299178 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-w7xhs"] Dec 07 16:04:34 crc kubenswrapper[4716]: I1207 16:04:34.299384 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-89v7m"] Dec 07 16:04:34 crc kubenswrapper[4716]: I1207 16:04:34.300771 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-6xp6f" podStartSLOduration=115.300751135 podStartE2EDuration="1m55.300751135s" podCreationTimestamp="2025-12-07 16:02:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:04:34.270260386 +0000 UTC m=+136.960545298" watchObservedRunningTime="2025-12-07 16:04:34.300751135 +0000 UTC m=+136.991036047" Dec 07 16:04:34 crc kubenswrapper[4716]: I1207 16:04:34.307379 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-h5l29" podStartSLOduration=115.307362115 podStartE2EDuration="1m55.307362115s" podCreationTimestamp="2025-12-07 16:02:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:04:34.299928272 +0000 UTC m=+136.990213184" watchObservedRunningTime="2025-12-07 16:04:34.307362115 +0000 UTC m=+136.997647027" Dec 07 16:04:34 crc kubenswrapper[4716]: I1207 16:04:34.327452 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 16:04:34 crc kubenswrapper[4716]: E1207 16:04:34.328419 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 16:04:34.828402892 +0000 UTC m=+137.518687804 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:04:34 crc kubenswrapper[4716]: I1207 16:04:34.343790 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zjwcl" podStartSLOduration=115.343775725 podStartE2EDuration="1m55.343775725s" podCreationTimestamp="2025-12-07 16:02:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:04:34.342706314 +0000 UTC m=+137.032991226" watchObservedRunningTime="2025-12-07 16:04:34.343775725 +0000 UTC m=+137.034060637" Dec 07 16:04:34 crc kubenswrapper[4716]: I1207 16:04:34.387143 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-l9dgz"] Dec 07 16:04:34 crc kubenswrapper[4716]: I1207 16:04:34.389376 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-h6gnn"] Dec 07 16:04:34 crc kubenswrapper[4716]: I1207 16:04:34.389609 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-j7zhr" podStartSLOduration=115.389600676 podStartE2EDuration="1m55.389600676s" podCreationTimestamp="2025-12-07 16:02:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:04:34.384593891 +0000 UTC m=+137.074878803" watchObservedRunningTime="2025-12-07 16:04:34.389600676 +0000 UTC m=+137.079885588" Dec 07 16:04:34 crc kubenswrapper[4716]: I1207 16:04:34.429863 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d257\" (UID: \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d257" Dec 07 16:04:34 crc kubenswrapper[4716]: E1207 16:04:34.430987 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 16:04:34.930976108 +0000 UTC m=+137.621261020 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d257" (UID: "75569b15-b1c1-46bf-83f2-cd6eb0a38381") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:04:34 crc kubenswrapper[4716]: W1207 16:04:34.472325 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6b4e49ef_27ab_417c_9c6b_1ad7f215604c.slice/crio-31044f4aba44d3f483ae8077aa8945059188c05b78db3927113fcc76108b3886 WatchSource:0}: Error finding container 31044f4aba44d3f483ae8077aa8945059188c05b78db3927113fcc76108b3886: Status 404 returned error can't find the container with id 31044f4aba44d3f483ae8077aa8945059188c05b78db3927113fcc76108b3886 Dec 07 16:04:34 crc kubenswrapper[4716]: I1207 16:04:34.490340 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-h5l29" Dec 07 16:04:34 crc kubenswrapper[4716]: I1207 16:04:34.495433 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rb966"] Dec 07 16:04:34 crc kubenswrapper[4716]: I1207 16:04:34.505954 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-z587m" podStartSLOduration=115.505936608 podStartE2EDuration="1m55.505936608s" podCreationTimestamp="2025-12-07 16:02:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:04:34.485182861 +0000 UTC m=+137.175467773" watchObservedRunningTime="2025-12-07 16:04:34.505936608 +0000 UTC m=+137.196221520" Dec 07 16:04:34 crc kubenswrapper[4716]: I1207 16:04:34.540290 4716 patch_prober.go:28] interesting pod/router-default-5444994796-h5l29 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 07 16:04:34 crc kubenswrapper[4716]: [-]has-synced failed: reason withheld Dec 07 16:04:34 crc kubenswrapper[4716]: [+]process-running ok Dec 07 16:04:34 crc kubenswrapper[4716]: healthz check failed Dec 07 16:04:34 crc kubenswrapper[4716]: I1207 16:04:34.540359 4716 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h5l29" podUID="85bc0929-de17-4d2d-8b03-1a05d207fb55" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 07 16:04:34 crc kubenswrapper[4716]: I1207 16:04:34.546670 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29418720-x5rqz"] Dec 07 16:04:34 crc kubenswrapper[4716]: I1207 16:04:34.547701 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 16:04:34 crc kubenswrapper[4716]: E1207 16:04:34.548137 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 16:04:35.048118104 +0000 UTC m=+137.738403016 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:04:34 crc kubenswrapper[4716]: I1207 16:04:34.573975 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-v9d58"] Dec 07 16:04:34 crc kubenswrapper[4716]: I1207 16:04:34.656681 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-n4492"] Dec 07 16:04:34 crc kubenswrapper[4716]: I1207 16:04:34.658237 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d257\" (UID: \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d257" Dec 07 16:04:34 crc kubenswrapper[4716]: E1207 16:04:34.658567 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 16:04:35.158554977 +0000 UTC m=+137.848839889 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d257" (UID: "75569b15-b1c1-46bf-83f2-cd6eb0a38381") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:04:34 crc kubenswrapper[4716]: W1207 16:04:34.712894 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod98efd33b_de1f_4414_b9cc_55fb0d423784.slice/crio-9c914c728e0bde869d16f15f74ade82d4b5d12a2b84894b9f9af328f85649956 WatchSource:0}: Error finding container 9c914c728e0bde869d16f15f74ade82d4b5d12a2b84894b9f9af328f85649956: Status 404 returned error can't find the container with id 9c914c728e0bde869d16f15f74ade82d4b5d12a2b84894b9f9af328f85649956 Dec 07 16:04:34 crc kubenswrapper[4716]: I1207 16:04:34.763269 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 16:04:34 crc kubenswrapper[4716]: E1207 16:04:34.763624 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 16:04:35.263608574 +0000 UTC m=+137.953893486 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:04:34 crc kubenswrapper[4716]: I1207 16:04:34.865766 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d257\" (UID: \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d257" Dec 07 16:04:34 crc kubenswrapper[4716]: E1207 16:04:34.866437 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 16:04:35.366426108 +0000 UTC m=+138.056711010 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d257" (UID: "75569b15-b1c1-46bf-83f2-cd6eb0a38381") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:04:34 crc kubenswrapper[4716]: I1207 16:04:34.966860 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 16:04:34 crc kubenswrapper[4716]: E1207 16:04:34.967064 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 16:04:35.467030807 +0000 UTC m=+138.157315719 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:04:34 crc kubenswrapper[4716]: I1207 16:04:34.967176 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d257\" (UID: \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d257" Dec 07 16:04:34 crc kubenswrapper[4716]: E1207 16:04:34.967448 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 16:04:35.467435339 +0000 UTC m=+138.157720251 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d257" (UID: "75569b15-b1c1-46bf-83f2-cd6eb0a38381") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:04:35 crc kubenswrapper[4716]: I1207 16:04:35.067858 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 16:04:35 crc kubenswrapper[4716]: E1207 16:04:35.067954 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 16:04:35.567933625 +0000 UTC m=+138.258218537 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:04:35 crc kubenswrapper[4716]: I1207 16:04:35.068431 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d257\" (UID: \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d257" Dec 07 16:04:35 crc kubenswrapper[4716]: E1207 16:04:35.068707 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 16:04:35.568696477 +0000 UTC m=+138.258981389 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d257" (UID: "75569b15-b1c1-46bf-83f2-cd6eb0a38381") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:04:35 crc kubenswrapper[4716]: I1207 16:04:35.176654 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 16:04:35 crc kubenswrapper[4716]: E1207 16:04:35.176990 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 16:04:35.676976108 +0000 UTC m=+138.367261020 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:04:35 crc kubenswrapper[4716]: I1207 16:04:35.279431 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d257\" (UID: \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d257" Dec 07 16:04:35 crc kubenswrapper[4716]: E1207 16:04:35.280019 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 16:04:35.780006137 +0000 UTC m=+138.470291049 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d257" (UID: "75569b15-b1c1-46bf-83f2-cd6eb0a38381") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:04:35 crc kubenswrapper[4716]: I1207 16:04:35.334159 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mmfjq" event={"ID":"db812416-33e4-462c-a28f-ec40b629dc37","Type":"ContainerStarted","Data":"eef9264e2615542e14d71449af433ed7ed3cfa56c489d3b7aea7501ab6f27697"} Dec 07 16:04:35 crc kubenswrapper[4716]: I1207 16:04:35.334224 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mmfjq" event={"ID":"db812416-33e4-462c-a28f-ec40b629dc37","Type":"ContainerStarted","Data":"ca996e2eec90ede56a32c075439ae003bb851c48944a7f53af8acbbb5fbca73f"} Dec 07 16:04:35 crc kubenswrapper[4716]: I1207 16:04:35.340402 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-n4492" event={"ID":"db753ef2-2939-42dd-886d-17a223c47360","Type":"ContainerStarted","Data":"a877628e1e068b332f356c99addd20e4b048c9a2606b822f288f4e3b107dec5e"} Dec 07 16:04:35 crc kubenswrapper[4716]: I1207 16:04:35.362524 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-jjn8w" event={"ID":"cf1baeba-a0c2-4d91-ae76-eb4f0ceccd60","Type":"ContainerStarted","Data":"94b33a0eb39ff644ac874a13d3c6e2a74e26848d5251d8c340c4aa05b277a84b"} Dec 07 16:04:35 crc kubenswrapper[4716]: I1207 16:04:35.362577 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-jjn8w" event={"ID":"cf1baeba-a0c2-4d91-ae76-eb4f0ceccd60","Type":"ContainerStarted","Data":"0621cfa5c3d0380c326130672383a14e8e35e1f56e7aa263bd508a37f83f67c9"} Dec 07 16:04:35 crc kubenswrapper[4716]: I1207 16:04:35.381576 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 16:04:35 crc kubenswrapper[4716]: E1207 16:04:35.381903 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 16:04:35.881845432 +0000 UTC m=+138.572130354 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:04:35 crc kubenswrapper[4716]: I1207 16:04:35.382460 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d257\" (UID: \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d257" Dec 07 16:04:35 crc kubenswrapper[4716]: E1207 16:04:35.383146 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 16:04:35.883138049 +0000 UTC m=+138.573422961 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d257" (UID: "75569b15-b1c1-46bf-83f2-cd6eb0a38381") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:04:35 crc kubenswrapper[4716]: I1207 16:04:35.399292 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-8sv58" event={"ID":"5b5665d7-a97a-4d47-b37e-bd3f1643c7cf","Type":"ContainerStarted","Data":"8d68fc0214ef7183adb67274a146dfa1652eb2bc20e4bf5b996c5c8863192b30"} Dec 07 16:04:35 crc kubenswrapper[4716]: I1207 16:04:35.404566 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mmfjq" podStartSLOduration=116.404536916 podStartE2EDuration="1m56.404536916s" podCreationTimestamp="2025-12-07 16:02:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:04:35.400960944 +0000 UTC m=+138.091245856" watchObservedRunningTime="2025-12-07 16:04:35.404536916 +0000 UTC m=+138.094821828" Dec 07 16:04:35 crc kubenswrapper[4716]: I1207 16:04:35.451041 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-v6899" event={"ID":"7a9cc84d-c02b-466f-9c1c-c1ae60e697ef","Type":"ContainerStarted","Data":"4d025fd1d535955cc38cb815b9eaf0aba61650c6a85b46e10fdbf4392568810a"} Dec 07 16:04:35 crc kubenswrapper[4716]: I1207 16:04:35.451344 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-v6899" Dec 07 16:04:35 crc kubenswrapper[4716]: I1207 16:04:35.467674 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7ff64" event={"ID":"6e5a490b-b9b3-4f19-aebf-14f6c1efbaa2","Type":"ContainerStarted","Data":"88cb3312da63a2bc61316571dad211cf00699e760cc73b8193423f17b48aae50"} Dec 07 16:04:35 crc kubenswrapper[4716]: I1207 16:04:35.483692 4716 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-v6899 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.32:8080/healthz\": dial tcp 10.217.0.32:8080: connect: connection refused" start-of-body= Dec 07 16:04:35 crc kubenswrapper[4716]: I1207 16:04:35.483744 4716 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-v6899" podUID="7a9cc84d-c02b-466f-9c1c-c1ae60e697ef" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.32:8080/healthz\": dial tcp 10.217.0.32:8080: connect: connection refused" Dec 07 16:04:35 crc kubenswrapper[4716]: I1207 16:04:35.484144 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 16:04:35 crc kubenswrapper[4716]: E1207 16:04:35.485093 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 16:04:35.985044757 +0000 UTC m=+138.675329669 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:04:35 crc kubenswrapper[4716]: I1207 16:04:35.490722 4716 patch_prober.go:28] interesting pod/router-default-5444994796-h5l29 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 07 16:04:35 crc kubenswrapper[4716]: [-]has-synced failed: reason withheld Dec 07 16:04:35 crc kubenswrapper[4716]: [+]process-running ok Dec 07 16:04:35 crc kubenswrapper[4716]: healthz check failed Dec 07 16:04:35 crc kubenswrapper[4716]: I1207 16:04:35.490945 4716 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h5l29" podUID="85bc0929-de17-4d2d-8b03-1a05d207fb55" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 07 16:04:35 crc kubenswrapper[4716]: I1207 16:04:35.491032 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-l9dgz" event={"ID":"55935fe0-005a-4f9f-80cd-9fa3348cc90a","Type":"ContainerStarted","Data":"f7701cd0e8230e1334245557b6a9ad11cb149e9df79dd6a2717c254aa6b1bd5f"} Dec 07 16:04:35 crc kubenswrapper[4716]: I1207 16:04:35.491506 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-l9dgz" Dec 07 16:04:35 crc kubenswrapper[4716]: I1207 16:04:35.504836 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-9bs98" event={"ID":"c94992de-685b-42c3-8100-93996239d0b0","Type":"ContainerStarted","Data":"466d985974bda0651e3fdb3d99e117c829bfe97d8bf2589fadc80ec68568c669"} Dec 07 16:04:35 crc kubenswrapper[4716]: I1207 16:04:35.505848 4716 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-l9dgz container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.38:5443/healthz\": dial tcp 10.217.0.38:5443: connect: connection refused" start-of-body= Dec 07 16:04:35 crc kubenswrapper[4716]: I1207 16:04:35.505890 4716 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-l9dgz" podUID="55935fe0-005a-4f9f-80cd-9fa3348cc90a" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.38:5443/healthz\": dial tcp 10.217.0.38:5443: connect: connection refused" Dec 07 16:04:35 crc kubenswrapper[4716]: I1207 16:04:35.554432 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-j7zhr" event={"ID":"4df2ba74-d18e-4e94-bee1-703645b1a463","Type":"ContainerStarted","Data":"1aa62f0d943838ad33bdf6f4aa9bc0f288bac95ab14bcd55bf44fd8e1afc8ff2"} Dec 07 16:04:35 crc kubenswrapper[4716]: I1207 16:04:35.580015 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29418720-x5rqz" event={"ID":"98efd33b-de1f-4414-b9cc-55fb0d423784","Type":"ContainerStarted","Data":"9c914c728e0bde869d16f15f74ade82d4b5d12a2b84894b9f9af328f85649956"} Dec 07 16:04:35 crc kubenswrapper[4716]: I1207 16:04:35.585894 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d257\" (UID: \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d257" Dec 07 16:04:35 crc kubenswrapper[4716]: I1207 16:04:35.587569 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-x2nd8" event={"ID":"23992ff8-b006-40d1-9048-1950f10fff7d","Type":"ContainerStarted","Data":"6114a8c12e90ede47b967299c8c9dc2fdc0e2ac017a9577f4b947dc87b584973"} Dec 07 16:04:35 crc kubenswrapper[4716]: I1207 16:04:35.594381 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-x2nd8" Dec 07 16:04:35 crc kubenswrapper[4716]: E1207 16:04:35.587628 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 16:04:36.087615902 +0000 UTC m=+138.777900814 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d257" (UID: "75569b15-b1c1-46bf-83f2-cd6eb0a38381") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:04:35 crc kubenswrapper[4716]: I1207 16:04:35.596141 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-zp6l4" event={"ID":"fc892d74-c6d2-431f-a52c-95a8f4e2d03c","Type":"ContainerStarted","Data":"7dd6ad63ca1474893027d94498cd687716d4b99552ad2e78e5444823c5a849ae"} Dec 07 16:04:35 crc kubenswrapper[4716]: I1207 16:04:35.610876 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-w7xhs" event={"ID":"38ef2332-5fee-44a2-b56c-bc6db5fad2e0","Type":"ContainerStarted","Data":"59f42ce5afc71d5c70b4e12228b9488b2bc65ae4a2de69214e35477c95de5d0a"} Dec 07 16:04:35 crc kubenswrapper[4716]: I1207 16:04:35.613385 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-gcr9f" event={"ID":"372517a0-98df-460c-a7ed-a01624cb747a","Type":"ContainerStarted","Data":"3a2371563594ee692e232c27e8edc16ae5daa40fef90cbf13e6ac3855378cf6b"} Dec 07 16:04:35 crc kubenswrapper[4716]: I1207 16:04:35.641525 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-89v7m" event={"ID":"6b4e49ef-27ab-417c-9c6b-1ad7f215604c","Type":"ContainerStarted","Data":"bdaf469a8b0a5db58f4a5bf18958307d54300fec9d2c44623e84b952dc28735c"} Dec 07 16:04:35 crc kubenswrapper[4716]: I1207 16:04:35.641569 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-89v7m" event={"ID":"6b4e49ef-27ab-417c-9c6b-1ad7f215604c","Type":"ContainerStarted","Data":"31044f4aba44d3f483ae8077aa8945059188c05b78db3927113fcc76108b3886"} Dec 07 16:04:35 crc kubenswrapper[4716]: I1207 16:04:35.674627 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-jjn8w" podStartSLOduration=116.674610439 podStartE2EDuration="1m56.674610439s" podCreationTimestamp="2025-12-07 16:02:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:04:35.605719915 +0000 UTC m=+138.296004837" watchObservedRunningTime="2025-12-07 16:04:35.674610439 +0000 UTC m=+138.364895351" Dec 07 16:04:35 crc kubenswrapper[4716]: I1207 16:04:35.676451 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-8sv58" podStartSLOduration=115.676442673 podStartE2EDuration="1m55.676442673s" podCreationTimestamp="2025-12-07 16:02:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:04:35.673999232 +0000 UTC m=+138.364284134" watchObservedRunningTime="2025-12-07 16:04:35.676442673 +0000 UTC m=+138.366727585" Dec 07 16:04:35 crc kubenswrapper[4716]: I1207 16:04:35.686427 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 16:04:35 crc kubenswrapper[4716]: E1207 16:04:35.687759 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 16:04:36.187742498 +0000 UTC m=+138.878027410 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:04:35 crc kubenswrapper[4716]: I1207 16:04:35.695879 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j26br" event={"ID":"297ee903-c479-460e-80df-a0c508c45fbe","Type":"ContainerStarted","Data":"238f56c2418b13847a7aa50f2102ce231856650a63b2fd8717dd5ab9eaeb81f3"} Dec 07 16:04:35 crc kubenswrapper[4716]: I1207 16:04:35.709113 4716 generic.go:334] "Generic (PLEG): container finished" podID="6f7305ef-d76f-4be3-9151-7247f7590fe0" containerID="73dc15ef16ae836f17ea4705740b31502f81e524cda2fdc61ccddef640315a12" exitCode=0 Dec 07 16:04:35 crc kubenswrapper[4716]: I1207 16:04:35.709255 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-gc4cp" event={"ID":"6f7305ef-d76f-4be3-9151-7247f7590fe0","Type":"ContainerDied","Data":"73dc15ef16ae836f17ea4705740b31502f81e524cda2fdc61ccddef640315a12"} Dec 07 16:04:35 crc kubenswrapper[4716]: I1207 16:04:35.724210 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-h6gnn" event={"ID":"74c41933-56ed-4f49-a808-1d7f1f1829b2","Type":"ContainerStarted","Data":"e768819258f477ecdd78e60bdebd36bed35aa999bcd3f734791da3f1938b71a8"} Dec 07 16:04:35 crc kubenswrapper[4716]: I1207 16:04:35.745265 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-x2nd8" podStartSLOduration=116.745249536 podStartE2EDuration="1m56.745249536s" podCreationTimestamp="2025-12-07 16:02:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:04:35.743535996 +0000 UTC m=+138.433820908" watchObservedRunningTime="2025-12-07 16:04:35.745249536 +0000 UTC m=+138.435534448" Dec 07 16:04:35 crc kubenswrapper[4716]: I1207 16:04:35.769134 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-kkvsk" event={"ID":"9345bd64-bb11-4d7b-96ca-9ae4a32b4e60","Type":"ContainerStarted","Data":"951c2c22d3954d0d1a1d8f86f3f035b50d4087bec50d15b9f3e1c94bd5869521"} Dec 07 16:04:35 crc kubenswrapper[4716]: I1207 16:04:35.769175 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-kkvsk" event={"ID":"9345bd64-bb11-4d7b-96ca-9ae4a32b4e60","Type":"ContainerStarted","Data":"73db741ec6b91c5653032d77d2a4a879043ccce26b33c8d338e7376bea722343"} Dec 07 16:04:35 crc kubenswrapper[4716]: I1207 16:04:35.796207 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d257\" (UID: \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d257" Dec 07 16:04:35 crc kubenswrapper[4716]: I1207 16:04:35.796288 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-v9d58" event={"ID":"5d2c595e-c4a6-446d-836a-766d1ec44516","Type":"ContainerStarted","Data":"dcb93418da467f42ddeeea5e9ff8e67accf69c10c5316bbaa3a5370b8e68cf74"} Dec 07 16:04:35 crc kubenswrapper[4716]: I1207 16:04:35.796726 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-7ff64" podStartSLOduration=116.796708429 podStartE2EDuration="1m56.796708429s" podCreationTimestamp="2025-12-07 16:02:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:04:35.795461482 +0000 UTC m=+138.485746394" watchObservedRunningTime="2025-12-07 16:04:35.796708429 +0000 UTC m=+138.486993341" Dec 07 16:04:35 crc kubenswrapper[4716]: E1207 16:04:35.797358 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 16:04:36.297346837 +0000 UTC m=+138.987631749 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d257" (UID: "75569b15-b1c1-46bf-83f2-cd6eb0a38381") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:04:35 crc kubenswrapper[4716]: I1207 16:04:35.804933 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6j7q2" event={"ID":"38c4a34b-864b-42d0-a3cb-90ede4226ccd","Type":"ContainerStarted","Data":"bc66700f6901f7a2562ee8d1fbc49ed66115ffab37c993b1413b9c02a6f324d4"} Dec 07 16:04:35 crc kubenswrapper[4716]: I1207 16:04:35.806271 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6j7q2" Dec 07 16:04:35 crc kubenswrapper[4716]: I1207 16:04:35.843847 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rb966" event={"ID":"060bb436-de36-408a-b742-f3768a575801","Type":"ContainerStarted","Data":"a3af4ff6ec09ffc89121f770170b90348b6b24873aca0b5095c04c7929d0f388"} Dec 07 16:04:35 crc kubenswrapper[4716]: I1207 16:04:35.855545 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6j7q2" Dec 07 16:04:35 crc kubenswrapper[4716]: I1207 16:04:35.871784 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29418720-x5rqz" podStartSLOduration=116.871758572 podStartE2EDuration="1m56.871758572s" podCreationTimestamp="2025-12-07 16:02:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:04:35.871365401 +0000 UTC m=+138.561650323" watchObservedRunningTime="2025-12-07 16:04:35.871758572 +0000 UTC m=+138.562043484" Dec 07 16:04:35 crc kubenswrapper[4716]: I1207 16:04:35.878153 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-v6899" podStartSLOduration=115.878136935 podStartE2EDuration="1m55.878136935s" podCreationTimestamp="2025-12-07 16:02:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:04:35.833473898 +0000 UTC m=+138.523758810" watchObservedRunningTime="2025-12-07 16:04:35.878136935 +0000 UTC m=+138.568421847" Dec 07 16:04:35 crc kubenswrapper[4716]: I1207 16:04:35.893403 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-64rzd" event={"ID":"580ae9d7-004b-4d66-ba7b-1566de502f66","Type":"ContainerStarted","Data":"0f44a8ea95e054fb2bbf943c9980bfaf6024f520a587f0215a3382016b8a4303"} Dec 07 16:04:35 crc kubenswrapper[4716]: I1207 16:04:35.897573 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 16:04:35 crc kubenswrapper[4716]: E1207 16:04:35.897657 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 16:04:36.397637857 +0000 UTC m=+139.087922769 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:04:35 crc kubenswrapper[4716]: I1207 16:04:35.903223 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bcfw4" event={"ID":"a51dadb3-2af5-4c6c-a973-cdceb34cd79a","Type":"ContainerStarted","Data":"807e4f3a31a13a2740e94d6cfa919eb15b3a1f10d4b95e29be63c37066407f04"} Dec 07 16:04:35 crc kubenswrapper[4716]: I1207 16:04:35.904782 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d257\" (UID: \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d257" Dec 07 16:04:35 crc kubenswrapper[4716]: E1207 16:04:35.906282 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 16:04:36.406262976 +0000 UTC m=+139.096547888 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d257" (UID: "75569b15-b1c1-46bf-83f2-cd6eb0a38381") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:04:35 crc kubenswrapper[4716]: I1207 16:04:35.942532 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-gcr9f" podStartSLOduration=6.94250406 podStartE2EDuration="6.94250406s" podCreationTimestamp="2025-12-07 16:04:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:04:35.931896624 +0000 UTC m=+138.622181536" watchObservedRunningTime="2025-12-07 16:04:35.94250406 +0000 UTC m=+138.632788972" Dec 07 16:04:35 crc kubenswrapper[4716]: I1207 16:04:35.955892 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-sq5lw" event={"ID":"d27e02ff-86da-4078-934b-fd472b471f69","Type":"ContainerStarted","Data":"9f7a79b9524cd05a5a5e13fe851bf3a512e07fe786d45fa86bacf222cb4aa478"} Dec 07 16:04:35 crc kubenswrapper[4716]: I1207 16:04:35.955944 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-sq5lw" event={"ID":"d27e02ff-86da-4078-934b-fd472b471f69","Type":"ContainerStarted","Data":"d81622f2cb717bb091b4cc86a1451f5ec6e6afdf5d21817997fd10db2135e978"} Dec 07 16:04:35 crc kubenswrapper[4716]: I1207 16:04:35.978748 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-9bs98" podStartSLOduration=116.978726174 podStartE2EDuration="1m56.978726174s" podCreationTimestamp="2025-12-07 16:02:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:04:35.978330773 +0000 UTC m=+138.668615685" watchObservedRunningTime="2025-12-07 16:04:35.978726174 +0000 UTC m=+138.669011086" Dec 07 16:04:35 crc kubenswrapper[4716]: I1207 16:04:35.985924 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-d7c2c" event={"ID":"18deffad-cf79-40dd-95f3-ed8410002584","Type":"ContainerStarted","Data":"9e16e013db219487042c12ee3d2ad62d947d0555b3f866f1736ba47a8a961246"} Dec 07 16:04:35 crc kubenswrapper[4716]: I1207 16:04:35.990226 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wrw67" event={"ID":"92724c2e-d7fd-4673-ae84-8929541f2a13","Type":"ContainerStarted","Data":"b8f185f15811a342285f5b5f9f377637fc464a1434b2932be17ff94675ba9440"} Dec 07 16:04:36 crc kubenswrapper[4716]: I1207 16:04:36.013421 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 16:04:36 crc kubenswrapper[4716]: E1207 16:04:36.014548 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 16:04:36.514531987 +0000 UTC m=+139.204816899 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:04:36 crc kubenswrapper[4716]: I1207 16:04:36.018461 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j26br" Dec 07 16:04:36 crc kubenswrapper[4716]: I1207 16:04:36.018521 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j26br" Dec 07 16:04:36 crc kubenswrapper[4716]: I1207 16:04:36.025523 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-l9dgz" podStartSLOduration=116.025499203 podStartE2EDuration="1m56.025499203s" podCreationTimestamp="2025-12-07 16:02:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:04:36.016261986 +0000 UTC m=+138.706546898" watchObservedRunningTime="2025-12-07 16:04:36.025499203 +0000 UTC m=+138.715784115" Dec 07 16:04:36 crc kubenswrapper[4716]: I1207 16:04:36.044063 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-6j7q2" podStartSLOduration=116.044044046 podStartE2EDuration="1m56.044044046s" podCreationTimestamp="2025-12-07 16:02:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:04:36.040272678 +0000 UTC m=+138.730557590" watchObservedRunningTime="2025-12-07 16:04:36.044044046 +0000 UTC m=+138.734328958" Dec 07 16:04:36 crc kubenswrapper[4716]: I1207 16:04:36.106105 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bwsl4" event={"ID":"fe74013f-e94b-4600-8d02-6f73efbd4212","Type":"ContainerStarted","Data":"abcca99d7e17538c36959a23ca03e7b10f8facc9c47b014cd4a69f4ac0de3de7"} Dec 07 16:04:36 crc kubenswrapper[4716]: I1207 16:04:36.106542 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bwsl4" Dec 07 16:04:36 crc kubenswrapper[4716]: I1207 16:04:36.106562 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bwsl4" event={"ID":"fe74013f-e94b-4600-8d02-6f73efbd4212","Type":"ContainerStarted","Data":"4f21e3081c2976047e58a31f253b35ee55c0bec5286da4cd2c03e5ee0a943dcc"} Dec 07 16:04:36 crc kubenswrapper[4716]: I1207 16:04:36.107475 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j26br" Dec 07 16:04:36 crc kubenswrapper[4716]: I1207 16:04:36.108718 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wrw67" podStartSLOduration=117.108694 podStartE2EDuration="1m57.108694s" podCreationTimestamp="2025-12-07 16:02:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:04:36.104841169 +0000 UTC m=+138.795126081" watchObservedRunningTime="2025-12-07 16:04:36.108694 +0000 UTC m=+138.798978912" Dec 07 16:04:36 crc kubenswrapper[4716]: I1207 16:04:36.112430 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-tqctv" event={"ID":"1e623490-3813-4a3a-b7bf-beb8b870ed9e","Type":"ContainerStarted","Data":"7055150d91a85bd75ece39587ba77874aef5c24bee257553c257c2d8f7220e62"} Dec 07 16:04:36 crc kubenswrapper[4716]: I1207 16:04:36.112474 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-tqctv" event={"ID":"1e623490-3813-4a3a-b7bf-beb8b870ed9e","Type":"ContainerStarted","Data":"2811668a961c458bc2c690d2130edbd27e173502d53ca775d7b968e05cee101d"} Dec 07 16:04:36 crc kubenswrapper[4716]: I1207 16:04:36.116383 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d257\" (UID: \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d257" Dec 07 16:04:36 crc kubenswrapper[4716]: E1207 16:04:36.118801 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 16:04:36.618786001 +0000 UTC m=+139.309070913 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d257" (UID: "75569b15-b1c1-46bf-83f2-cd6eb0a38381") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:04:36 crc kubenswrapper[4716]: I1207 16:04:36.124676 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-82bvt" event={"ID":"af51b694-cd0f-4c33-ba6e-1f4666178820","Type":"ContainerStarted","Data":"fd6841b4abc42f8242997efe6e7a2d4294f4e1b609490427d722a146c2db5864"} Dec 07 16:04:36 crc kubenswrapper[4716]: I1207 16:04:36.150108 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-v9d58" podStartSLOduration=8.150093783 podStartE2EDuration="8.150093783s" podCreationTimestamp="2025-12-07 16:04:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:04:36.148636042 +0000 UTC m=+138.838920954" watchObservedRunningTime="2025-12-07 16:04:36.150093783 +0000 UTC m=+138.840378695" Dec 07 16:04:36 crc kubenswrapper[4716]: I1207 16:04:36.162214 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-d49gn" event={"ID":"8439dba6-4f32-4c02-9aad-27b9c40f7575","Type":"ContainerStarted","Data":"0d5ef69a6e678fbb5dcd566cdf11023fa2204612f02c9a78277b7b1e416b518b"} Dec 07 16:04:36 crc kubenswrapper[4716]: I1207 16:04:36.180860 4716 patch_prober.go:28] interesting pod/downloads-7954f5f757-b257c container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" start-of-body= Dec 07 16:04:36 crc kubenswrapper[4716]: I1207 16:04:36.180892 4716 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-b257c" podUID="1306f3b2-2982-494e-9db0-86c2e05b7a82" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" Dec 07 16:04:36 crc kubenswrapper[4716]: I1207 16:04:36.190243 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-pszqv" Dec 07 16:04:36 crc kubenswrapper[4716]: I1207 16:04:36.195349 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bwsl4" Dec 07 16:04:36 crc kubenswrapper[4716]: I1207 16:04:36.197865 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j26br" podStartSLOduration=116.197855919 podStartE2EDuration="1m56.197855919s" podCreationTimestamp="2025-12-07 16:02:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:04:36.19647874 +0000 UTC m=+138.886763652" watchObservedRunningTime="2025-12-07 16:04:36.197855919 +0000 UTC m=+138.888140821" Dec 07 16:04:36 crc kubenswrapper[4716]: I1207 16:04:36.224994 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 16:04:36 crc kubenswrapper[4716]: E1207 16:04:36.225907 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 16:04:36.725876337 +0000 UTC m=+139.416161249 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:04:36 crc kubenswrapper[4716]: I1207 16:04:36.265173 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bcfw4" podStartSLOduration=117.265147559 podStartE2EDuration="1m57.265147559s" podCreationTimestamp="2025-12-07 16:02:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:04:36.262493542 +0000 UTC m=+138.952778454" watchObservedRunningTime="2025-12-07 16:04:36.265147559 +0000 UTC m=+138.955432471" Dec 07 16:04:36 crc kubenswrapper[4716]: I1207 16:04:36.315732 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-d7c2c" podStartSLOduration=117.315715166 podStartE2EDuration="1m57.315715166s" podCreationTimestamp="2025-12-07 16:02:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:04:36.315696796 +0000 UTC m=+139.005981708" watchObservedRunningTime="2025-12-07 16:04:36.315715166 +0000 UTC m=+139.006000078" Dec 07 16:04:36 crc kubenswrapper[4716]: I1207 16:04:36.326059 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d257\" (UID: \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d257" Dec 07 16:04:36 crc kubenswrapper[4716]: E1207 16:04:36.328948 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 16:04:36.828897166 +0000 UTC m=+139.519182078 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d257" (UID: "75569b15-b1c1-46bf-83f2-cd6eb0a38381") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:04:36 crc kubenswrapper[4716]: I1207 16:04:36.350108 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-kkvsk" podStartSLOduration=117.350090277 podStartE2EDuration="1m57.350090277s" podCreationTimestamp="2025-12-07 16:02:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:04:36.34739088 +0000 UTC m=+139.037675792" watchObservedRunningTime="2025-12-07 16:04:36.350090277 +0000 UTC m=+139.040375189" Dec 07 16:04:36 crc kubenswrapper[4716]: I1207 16:04:36.430769 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-64rzd" podStartSLOduration=117.430752421 podStartE2EDuration="1m57.430752421s" podCreationTimestamp="2025-12-07 16:02:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:04:36.388590796 +0000 UTC m=+139.078875708" watchObservedRunningTime="2025-12-07 16:04:36.430752421 +0000 UTC m=+139.121037333" Dec 07 16:04:36 crc kubenswrapper[4716]: I1207 16:04:36.455721 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 16:04:36 crc kubenswrapper[4716]: E1207 16:04:36.455817 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 16:04:36.955796263 +0000 UTC m=+139.646081195 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:04:36 crc kubenswrapper[4716]: I1207 16:04:36.455863 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d257\" (UID: \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d257" Dec 07 16:04:36 crc kubenswrapper[4716]: E1207 16:04:36.456199 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 16:04:36.956191265 +0000 UTC m=+139.646476177 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d257" (UID: "75569b15-b1c1-46bf-83f2-cd6eb0a38381") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:04:36 crc kubenswrapper[4716]: I1207 16:04:36.509493 4716 patch_prober.go:28] interesting pod/router-default-5444994796-h5l29 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 07 16:04:36 crc kubenswrapper[4716]: [-]has-synced failed: reason withheld Dec 07 16:04:36 crc kubenswrapper[4716]: [+]process-running ok Dec 07 16:04:36 crc kubenswrapper[4716]: healthz check failed Dec 07 16:04:36 crc kubenswrapper[4716]: I1207 16:04:36.509553 4716 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h5l29" podUID="85bc0929-de17-4d2d-8b03-1a05d207fb55" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 07 16:04:36 crc kubenswrapper[4716]: I1207 16:04:36.515908 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bwsl4" podStartSLOduration=116.515884215 podStartE2EDuration="1m56.515884215s" podCreationTimestamp="2025-12-07 16:02:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:04:36.509538262 +0000 UTC m=+139.199823174" watchObservedRunningTime="2025-12-07 16:04:36.515884215 +0000 UTC m=+139.206169127" Dec 07 16:04:36 crc kubenswrapper[4716]: I1207 16:04:36.527104 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-d49gn" podStartSLOduration=117.527064238 podStartE2EDuration="1m57.527064238s" podCreationTimestamp="2025-12-07 16:02:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:04:36.526715718 +0000 UTC m=+139.217000630" watchObservedRunningTime="2025-12-07 16:04:36.527064238 +0000 UTC m=+139.217349150" Dec 07 16:04:36 crc kubenswrapper[4716]: I1207 16:04:36.551048 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-tqctv" podStartSLOduration=116.551028528 podStartE2EDuration="1m56.551028528s" podCreationTimestamp="2025-12-07 16:02:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:04:36.550448011 +0000 UTC m=+139.240732913" watchObservedRunningTime="2025-12-07 16:04:36.551028528 +0000 UTC m=+139.241313440" Dec 07 16:04:36 crc kubenswrapper[4716]: I1207 16:04:36.559428 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 16:04:36 crc kubenswrapper[4716]: E1207 16:04:36.559897 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 16:04:37.059881443 +0000 UTC m=+139.750166355 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:04:36 crc kubenswrapper[4716]: I1207 16:04:36.661306 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d257\" (UID: \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d257" Dec 07 16:04:36 crc kubenswrapper[4716]: E1207 16:04:36.661823 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 16:04:37.161807631 +0000 UTC m=+139.852092533 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d257" (UID: "75569b15-b1c1-46bf-83f2-cd6eb0a38381") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:04:36 crc kubenswrapper[4716]: I1207 16:04:36.762472 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 16:04:36 crc kubenswrapper[4716]: E1207 16:04:36.762630 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 16:04:37.262610456 +0000 UTC m=+139.952895368 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:04:36 crc kubenswrapper[4716]: I1207 16:04:36.762746 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d257\" (UID: \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d257" Dec 07 16:04:36 crc kubenswrapper[4716]: E1207 16:04:36.763092 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 16:04:37.263070599 +0000 UTC m=+139.953355511 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d257" (UID: "75569b15-b1c1-46bf-83f2-cd6eb0a38381") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:04:36 crc kubenswrapper[4716]: I1207 16:04:36.864146 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 16:04:36 crc kubenswrapper[4716]: E1207 16:04:36.864295 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 16:04:37.364278626 +0000 UTC m=+140.054563528 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:04:36 crc kubenswrapper[4716]: I1207 16:04:36.864742 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d257\" (UID: \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d257" Dec 07 16:04:36 crc kubenswrapper[4716]: E1207 16:04:36.865060 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 16:04:37.365052598 +0000 UTC m=+140.055337510 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d257" (UID: "75569b15-b1c1-46bf-83f2-cd6eb0a38381") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:04:36 crc kubenswrapper[4716]: I1207 16:04:36.966122 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 16:04:36 crc kubenswrapper[4716]: E1207 16:04:36.966250 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 16:04:37.466235264 +0000 UTC m=+140.156520166 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:04:36 crc kubenswrapper[4716]: I1207 16:04:36.966602 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d257\" (UID: \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d257" Dec 07 16:04:36 crc kubenswrapper[4716]: E1207 16:04:36.966934 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 16:04:37.466922054 +0000 UTC m=+140.157206966 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d257" (UID: "75569b15-b1c1-46bf-83f2-cd6eb0a38381") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.067752 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 16:04:37 crc kubenswrapper[4716]: E1207 16:04:37.068142 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 16:04:37.568116531 +0000 UTC m=+140.258401443 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.122910 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-x2nd8" Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.169708 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d257\" (UID: \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d257" Dec 07 16:04:37 crc kubenswrapper[4716]: E1207 16:04:37.170097 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 16:04:37.670071079 +0000 UTC m=+140.360355991 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d257" (UID: "75569b15-b1c1-46bf-83f2-cd6eb0a38381") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.170616 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-sq5lw" event={"ID":"d27e02ff-86da-4078-934b-fd472b471f69","Type":"ContainerStarted","Data":"b1767f9e51c7bd963dd8ec8a1785dd8ad2bd00997bb2c91df555e7df8c92635f"} Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.172418 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-v9d58" event={"ID":"5d2c595e-c4a6-446d-836a-766d1ec44516","Type":"ContainerStarted","Data":"5af2f646b5c672fd77a22885e6a391e97db8a9b2ada347d5331874c389f7cbf6"} Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.175168 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rb966" event={"ID":"060bb436-de36-408a-b742-f3768a575801","Type":"ContainerStarted","Data":"66933e918e00cc5c4bc2d1dc0a3b2bb85f40bf0b944bd233c6633e15dd0db090"} Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.177595 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-n4492" event={"ID":"db753ef2-2939-42dd-886d-17a223c47360","Type":"ContainerStarted","Data":"e8962f558dcf1b30f3489ec6095ef50d0ca9832f4f373300453152dd70acef20"} Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.177626 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-n4492" event={"ID":"db753ef2-2939-42dd-886d-17a223c47360","Type":"ContainerStarted","Data":"90ca326eb3b219528d2cff40f51ba149d9c600f21e7c129390eb08e0ef48a35a"} Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.177843 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-n4492" Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.183353 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29418720-x5rqz" event={"ID":"98efd33b-de1f-4414-b9cc-55fb0d423784","Type":"ContainerStarted","Data":"efc3cdf3d0d1568d81c77d3121f4fc0f37b2c7bfba87382c2b7631532f2d134b"} Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.188360 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-h6gnn" event={"ID":"74c41933-56ed-4f49-a808-1d7f1f1829b2","Type":"ContainerStarted","Data":"a0c2a90668d354d1330651c7025f41950e5e4cb88a1b5caf91afe1aa8786dcea"} Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.189052 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-h6gnn" event={"ID":"74c41933-56ed-4f49-a808-1d7f1f1829b2","Type":"ContainerStarted","Data":"0642fd23bc35952d8cc02f992ad2fcedbb9a9c47b5acca307e3db5ece5f078ce"} Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.219896 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-89v7m" event={"ID":"6b4e49ef-27ab-417c-9c6b-1ad7f215604c","Type":"ContainerStarted","Data":"342e51240282a5bf7e6a8eec863906ee13db2d35c760f03711da3dbf1a80fea0"} Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.220290 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-89v7m" Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.235401 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-l9dgz" event={"ID":"55935fe0-005a-4f9f-80cd-9fa3348cc90a","Type":"ContainerStarted","Data":"c0145f1b5b72f9a9b348fb152ba5b1853d10a9b2cd79570da29cc0d77f5911de"} Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.242259 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-sq5lw" podStartSLOduration=117.242242229 podStartE2EDuration="1m57.242242229s" podCreationTimestamp="2025-12-07 16:02:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:04:37.210062071 +0000 UTC m=+139.900346983" watchObservedRunningTime="2025-12-07 16:04:37.242242229 +0000 UTC m=+139.932527141" Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.243850 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-n4492" podStartSLOduration=9.243846316 podStartE2EDuration="9.243846316s" podCreationTimestamp="2025-12-07 16:04:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:04:37.242072134 +0000 UTC m=+139.932357046" watchObservedRunningTime="2025-12-07 16:04:37.243846316 +0000 UTC m=+139.934131228" Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.261588 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-zp6l4" event={"ID":"fc892d74-c6d2-431f-a52c-95a8f4e2d03c","Type":"ContainerStarted","Data":"a7a6691848c46803004346e63af5a312f3ccddc4310a6dfc360a9379c80aee96"} Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.272632 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.275113 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rb966" podStartSLOduration=118.275062285 podStartE2EDuration="1m58.275062285s" podCreationTimestamp="2025-12-07 16:02:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:04:37.273269413 +0000 UTC m=+139.963554325" watchObservedRunningTime="2025-12-07 16:04:37.275062285 +0000 UTC m=+139.965347197" Dec 07 16:04:37 crc kubenswrapper[4716]: E1207 16:04:37.275603 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 16:04:37.77557403 +0000 UTC m=+140.465858932 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.289940 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-w7xhs" event={"ID":"38ef2332-5fee-44a2-b56c-bc6db5fad2e0","Type":"ContainerStarted","Data":"d69cb005631a6ab0a57d79713760d97d43e56477d5ecce0dbd67356effb42388"} Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.289993 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-w7xhs" event={"ID":"38ef2332-5fee-44a2-b56c-bc6db5fad2e0","Type":"ContainerStarted","Data":"087dcc37530410874b9d8816675dd11a7cc88ad49b93e9296d42881f3cbebbc2"} Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.315427 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-h6gnn" podStartSLOduration=118.315405237 podStartE2EDuration="1m58.315405237s" podCreationTimestamp="2025-12-07 16:02:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:04:37.312699569 +0000 UTC m=+140.002984481" watchObservedRunningTime="2025-12-07 16:04:37.315405237 +0000 UTC m=+140.005690139" Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.335143 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-gc4cp" event={"ID":"6f7305ef-d76f-4be3-9151-7247f7590fe0","Type":"ContainerStarted","Data":"9a0f991f2dc9ab8506592dc4c075a778d3fbd62c03e1d1b59b25ab0176429d95"} Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.335187 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-gc4cp" event={"ID":"6f7305ef-d76f-4be3-9151-7247f7590fe0","Type":"ContainerStarted","Data":"ef0e4f4c6ad47c2bba3ca3e0fd13735395e9f45a784cd4077d6ca539327d9506"} Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.347499 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-zp6l4" podStartSLOduration=118.347479582 podStartE2EDuration="1m58.347479582s" podCreationTimestamp="2025-12-07 16:02:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:04:37.344727013 +0000 UTC m=+140.035011915" watchObservedRunningTime="2025-12-07 16:04:37.347479582 +0000 UTC m=+140.037764484" Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.376025 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d257\" (UID: \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d257" Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.378743 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-w7xhs" podStartSLOduration=118.378732893 podStartE2EDuration="1m58.378732893s" podCreationTimestamp="2025-12-07 16:02:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:04:37.376671463 +0000 UTC m=+140.066956375" watchObservedRunningTime="2025-12-07 16:04:37.378732893 +0000 UTC m=+140.069017795" Dec 07 16:04:37 crc kubenswrapper[4716]: E1207 16:04:37.379898 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 16:04:37.879880936 +0000 UTC m=+140.570165848 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d257" (UID: "75569b15-b1c1-46bf-83f2-cd6eb0a38381") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.391034 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-82bvt" event={"ID":"af51b694-cd0f-4c33-ba6e-1f4666178820","Type":"ContainerStarted","Data":"32b3c0fc8bd8905012501b42c4a7049af9dd4fc4974f6bfe5933c2b1ed580746"} Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.391092 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-82bvt" event={"ID":"af51b694-cd0f-4c33-ba6e-1f4666178820","Type":"ContainerStarted","Data":"293ea8bda76a24e44032d365df6227a5aa752575883ac302941f43319293cde9"} Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.393170 4716 patch_prober.go:28] interesting pod/downloads-7954f5f757-b257c container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" start-of-body= Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.393238 4716 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-b257c" podUID="1306f3b2-2982-494e-9db0-86c2e05b7a82" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.402667 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-v6899" Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.411474 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-j26br" Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.421387 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-89v7m" podStartSLOduration=117.421371331 podStartE2EDuration="1m57.421371331s" podCreationTimestamp="2025-12-07 16:02:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:04:37.420501037 +0000 UTC m=+140.110785949" watchObservedRunningTime="2025-12-07 16:04:37.421371331 +0000 UTC m=+140.111656243" Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.479627 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 16:04:37 crc kubenswrapper[4716]: E1207 16:04:37.481645 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 16:04:37.981630358 +0000 UTC m=+140.671915270 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.492287 4716 patch_prober.go:28] interesting pod/router-default-5444994796-h5l29 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 07 16:04:37 crc kubenswrapper[4716]: [-]has-synced failed: reason withheld Dec 07 16:04:37 crc kubenswrapper[4716]: [+]process-running ok Dec 07 16:04:37 crc kubenswrapper[4716]: healthz check failed Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.492341 4716 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h5l29" podUID="85bc0929-de17-4d2d-8b03-1a05d207fb55" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.503262 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-cd798"] Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.504344 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cd798" Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.507969 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.531047 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cd798"] Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.551758 4716 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.569959 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-gc4cp" podStartSLOduration=118.569933163 podStartE2EDuration="1m58.569933163s" podCreationTimestamp="2025-12-07 16:02:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:04:37.560313015 +0000 UTC m=+140.250597947" watchObservedRunningTime="2025-12-07 16:04:37.569933163 +0000 UTC m=+140.260218075" Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.581739 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45c781c9-df95-4b6e-bed3-fc1f22f079a3-utilities\") pod \"certified-operators-cd798\" (UID: \"45c781c9-df95-4b6e-bed3-fc1f22f079a3\") " pod="openshift-marketplace/certified-operators-cd798" Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.581807 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d257\" (UID: \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d257" Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.581921 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45c781c9-df95-4b6e-bed3-fc1f22f079a3-catalog-content\") pod \"certified-operators-cd798\" (UID: \"45c781c9-df95-4b6e-bed3-fc1f22f079a3\") " pod="openshift-marketplace/certified-operators-cd798" Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.582005 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wn54r\" (UniqueName: \"kubernetes.io/projected/45c781c9-df95-4b6e-bed3-fc1f22f079a3-kube-api-access-wn54r\") pod \"certified-operators-cd798\" (UID: \"45c781c9-df95-4b6e-bed3-fc1f22f079a3\") " pod="openshift-marketplace/certified-operators-cd798" Dec 07 16:04:37 crc kubenswrapper[4716]: E1207 16:04:37.582414 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 16:04:38.082394142 +0000 UTC m=+140.772679064 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d257" (UID: "75569b15-b1c1-46bf-83f2-cd6eb0a38381") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.686878 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.686900 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-d92nn"] Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.687124 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45c781c9-df95-4b6e-bed3-fc1f22f079a3-utilities\") pod \"certified-operators-cd798\" (UID: \"45c781c9-df95-4b6e-bed3-fc1f22f079a3\") " pod="openshift-marketplace/certified-operators-cd798" Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.687213 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45c781c9-df95-4b6e-bed3-fc1f22f079a3-catalog-content\") pod \"certified-operators-cd798\" (UID: \"45c781c9-df95-4b6e-bed3-fc1f22f079a3\") " pod="openshift-marketplace/certified-operators-cd798" Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.687264 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wn54r\" (UniqueName: \"kubernetes.io/projected/45c781c9-df95-4b6e-bed3-fc1f22f079a3-kube-api-access-wn54r\") pod \"certified-operators-cd798\" (UID: \"45c781c9-df95-4b6e-bed3-fc1f22f079a3\") " pod="openshift-marketplace/certified-operators-cd798" Dec 07 16:04:37 crc kubenswrapper[4716]: E1207 16:04:37.687619 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-07 16:04:38.187603464 +0000 UTC m=+140.877888376 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.687981 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45c781c9-df95-4b6e-bed3-fc1f22f079a3-utilities\") pod \"certified-operators-cd798\" (UID: \"45c781c9-df95-4b6e-bed3-fc1f22f079a3\") " pod="openshift-marketplace/certified-operators-cd798" Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.688217 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45c781c9-df95-4b6e-bed3-fc1f22f079a3-catalog-content\") pod \"certified-operators-cd798\" (UID: \"45c781c9-df95-4b6e-bed3-fc1f22f079a3\") " pod="openshift-marketplace/certified-operators-cd798" Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.688246 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d92nn" Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.692242 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.715328 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wn54r\" (UniqueName: \"kubernetes.io/projected/45c781c9-df95-4b6e-bed3-fc1f22f079a3-kube-api-access-wn54r\") pod \"certified-operators-cd798\" (UID: \"45c781c9-df95-4b6e-bed3-fc1f22f079a3\") " pod="openshift-marketplace/certified-operators-cd798" Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.716561 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-d92nn"] Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.787857 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/074b6ff7-7e20-4054-811d-34050c07e74e-catalog-content\") pod \"community-operators-d92nn\" (UID: \"074b6ff7-7e20-4054-811d-34050c07e74e\") " pod="openshift-marketplace/community-operators-d92nn" Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.787908 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d257\" (UID: \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d257" Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.788115 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gvh45\" (UniqueName: \"kubernetes.io/projected/074b6ff7-7e20-4054-811d-34050c07e74e-kube-api-access-gvh45\") pod \"community-operators-d92nn\" (UID: \"074b6ff7-7e20-4054-811d-34050c07e74e\") " pod="openshift-marketplace/community-operators-d92nn" Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.788144 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/074b6ff7-7e20-4054-811d-34050c07e74e-utilities\") pod \"community-operators-d92nn\" (UID: \"074b6ff7-7e20-4054-811d-34050c07e74e\") " pod="openshift-marketplace/community-operators-d92nn" Dec 07 16:04:37 crc kubenswrapper[4716]: E1207 16:04:37.788224 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-07 16:04:38.288212894 +0000 UTC m=+140.978497806 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4d257" (UID: "75569b15-b1c1-46bf-83f2-cd6eb0a38381") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.816931 4716 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-12-07T16:04:37.55178778Z","Handler":null,"Name":""} Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.820295 4716 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.820339 4716 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.842097 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cd798" Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.866035 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-chrks"] Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.867586 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-chrks" Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.889658 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.889871 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9tbdw\" (UniqueName: \"kubernetes.io/projected/de1318df-206b-4ad8-98c7-b9c943560e44-kube-api-access-9tbdw\") pod \"certified-operators-chrks\" (UID: \"de1318df-206b-4ad8-98c7-b9c943560e44\") " pod="openshift-marketplace/certified-operators-chrks" Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.889972 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gvh45\" (UniqueName: \"kubernetes.io/projected/074b6ff7-7e20-4054-811d-34050c07e74e-kube-api-access-gvh45\") pod \"community-operators-d92nn\" (UID: \"074b6ff7-7e20-4054-811d-34050c07e74e\") " pod="openshift-marketplace/community-operators-d92nn" Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.890001 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/074b6ff7-7e20-4054-811d-34050c07e74e-utilities\") pod \"community-operators-d92nn\" (UID: \"074b6ff7-7e20-4054-811d-34050c07e74e\") " pod="openshift-marketplace/community-operators-d92nn" Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.890021 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de1318df-206b-4ad8-98c7-b9c943560e44-utilities\") pod \"certified-operators-chrks\" (UID: \"de1318df-206b-4ad8-98c7-b9c943560e44\") " pod="openshift-marketplace/certified-operators-chrks" Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.890045 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/074b6ff7-7e20-4054-811d-34050c07e74e-catalog-content\") pod \"community-operators-d92nn\" (UID: \"074b6ff7-7e20-4054-811d-34050c07e74e\") " pod="openshift-marketplace/community-operators-d92nn" Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.890088 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de1318df-206b-4ad8-98c7-b9c943560e44-catalog-content\") pod \"certified-operators-chrks\" (UID: \"de1318df-206b-4ad8-98c7-b9c943560e44\") " pod="openshift-marketplace/certified-operators-chrks" Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.890995 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/074b6ff7-7e20-4054-811d-34050c07e74e-utilities\") pod \"community-operators-d92nn\" (UID: \"074b6ff7-7e20-4054-811d-34050c07e74e\") " pod="openshift-marketplace/community-operators-d92nn" Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.891310 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/074b6ff7-7e20-4054-811d-34050c07e74e-catalog-content\") pod \"community-operators-d92nn\" (UID: \"074b6ff7-7e20-4054-811d-34050c07e74e\") " pod="openshift-marketplace/community-operators-d92nn" Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.903766 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-chrks"] Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.914186 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.931312 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gvh45\" (UniqueName: \"kubernetes.io/projected/074b6ff7-7e20-4054-811d-34050c07e74e-kube-api-access-gvh45\") pod \"community-operators-d92nn\" (UID: \"074b6ff7-7e20-4054-811d-34050c07e74e\") " pod="openshift-marketplace/community-operators-d92nn" Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.990875 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de1318df-206b-4ad8-98c7-b9c943560e44-utilities\") pod \"certified-operators-chrks\" (UID: \"de1318df-206b-4ad8-98c7-b9c943560e44\") " pod="openshift-marketplace/certified-operators-chrks" Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.990945 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d257\" (UID: \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d257" Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.990965 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de1318df-206b-4ad8-98c7-b9c943560e44-catalog-content\") pod \"certified-operators-chrks\" (UID: \"de1318df-206b-4ad8-98c7-b9c943560e44\") " pod="openshift-marketplace/certified-operators-chrks" Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.990998 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9tbdw\" (UniqueName: \"kubernetes.io/projected/de1318df-206b-4ad8-98c7-b9c943560e44-kube-api-access-9tbdw\") pod \"certified-operators-chrks\" (UID: \"de1318df-206b-4ad8-98c7-b9c943560e44\") " pod="openshift-marketplace/certified-operators-chrks" Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.991834 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de1318df-206b-4ad8-98c7-b9c943560e44-utilities\") pod \"certified-operators-chrks\" (UID: \"de1318df-206b-4ad8-98c7-b9c943560e44\") " pod="openshift-marketplace/certified-operators-chrks" Dec 07 16:04:37 crc kubenswrapper[4716]: I1207 16:04:37.992623 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de1318df-206b-4ad8-98c7-b9c943560e44-catalog-content\") pod \"certified-operators-chrks\" (UID: \"de1318df-206b-4ad8-98c7-b9c943560e44\") " pod="openshift-marketplace/certified-operators-chrks" Dec 07 16:04:38 crc kubenswrapper[4716]: I1207 16:04:38.000790 4716 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 07 16:04:38 crc kubenswrapper[4716]: I1207 16:04:38.000851 4716 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d257\" (UID: \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-4d257" Dec 07 16:04:38 crc kubenswrapper[4716]: I1207 16:04:38.009890 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9tbdw\" (UniqueName: \"kubernetes.io/projected/de1318df-206b-4ad8-98c7-b9c943560e44-kube-api-access-9tbdw\") pod \"certified-operators-chrks\" (UID: \"de1318df-206b-4ad8-98c7-b9c943560e44\") " pod="openshift-marketplace/certified-operators-chrks" Dec 07 16:04:38 crc kubenswrapper[4716]: I1207 16:04:38.013126 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d92nn" Dec 07 16:04:38 crc kubenswrapper[4716]: I1207 16:04:38.074921 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-6wb5h"] Dec 07 16:04:38 crc kubenswrapper[4716]: I1207 16:04:38.076726 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6wb5h" Dec 07 16:04:38 crc kubenswrapper[4716]: I1207 16:04:38.092406 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96a76712-c754-4978-902a-fa07f6c3c634-utilities\") pod \"community-operators-6wb5h\" (UID: \"96a76712-c754-4978-902a-fa07f6c3c634\") " pod="openshift-marketplace/community-operators-6wb5h" Dec 07 16:04:38 crc kubenswrapper[4716]: I1207 16:04:38.092554 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9lbc5\" (UniqueName: \"kubernetes.io/projected/96a76712-c754-4978-902a-fa07f6c3c634-kube-api-access-9lbc5\") pod \"community-operators-6wb5h\" (UID: \"96a76712-c754-4978-902a-fa07f6c3c634\") " pod="openshift-marketplace/community-operators-6wb5h" Dec 07 16:04:38 crc kubenswrapper[4716]: I1207 16:04:38.092601 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96a76712-c754-4978-902a-fa07f6c3c634-catalog-content\") pod \"community-operators-6wb5h\" (UID: \"96a76712-c754-4978-902a-fa07f6c3c634\") " pod="openshift-marketplace/community-operators-6wb5h" Dec 07 16:04:38 crc kubenswrapper[4716]: I1207 16:04:38.092709 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6wb5h"] Dec 07 16:04:38 crc kubenswrapper[4716]: I1207 16:04:38.194228 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96a76712-c754-4978-902a-fa07f6c3c634-utilities\") pod \"community-operators-6wb5h\" (UID: \"96a76712-c754-4978-902a-fa07f6c3c634\") " pod="openshift-marketplace/community-operators-6wb5h" Dec 07 16:04:38 crc kubenswrapper[4716]: I1207 16:04:38.194346 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9lbc5\" (UniqueName: \"kubernetes.io/projected/96a76712-c754-4978-902a-fa07f6c3c634-kube-api-access-9lbc5\") pod \"community-operators-6wb5h\" (UID: \"96a76712-c754-4978-902a-fa07f6c3c634\") " pod="openshift-marketplace/community-operators-6wb5h" Dec 07 16:04:38 crc kubenswrapper[4716]: I1207 16:04:38.194376 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96a76712-c754-4978-902a-fa07f6c3c634-catalog-content\") pod \"community-operators-6wb5h\" (UID: \"96a76712-c754-4978-902a-fa07f6c3c634\") " pod="openshift-marketplace/community-operators-6wb5h" Dec 07 16:04:38 crc kubenswrapper[4716]: I1207 16:04:38.194946 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96a76712-c754-4978-902a-fa07f6c3c634-catalog-content\") pod \"community-operators-6wb5h\" (UID: \"96a76712-c754-4978-902a-fa07f6c3c634\") " pod="openshift-marketplace/community-operators-6wb5h" Dec 07 16:04:38 crc kubenswrapper[4716]: I1207 16:04:38.195185 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96a76712-c754-4978-902a-fa07f6c3c634-utilities\") pod \"community-operators-6wb5h\" (UID: \"96a76712-c754-4978-902a-fa07f6c3c634\") " pod="openshift-marketplace/community-operators-6wb5h" Dec 07 16:04:38 crc kubenswrapper[4716]: I1207 16:04:38.206740 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-chrks" Dec 07 16:04:38 crc kubenswrapper[4716]: I1207 16:04:38.221156 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9lbc5\" (UniqueName: \"kubernetes.io/projected/96a76712-c754-4978-902a-fa07f6c3c634-kube-api-access-9lbc5\") pod \"community-operators-6wb5h\" (UID: \"96a76712-c754-4978-902a-fa07f6c3c634\") " pod="openshift-marketplace/community-operators-6wb5h" Dec 07 16:04:38 crc kubenswrapper[4716]: I1207 16:04:38.236989 4716 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-l9dgz container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.38:5443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 07 16:04:38 crc kubenswrapper[4716]: I1207 16:04:38.237050 4716 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-l9dgz" podUID="55935fe0-005a-4f9f-80cd-9fa3348cc90a" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.38:5443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 07 16:04:38 crc kubenswrapper[4716]: I1207 16:04:38.247911 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cd798"] Dec 07 16:04:38 crc kubenswrapper[4716]: I1207 16:04:38.306138 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-d92nn"] Dec 07 16:04:38 crc kubenswrapper[4716]: I1207 16:04:38.351694 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4d257\" (UID: \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\") " pod="openshift-image-registry/image-registry-697d97f7c8-4d257" Dec 07 16:04:38 crc kubenswrapper[4716]: I1207 16:04:38.400446 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6wb5h" Dec 07 16:04:38 crc kubenswrapper[4716]: I1207 16:04:38.420366 4716 generic.go:334] "Generic (PLEG): container finished" podID="98efd33b-de1f-4414-b9cc-55fb0d423784" containerID="efc3cdf3d0d1568d81c77d3121f4fc0f37b2c7bfba87382c2b7631532f2d134b" exitCode=0 Dec 07 16:04:38 crc kubenswrapper[4716]: I1207 16:04:38.420428 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29418720-x5rqz" event={"ID":"98efd33b-de1f-4414-b9cc-55fb0d423784","Type":"ContainerDied","Data":"efc3cdf3d0d1568d81c77d3121f4fc0f37b2c7bfba87382c2b7631532f2d134b"} Dec 07 16:04:38 crc kubenswrapper[4716]: I1207 16:04:38.428254 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-4d257" Dec 07 16:04:38 crc kubenswrapper[4716]: I1207 16:04:38.445714 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-82bvt" event={"ID":"af51b694-cd0f-4c33-ba6e-1f4666178820","Type":"ContainerStarted","Data":"e6b5eccee7f1af8a57ba74ebc1094e082d749fb30ae944f52ac93462275f3672"} Dec 07 16:04:38 crc kubenswrapper[4716]: I1207 16:04:38.449057 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cd798" event={"ID":"45c781c9-df95-4b6e-bed3-fc1f22f079a3","Type":"ContainerStarted","Data":"50fe4fc8b66ae902be0b4d4206384da12fa977147ca08eb352aec9efbb98cdbb"} Dec 07 16:04:38 crc kubenswrapper[4716]: I1207 16:04:38.473860 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d92nn" event={"ID":"074b6ff7-7e20-4054-811d-34050c07e74e","Type":"ContainerStarted","Data":"b4c6a8dad58efff23ae8852394652eb42a704a02d7de36cb77fc9ab189dd5221"} Dec 07 16:04:38 crc kubenswrapper[4716]: I1207 16:04:38.482885 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-l9dgz" Dec 07 16:04:38 crc kubenswrapper[4716]: I1207 16:04:38.499565 4716 patch_prober.go:28] interesting pod/router-default-5444994796-h5l29 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 07 16:04:38 crc kubenswrapper[4716]: [-]has-synced failed: reason withheld Dec 07 16:04:38 crc kubenswrapper[4716]: [+]process-running ok Dec 07 16:04:38 crc kubenswrapper[4716]: healthz check failed Dec 07 16:04:38 crc kubenswrapper[4716]: I1207 16:04:38.499633 4716 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h5l29" podUID="85bc0929-de17-4d2d-8b03-1a05d207fb55" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 07 16:04:38 crc kubenswrapper[4716]: I1207 16:04:38.587500 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-chrks"] Dec 07 16:04:38 crc kubenswrapper[4716]: I1207 16:04:38.991735 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-4d257"] Dec 07 16:04:39 crc kubenswrapper[4716]: I1207 16:04:39.034901 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6wb5h"] Dec 07 16:04:39 crc kubenswrapper[4716]: W1207 16:04:39.052762 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod96a76712_c754_4978_902a_fa07f6c3c634.slice/crio-5d68160f8fb287da1d047a59a81beaf4513bf3a9451850aada38fb9c5fc6d6d2 WatchSource:0}: Error finding container 5d68160f8fb287da1d047a59a81beaf4513bf3a9451850aada38fb9c5fc6d6d2: Status 404 returned error can't find the container with id 5d68160f8fb287da1d047a59a81beaf4513bf3a9451850aada38fb9c5fc6d6d2 Dec 07 16:04:39 crc kubenswrapper[4716]: W1207 16:04:39.053535 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod75569b15_b1c1_46bf_83f2_cd6eb0a38381.slice/crio-84f55b0d6385b8e76bcf9155671e73688b1d730161e0cace0ed36c4b4b026162 WatchSource:0}: Error finding container 84f55b0d6385b8e76bcf9155671e73688b1d730161e0cace0ed36c4b4b026162: Status 404 returned error can't find the container with id 84f55b0d6385b8e76bcf9155671e73688b1d730161e0cace0ed36c4b4b026162 Dec 07 16:04:39 crc kubenswrapper[4716]: I1207 16:04:39.274882 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 07 16:04:39 crc kubenswrapper[4716]: I1207 16:04:39.275515 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 07 16:04:39 crc kubenswrapper[4716]: I1207 16:04:39.277912 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Dec 07 16:04:39 crc kubenswrapper[4716]: I1207 16:04:39.278281 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Dec 07 16:04:39 crc kubenswrapper[4716]: I1207 16:04:39.284577 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 07 16:04:39 crc kubenswrapper[4716]: I1207 16:04:39.364155 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/15c001b6-abae-4fb3-ac0e-c23c2e006ff1-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"15c001b6-abae-4fb3-ac0e-c23c2e006ff1\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 07 16:04:39 crc kubenswrapper[4716]: I1207 16:04:39.364370 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/15c001b6-abae-4fb3-ac0e-c23c2e006ff1-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"15c001b6-abae-4fb3-ac0e-c23c2e006ff1\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 07 16:04:39 crc kubenswrapper[4716]: I1207 16:04:39.462517 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-fnkww"] Dec 07 16:04:39 crc kubenswrapper[4716]: I1207 16:04:39.463529 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fnkww" Dec 07 16:04:39 crc kubenswrapper[4716]: I1207 16:04:39.465575 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/15c001b6-abae-4fb3-ac0e-c23c2e006ff1-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"15c001b6-abae-4fb3-ac0e-c23c2e006ff1\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 07 16:04:39 crc kubenswrapper[4716]: I1207 16:04:39.465720 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/15c001b6-abae-4fb3-ac0e-c23c2e006ff1-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"15c001b6-abae-4fb3-ac0e-c23c2e006ff1\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 07 16:04:39 crc kubenswrapper[4716]: I1207 16:04:39.465716 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/15c001b6-abae-4fb3-ac0e-c23c2e006ff1-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"15c001b6-abae-4fb3-ac0e-c23c2e006ff1\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 07 16:04:39 crc kubenswrapper[4716]: I1207 16:04:39.465869 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 07 16:04:39 crc kubenswrapper[4716]: I1207 16:04:39.475604 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fnkww"] Dec 07 16:04:39 crc kubenswrapper[4716]: I1207 16:04:39.483994 4716 generic.go:334] "Generic (PLEG): container finished" podID="de1318df-206b-4ad8-98c7-b9c943560e44" containerID="1fe9014ba1b4e5f8f3b0c1d316017607984f88799209849528531f460b024af7" exitCode=0 Dec 07 16:04:39 crc kubenswrapper[4716]: I1207 16:04:39.484074 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-chrks" event={"ID":"de1318df-206b-4ad8-98c7-b9c943560e44","Type":"ContainerDied","Data":"1fe9014ba1b4e5f8f3b0c1d316017607984f88799209849528531f460b024af7"} Dec 07 16:04:39 crc kubenswrapper[4716]: I1207 16:04:39.484123 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-chrks" event={"ID":"de1318df-206b-4ad8-98c7-b9c943560e44","Type":"ContainerStarted","Data":"34c87a6f423d4af364ae035ff900e68134f28aa0fac938106f075adac4214216"} Dec 07 16:04:39 crc kubenswrapper[4716]: I1207 16:04:39.485847 4716 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 07 16:04:39 crc kubenswrapper[4716]: I1207 16:04:39.488645 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-82bvt" event={"ID":"af51b694-cd0f-4c33-ba6e-1f4666178820","Type":"ContainerStarted","Data":"56c10f8966f220a0b64c2068be94d2756490c8e1a37ec40d2bea256f7170371c"} Dec 07 16:04:39 crc kubenswrapper[4716]: I1207 16:04:39.496281 4716 patch_prober.go:28] interesting pod/router-default-5444994796-h5l29 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 07 16:04:39 crc kubenswrapper[4716]: [-]has-synced failed: reason withheld Dec 07 16:04:39 crc kubenswrapper[4716]: [+]process-running ok Dec 07 16:04:39 crc kubenswrapper[4716]: healthz check failed Dec 07 16:04:39 crc kubenswrapper[4716]: I1207 16:04:39.496337 4716 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h5l29" podUID="85bc0929-de17-4d2d-8b03-1a05d207fb55" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 07 16:04:39 crc kubenswrapper[4716]: I1207 16:04:39.496717 4716 generic.go:334] "Generic (PLEG): container finished" podID="45c781c9-df95-4b6e-bed3-fc1f22f079a3" containerID="1063c57a5e4100898b9507ac1f578155291d67907813fbb7c8def6e42f86febc" exitCode=0 Dec 07 16:04:39 crc kubenswrapper[4716]: I1207 16:04:39.496795 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cd798" event={"ID":"45c781c9-df95-4b6e-bed3-fc1f22f079a3","Type":"ContainerDied","Data":"1063c57a5e4100898b9507ac1f578155291d67907813fbb7c8def6e42f86febc"} Dec 07 16:04:39 crc kubenswrapper[4716]: I1207 16:04:39.500234 4716 generic.go:334] "Generic (PLEG): container finished" podID="96a76712-c754-4978-902a-fa07f6c3c634" containerID="d772d9a9b3f4a67928c244a2bc14bd2164e8db4fe0f4365713849cebdfa09b53" exitCode=0 Dec 07 16:04:39 crc kubenswrapper[4716]: I1207 16:04:39.500303 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6wb5h" event={"ID":"96a76712-c754-4978-902a-fa07f6c3c634","Type":"ContainerDied","Data":"d772d9a9b3f4a67928c244a2bc14bd2164e8db4fe0f4365713849cebdfa09b53"} Dec 07 16:04:39 crc kubenswrapper[4716]: I1207 16:04:39.500328 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6wb5h" event={"ID":"96a76712-c754-4978-902a-fa07f6c3c634","Type":"ContainerStarted","Data":"5d68160f8fb287da1d047a59a81beaf4513bf3a9451850aada38fb9c5fc6d6d2"} Dec 07 16:04:39 crc kubenswrapper[4716]: I1207 16:04:39.501416 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-4d257" event={"ID":"75569b15-b1c1-46bf-83f2-cd6eb0a38381","Type":"ContainerStarted","Data":"aba2079f6d3cf016e5420b1261ba461a13577e9a5f66685aeb124feec845dad6"} Dec 07 16:04:39 crc kubenswrapper[4716]: I1207 16:04:39.501444 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-4d257" event={"ID":"75569b15-b1c1-46bf-83f2-cd6eb0a38381","Type":"ContainerStarted","Data":"84f55b0d6385b8e76bcf9155671e73688b1d730161e0cace0ed36c4b4b026162"} Dec 07 16:04:39 crc kubenswrapper[4716]: I1207 16:04:39.501650 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-4d257" Dec 07 16:04:39 crc kubenswrapper[4716]: I1207 16:04:39.508015 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/15c001b6-abae-4fb3-ac0e-c23c2e006ff1-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"15c001b6-abae-4fb3-ac0e-c23c2e006ff1\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 07 16:04:39 crc kubenswrapper[4716]: I1207 16:04:39.512067 4716 generic.go:334] "Generic (PLEG): container finished" podID="074b6ff7-7e20-4054-811d-34050c07e74e" containerID="fdb1818e7e1be2170714671f582dd8cfac64141be6c0bc6c67d635a72e4e3aff" exitCode=0 Dec 07 16:04:39 crc kubenswrapper[4716]: I1207 16:04:39.512192 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d92nn" event={"ID":"074b6ff7-7e20-4054-811d-34050c07e74e","Type":"ContainerDied","Data":"fdb1818e7e1be2170714671f582dd8cfac64141be6c0bc6c67d635a72e4e3aff"} Dec 07 16:04:39 crc kubenswrapper[4716]: I1207 16:04:39.566902 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/12e72f4d-2887-4c37-bf16-403f124d6f68-utilities\") pod \"redhat-marketplace-fnkww\" (UID: \"12e72f4d-2887-4c37-bf16-403f124d6f68\") " pod="openshift-marketplace/redhat-marketplace-fnkww" Dec 07 16:04:39 crc kubenswrapper[4716]: I1207 16:04:39.567054 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/12e72f4d-2887-4c37-bf16-403f124d6f68-catalog-content\") pod \"redhat-marketplace-fnkww\" (UID: \"12e72f4d-2887-4c37-bf16-403f124d6f68\") " pod="openshift-marketplace/redhat-marketplace-fnkww" Dec 07 16:04:39 crc kubenswrapper[4716]: I1207 16:04:39.567086 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fx9hn\" (UniqueName: \"kubernetes.io/projected/12e72f4d-2887-4c37-bf16-403f124d6f68-kube-api-access-fx9hn\") pod \"redhat-marketplace-fnkww\" (UID: \"12e72f4d-2887-4c37-bf16-403f124d6f68\") " pod="openshift-marketplace/redhat-marketplace-fnkww" Dec 07 16:04:39 crc kubenswrapper[4716]: I1207 16:04:39.584510 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-4d257" podStartSLOduration=120.584491503 podStartE2EDuration="2m0.584491503s" podCreationTimestamp="2025-12-07 16:02:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:04:39.582740162 +0000 UTC m=+142.273025074" watchObservedRunningTime="2025-12-07 16:04:39.584491503 +0000 UTC m=+142.274776415" Dec 07 16:04:39 crc kubenswrapper[4716]: I1207 16:04:39.625504 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-82bvt" podStartSLOduration=11.625482984 podStartE2EDuration="11.625482984s" podCreationTimestamp="2025-12-07 16:04:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:04:39.624224928 +0000 UTC m=+142.314509840" watchObservedRunningTime="2025-12-07 16:04:39.625482984 +0000 UTC m=+142.315767896" Dec 07 16:04:39 crc kubenswrapper[4716]: I1207 16:04:39.642741 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 07 16:04:39 crc kubenswrapper[4716]: I1207 16:04:39.673519 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/12e72f4d-2887-4c37-bf16-403f124d6f68-utilities\") pod \"redhat-marketplace-fnkww\" (UID: \"12e72f4d-2887-4c37-bf16-403f124d6f68\") " pod="openshift-marketplace/redhat-marketplace-fnkww" Dec 07 16:04:39 crc kubenswrapper[4716]: I1207 16:04:39.673714 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/12e72f4d-2887-4c37-bf16-403f124d6f68-catalog-content\") pod \"redhat-marketplace-fnkww\" (UID: \"12e72f4d-2887-4c37-bf16-403f124d6f68\") " pod="openshift-marketplace/redhat-marketplace-fnkww" Dec 07 16:04:39 crc kubenswrapper[4716]: I1207 16:04:39.673732 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fx9hn\" (UniqueName: \"kubernetes.io/projected/12e72f4d-2887-4c37-bf16-403f124d6f68-kube-api-access-fx9hn\") pod \"redhat-marketplace-fnkww\" (UID: \"12e72f4d-2887-4c37-bf16-403f124d6f68\") " pod="openshift-marketplace/redhat-marketplace-fnkww" Dec 07 16:04:39 crc kubenswrapper[4716]: I1207 16:04:39.689573 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Dec 07 16:04:39 crc kubenswrapper[4716]: I1207 16:04:39.690421 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/12e72f4d-2887-4c37-bf16-403f124d6f68-catalog-content\") pod \"redhat-marketplace-fnkww\" (UID: \"12e72f4d-2887-4c37-bf16-403f124d6f68\") " pod="openshift-marketplace/redhat-marketplace-fnkww" Dec 07 16:04:39 crc kubenswrapper[4716]: I1207 16:04:39.690815 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/12e72f4d-2887-4c37-bf16-403f124d6f68-utilities\") pod \"redhat-marketplace-fnkww\" (UID: \"12e72f4d-2887-4c37-bf16-403f124d6f68\") " pod="openshift-marketplace/redhat-marketplace-fnkww" Dec 07 16:04:39 crc kubenswrapper[4716]: I1207 16:04:39.711908 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fx9hn\" (UniqueName: \"kubernetes.io/projected/12e72f4d-2887-4c37-bf16-403f124d6f68-kube-api-access-fx9hn\") pod \"redhat-marketplace-fnkww\" (UID: \"12e72f4d-2887-4c37-bf16-403f124d6f68\") " pod="openshift-marketplace/redhat-marketplace-fnkww" Dec 07 16:04:39 crc kubenswrapper[4716]: I1207 16:04:39.781783 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fnkww" Dec 07 16:04:39 crc kubenswrapper[4716]: I1207 16:04:39.869561 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-8vsbm"] Dec 07 16:04:39 crc kubenswrapper[4716]: I1207 16:04:39.876598 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8vsbm" Dec 07 16:04:39 crc kubenswrapper[4716]: I1207 16:04:39.908235 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8vsbm"] Dec 07 16:04:39 crc kubenswrapper[4716]: I1207 16:04:39.941116 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29418720-x5rqz" Dec 07 16:04:39 crc kubenswrapper[4716]: I1207 16:04:39.980259 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a72aa026-8e89-45c7-86e7-8a5063352df7-catalog-content\") pod \"redhat-marketplace-8vsbm\" (UID: \"a72aa026-8e89-45c7-86e7-8a5063352df7\") " pod="openshift-marketplace/redhat-marketplace-8vsbm" Dec 07 16:04:39 crc kubenswrapper[4716]: I1207 16:04:39.980752 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a72aa026-8e89-45c7-86e7-8a5063352df7-utilities\") pod \"redhat-marketplace-8vsbm\" (UID: \"a72aa026-8e89-45c7-86e7-8a5063352df7\") " pod="openshift-marketplace/redhat-marketplace-8vsbm" Dec 07 16:04:39 crc kubenswrapper[4716]: I1207 16:04:39.980812 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tz647\" (UniqueName: \"kubernetes.io/projected/a72aa026-8e89-45c7-86e7-8a5063352df7-kube-api-access-tz647\") pod \"redhat-marketplace-8vsbm\" (UID: \"a72aa026-8e89-45c7-86e7-8a5063352df7\") " pod="openshift-marketplace/redhat-marketplace-8vsbm" Dec 07 16:04:40 crc kubenswrapper[4716]: I1207 16:04:40.050775 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fnkww"] Dec 07 16:04:40 crc kubenswrapper[4716]: W1207 16:04:40.059518 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod12e72f4d_2887_4c37_bf16_403f124d6f68.slice/crio-54ab04ae5b53e4cd7e6ad02374b0fe0ae54d32e9f91897ef9129dc7d43995839 WatchSource:0}: Error finding container 54ab04ae5b53e4cd7e6ad02374b0fe0ae54d32e9f91897ef9129dc7d43995839: Status 404 returned error can't find the container with id 54ab04ae5b53e4cd7e6ad02374b0fe0ae54d32e9f91897ef9129dc7d43995839 Dec 07 16:04:40 crc kubenswrapper[4716]: I1207 16:04:40.081066 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jmt77\" (UniqueName: \"kubernetes.io/projected/98efd33b-de1f-4414-b9cc-55fb0d423784-kube-api-access-jmt77\") pod \"98efd33b-de1f-4414-b9cc-55fb0d423784\" (UID: \"98efd33b-de1f-4414-b9cc-55fb0d423784\") " Dec 07 16:04:40 crc kubenswrapper[4716]: I1207 16:04:40.081178 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/98efd33b-de1f-4414-b9cc-55fb0d423784-secret-volume\") pod \"98efd33b-de1f-4414-b9cc-55fb0d423784\" (UID: \"98efd33b-de1f-4414-b9cc-55fb0d423784\") " Dec 07 16:04:40 crc kubenswrapper[4716]: I1207 16:04:40.081239 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/98efd33b-de1f-4414-b9cc-55fb0d423784-config-volume\") pod \"98efd33b-de1f-4414-b9cc-55fb0d423784\" (UID: \"98efd33b-de1f-4414-b9cc-55fb0d423784\") " Dec 07 16:04:40 crc kubenswrapper[4716]: I1207 16:04:40.081390 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a72aa026-8e89-45c7-86e7-8a5063352df7-utilities\") pod \"redhat-marketplace-8vsbm\" (UID: \"a72aa026-8e89-45c7-86e7-8a5063352df7\") " pod="openshift-marketplace/redhat-marketplace-8vsbm" Dec 07 16:04:40 crc kubenswrapper[4716]: I1207 16:04:40.081453 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tz647\" (UniqueName: \"kubernetes.io/projected/a72aa026-8e89-45c7-86e7-8a5063352df7-kube-api-access-tz647\") pod \"redhat-marketplace-8vsbm\" (UID: \"a72aa026-8e89-45c7-86e7-8a5063352df7\") " pod="openshift-marketplace/redhat-marketplace-8vsbm" Dec 07 16:04:40 crc kubenswrapper[4716]: I1207 16:04:40.081529 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a72aa026-8e89-45c7-86e7-8a5063352df7-catalog-content\") pod \"redhat-marketplace-8vsbm\" (UID: \"a72aa026-8e89-45c7-86e7-8a5063352df7\") " pod="openshift-marketplace/redhat-marketplace-8vsbm" Dec 07 16:04:40 crc kubenswrapper[4716]: I1207 16:04:40.081994 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a72aa026-8e89-45c7-86e7-8a5063352df7-catalog-content\") pod \"redhat-marketplace-8vsbm\" (UID: \"a72aa026-8e89-45c7-86e7-8a5063352df7\") " pod="openshift-marketplace/redhat-marketplace-8vsbm" Dec 07 16:04:40 crc kubenswrapper[4716]: I1207 16:04:40.082149 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a72aa026-8e89-45c7-86e7-8a5063352df7-utilities\") pod \"redhat-marketplace-8vsbm\" (UID: \"a72aa026-8e89-45c7-86e7-8a5063352df7\") " pod="openshift-marketplace/redhat-marketplace-8vsbm" Dec 07 16:04:40 crc kubenswrapper[4716]: I1207 16:04:40.082580 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/98efd33b-de1f-4414-b9cc-55fb0d423784-config-volume" (OuterVolumeSpecName: "config-volume") pod "98efd33b-de1f-4414-b9cc-55fb0d423784" (UID: "98efd33b-de1f-4414-b9cc-55fb0d423784"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:04:40 crc kubenswrapper[4716]: I1207 16:04:40.087303 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98efd33b-de1f-4414-b9cc-55fb0d423784-kube-api-access-jmt77" (OuterVolumeSpecName: "kube-api-access-jmt77") pod "98efd33b-de1f-4414-b9cc-55fb0d423784" (UID: "98efd33b-de1f-4414-b9cc-55fb0d423784"). InnerVolumeSpecName "kube-api-access-jmt77". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:04:40 crc kubenswrapper[4716]: I1207 16:04:40.087637 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98efd33b-de1f-4414-b9cc-55fb0d423784-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "98efd33b-de1f-4414-b9cc-55fb0d423784" (UID: "98efd33b-de1f-4414-b9cc-55fb0d423784"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:04:40 crc kubenswrapper[4716]: I1207 16:04:40.100193 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tz647\" (UniqueName: \"kubernetes.io/projected/a72aa026-8e89-45c7-86e7-8a5063352df7-kube-api-access-tz647\") pod \"redhat-marketplace-8vsbm\" (UID: \"a72aa026-8e89-45c7-86e7-8a5063352df7\") " pod="openshift-marketplace/redhat-marketplace-8vsbm" Dec 07 16:04:40 crc kubenswrapper[4716]: I1207 16:04:40.172228 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 07 16:04:40 crc kubenswrapper[4716]: I1207 16:04:40.182487 4716 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/98efd33b-de1f-4414-b9cc-55fb0d423784-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 07 16:04:40 crc kubenswrapper[4716]: I1207 16:04:40.182524 4716 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/98efd33b-de1f-4414-b9cc-55fb0d423784-config-volume\") on node \"crc\" DevicePath \"\"" Dec 07 16:04:40 crc kubenswrapper[4716]: I1207 16:04:40.182534 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jmt77\" (UniqueName: \"kubernetes.io/projected/98efd33b-de1f-4414-b9cc-55fb0d423784-kube-api-access-jmt77\") on node \"crc\" DevicePath \"\"" Dec 07 16:04:40 crc kubenswrapper[4716]: W1207 16:04:40.194467 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod15c001b6_abae_4fb3_ac0e_c23c2e006ff1.slice/crio-639f60596f1ed3db30d93722bffd619ab929d7ba4cf1ace04e2a0f1375422e1d WatchSource:0}: Error finding container 639f60596f1ed3db30d93722bffd619ab929d7ba4cf1ace04e2a0f1375422e1d: Status 404 returned error can't find the container with id 639f60596f1ed3db30d93722bffd619ab929d7ba4cf1ace04e2a0f1375422e1d Dec 07 16:04:40 crc kubenswrapper[4716]: I1207 16:04:40.202039 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8vsbm" Dec 07 16:04:40 crc kubenswrapper[4716]: I1207 16:04:40.494035 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8vsbm"] Dec 07 16:04:40 crc kubenswrapper[4716]: I1207 16:04:40.500747 4716 patch_prober.go:28] interesting pod/router-default-5444994796-h5l29 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 07 16:04:40 crc kubenswrapper[4716]: [-]has-synced failed: reason withheld Dec 07 16:04:40 crc kubenswrapper[4716]: [+]process-running ok Dec 07 16:04:40 crc kubenswrapper[4716]: healthz check failed Dec 07 16:04:40 crc kubenswrapper[4716]: I1207 16:04:40.501369 4716 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h5l29" podUID="85bc0929-de17-4d2d-8b03-1a05d207fb55" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 07 16:04:40 crc kubenswrapper[4716]: I1207 16:04:40.544132 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"15c001b6-abae-4fb3-ac0e-c23c2e006ff1","Type":"ContainerStarted","Data":"639f60596f1ed3db30d93722bffd619ab929d7ba4cf1ace04e2a0f1375422e1d"} Dec 07 16:04:40 crc kubenswrapper[4716]: I1207 16:04:40.548540 4716 generic.go:334] "Generic (PLEG): container finished" podID="12e72f4d-2887-4c37-bf16-403f124d6f68" containerID="091f16f5b3fa9f7922dac579894eb8f866cf7bed963db6281ba83385ce94aa70" exitCode=0 Dec 07 16:04:40 crc kubenswrapper[4716]: I1207 16:04:40.548617 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fnkww" event={"ID":"12e72f4d-2887-4c37-bf16-403f124d6f68","Type":"ContainerDied","Data":"091f16f5b3fa9f7922dac579894eb8f866cf7bed963db6281ba83385ce94aa70"} Dec 07 16:04:40 crc kubenswrapper[4716]: I1207 16:04:40.548638 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fnkww" event={"ID":"12e72f4d-2887-4c37-bf16-403f124d6f68","Type":"ContainerStarted","Data":"54ab04ae5b53e4cd7e6ad02374b0fe0ae54d32e9f91897ef9129dc7d43995839"} Dec 07 16:04:40 crc kubenswrapper[4716]: I1207 16:04:40.553782 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29418720-x5rqz" Dec 07 16:04:40 crc kubenswrapper[4716]: I1207 16:04:40.556387 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29418720-x5rqz" event={"ID":"98efd33b-de1f-4414-b9cc-55fb0d423784","Type":"ContainerDied","Data":"9c914c728e0bde869d16f15f74ade82d4b5d12a2b84894b9f9af328f85649956"} Dec 07 16:04:40 crc kubenswrapper[4716]: I1207 16:04:40.556430 4716 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9c914c728e0bde869d16f15f74ade82d4b5d12a2b84894b9f9af328f85649956" Dec 07 16:04:40 crc kubenswrapper[4716]: I1207 16:04:40.667898 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-hqtd7"] Dec 07 16:04:40 crc kubenswrapper[4716]: E1207 16:04:40.668116 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98efd33b-de1f-4414-b9cc-55fb0d423784" containerName="collect-profiles" Dec 07 16:04:40 crc kubenswrapper[4716]: I1207 16:04:40.668129 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="98efd33b-de1f-4414-b9cc-55fb0d423784" containerName="collect-profiles" Dec 07 16:04:40 crc kubenswrapper[4716]: I1207 16:04:40.668275 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="98efd33b-de1f-4414-b9cc-55fb0d423784" containerName="collect-profiles" Dec 07 16:04:40 crc kubenswrapper[4716]: I1207 16:04:40.670703 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hqtd7" Dec 07 16:04:40 crc kubenswrapper[4716]: I1207 16:04:40.673526 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 07 16:04:40 crc kubenswrapper[4716]: I1207 16:04:40.682755 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hqtd7"] Dec 07 16:04:40 crc kubenswrapper[4716]: I1207 16:04:40.699941 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86df58f0-8580-4c66-9667-4f7cc285f3a0-catalog-content\") pod \"redhat-operators-hqtd7\" (UID: \"86df58f0-8580-4c66-9667-4f7cc285f3a0\") " pod="openshift-marketplace/redhat-operators-hqtd7" Dec 07 16:04:40 crc kubenswrapper[4716]: I1207 16:04:40.700174 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86df58f0-8580-4c66-9667-4f7cc285f3a0-utilities\") pod \"redhat-operators-hqtd7\" (UID: \"86df58f0-8580-4c66-9667-4f7cc285f3a0\") " pod="openshift-marketplace/redhat-operators-hqtd7" Dec 07 16:04:40 crc kubenswrapper[4716]: I1207 16:04:40.700431 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxznt\" (UniqueName: \"kubernetes.io/projected/86df58f0-8580-4c66-9667-4f7cc285f3a0-kube-api-access-qxznt\") pod \"redhat-operators-hqtd7\" (UID: \"86df58f0-8580-4c66-9667-4f7cc285f3a0\") " pod="openshift-marketplace/redhat-operators-hqtd7" Dec 07 16:04:40 crc kubenswrapper[4716]: I1207 16:04:40.802006 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxznt\" (UniqueName: \"kubernetes.io/projected/86df58f0-8580-4c66-9667-4f7cc285f3a0-kube-api-access-qxznt\") pod \"redhat-operators-hqtd7\" (UID: \"86df58f0-8580-4c66-9667-4f7cc285f3a0\") " pod="openshift-marketplace/redhat-operators-hqtd7" Dec 07 16:04:40 crc kubenswrapper[4716]: I1207 16:04:40.802111 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86df58f0-8580-4c66-9667-4f7cc285f3a0-catalog-content\") pod \"redhat-operators-hqtd7\" (UID: \"86df58f0-8580-4c66-9667-4f7cc285f3a0\") " pod="openshift-marketplace/redhat-operators-hqtd7" Dec 07 16:04:40 crc kubenswrapper[4716]: I1207 16:04:40.802179 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86df58f0-8580-4c66-9667-4f7cc285f3a0-utilities\") pod \"redhat-operators-hqtd7\" (UID: \"86df58f0-8580-4c66-9667-4f7cc285f3a0\") " pod="openshift-marketplace/redhat-operators-hqtd7" Dec 07 16:04:40 crc kubenswrapper[4716]: I1207 16:04:40.802697 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86df58f0-8580-4c66-9667-4f7cc285f3a0-utilities\") pod \"redhat-operators-hqtd7\" (UID: \"86df58f0-8580-4c66-9667-4f7cc285f3a0\") " pod="openshift-marketplace/redhat-operators-hqtd7" Dec 07 16:04:40 crc kubenswrapper[4716]: I1207 16:04:40.803011 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86df58f0-8580-4c66-9667-4f7cc285f3a0-catalog-content\") pod \"redhat-operators-hqtd7\" (UID: \"86df58f0-8580-4c66-9667-4f7cc285f3a0\") " pod="openshift-marketplace/redhat-operators-hqtd7" Dec 07 16:04:40 crc kubenswrapper[4716]: I1207 16:04:40.824916 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxznt\" (UniqueName: \"kubernetes.io/projected/86df58f0-8580-4c66-9667-4f7cc285f3a0-kube-api-access-qxznt\") pod \"redhat-operators-hqtd7\" (UID: \"86df58f0-8580-4c66-9667-4f7cc285f3a0\") " pod="openshift-marketplace/redhat-operators-hqtd7" Dec 07 16:04:41 crc kubenswrapper[4716]: I1207 16:04:41.007204 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hqtd7" Dec 07 16:04:41 crc kubenswrapper[4716]: I1207 16:04:41.083219 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-hwbw9"] Dec 07 16:04:41 crc kubenswrapper[4716]: I1207 16:04:41.087600 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hwbw9"] Dec 07 16:04:41 crc kubenswrapper[4716]: I1207 16:04:41.088037 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hwbw9" Dec 07 16:04:41 crc kubenswrapper[4716]: I1207 16:04:41.211999 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9ab4b15-9882-4125-ba5e-1f5f236c5b6b-catalog-content\") pod \"redhat-operators-hwbw9\" (UID: \"b9ab4b15-9882-4125-ba5e-1f5f236c5b6b\") " pod="openshift-marketplace/redhat-operators-hwbw9" Dec 07 16:04:41 crc kubenswrapper[4716]: I1207 16:04:41.212046 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9ab4b15-9882-4125-ba5e-1f5f236c5b6b-utilities\") pod \"redhat-operators-hwbw9\" (UID: \"b9ab4b15-9882-4125-ba5e-1f5f236c5b6b\") " pod="openshift-marketplace/redhat-operators-hwbw9" Dec 07 16:04:41 crc kubenswrapper[4716]: I1207 16:04:41.212233 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7gd98\" (UniqueName: \"kubernetes.io/projected/b9ab4b15-9882-4125-ba5e-1f5f236c5b6b-kube-api-access-7gd98\") pod \"redhat-operators-hwbw9\" (UID: \"b9ab4b15-9882-4125-ba5e-1f5f236c5b6b\") " pod="openshift-marketplace/redhat-operators-hwbw9" Dec 07 16:04:41 crc kubenswrapper[4716]: I1207 16:04:41.306954 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-gc4cp" Dec 07 16:04:41 crc kubenswrapper[4716]: I1207 16:04:41.307414 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-gc4cp" Dec 07 16:04:41 crc kubenswrapper[4716]: I1207 16:04:41.313624 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7gd98\" (UniqueName: \"kubernetes.io/projected/b9ab4b15-9882-4125-ba5e-1f5f236c5b6b-kube-api-access-7gd98\") pod \"redhat-operators-hwbw9\" (UID: \"b9ab4b15-9882-4125-ba5e-1f5f236c5b6b\") " pod="openshift-marketplace/redhat-operators-hwbw9" Dec 07 16:04:41 crc kubenswrapper[4716]: I1207 16:04:41.313826 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9ab4b15-9882-4125-ba5e-1f5f236c5b6b-catalog-content\") pod \"redhat-operators-hwbw9\" (UID: \"b9ab4b15-9882-4125-ba5e-1f5f236c5b6b\") " pod="openshift-marketplace/redhat-operators-hwbw9" Dec 07 16:04:41 crc kubenswrapper[4716]: I1207 16:04:41.313855 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9ab4b15-9882-4125-ba5e-1f5f236c5b6b-utilities\") pod \"redhat-operators-hwbw9\" (UID: \"b9ab4b15-9882-4125-ba5e-1f5f236c5b6b\") " pod="openshift-marketplace/redhat-operators-hwbw9" Dec 07 16:04:41 crc kubenswrapper[4716]: I1207 16:04:41.314499 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9ab4b15-9882-4125-ba5e-1f5f236c5b6b-utilities\") pod \"redhat-operators-hwbw9\" (UID: \"b9ab4b15-9882-4125-ba5e-1f5f236c5b6b\") " pod="openshift-marketplace/redhat-operators-hwbw9" Dec 07 16:04:41 crc kubenswrapper[4716]: I1207 16:04:41.315394 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9ab4b15-9882-4125-ba5e-1f5f236c5b6b-catalog-content\") pod \"redhat-operators-hwbw9\" (UID: \"b9ab4b15-9882-4125-ba5e-1f5f236c5b6b\") " pod="openshift-marketplace/redhat-operators-hwbw9" Dec 07 16:04:41 crc kubenswrapper[4716]: I1207 16:04:41.325561 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-gc4cp" Dec 07 16:04:41 crc kubenswrapper[4716]: I1207 16:04:41.328205 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hqtd7"] Dec 07 16:04:41 crc kubenswrapper[4716]: I1207 16:04:41.331274 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7gd98\" (UniqueName: \"kubernetes.io/projected/b9ab4b15-9882-4125-ba5e-1f5f236c5b6b-kube-api-access-7gd98\") pod \"redhat-operators-hwbw9\" (UID: \"b9ab4b15-9882-4125-ba5e-1f5f236c5b6b\") " pod="openshift-marketplace/redhat-operators-hwbw9" Dec 07 16:04:41 crc kubenswrapper[4716]: I1207 16:04:41.408375 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hwbw9" Dec 07 16:04:41 crc kubenswrapper[4716]: I1207 16:04:41.411765 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-ndp5v" Dec 07 16:04:41 crc kubenswrapper[4716]: I1207 16:04:41.412992 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-ndp5v" Dec 07 16:04:41 crc kubenswrapper[4716]: I1207 16:04:41.416580 4716 patch_prober.go:28] interesting pod/console-f9d7485db-ndp5v container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.12:8443/health\": dial tcp 10.217.0.12:8443: connect: connection refused" start-of-body= Dec 07 16:04:41 crc kubenswrapper[4716]: I1207 16:04:41.416635 4716 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-ndp5v" podUID="cbf3f592-bbb1-4596-8ae1-0e62aa366f0b" containerName="console" probeResult="failure" output="Get \"https://10.217.0.12:8443/health\": dial tcp 10.217.0.12:8443: connect: connection refused" Dec 07 16:04:41 crc kubenswrapper[4716]: I1207 16:04:41.460639 4716 patch_prober.go:28] interesting pod/downloads-7954f5f757-b257c container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" start-of-body= Dec 07 16:04:41 crc kubenswrapper[4716]: I1207 16:04:41.460696 4716 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-b257c" podUID="1306f3b2-2982-494e-9db0-86c2e05b7a82" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" Dec 07 16:04:41 crc kubenswrapper[4716]: I1207 16:04:41.460711 4716 patch_prober.go:28] interesting pod/downloads-7954f5f757-b257c container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" start-of-body= Dec 07 16:04:41 crc kubenswrapper[4716]: I1207 16:04:41.460790 4716 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-b257c" podUID="1306f3b2-2982-494e-9db0-86c2e05b7a82" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" Dec 07 16:04:41 crc kubenswrapper[4716]: I1207 16:04:41.484033 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-h5l29" Dec 07 16:04:41 crc kubenswrapper[4716]: I1207 16:04:41.486949 4716 patch_prober.go:28] interesting pod/router-default-5444994796-h5l29 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 07 16:04:41 crc kubenswrapper[4716]: [-]has-synced failed: reason withheld Dec 07 16:04:41 crc kubenswrapper[4716]: [+]process-running ok Dec 07 16:04:41 crc kubenswrapper[4716]: healthz check failed Dec 07 16:04:41 crc kubenswrapper[4716]: I1207 16:04:41.487012 4716 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h5l29" podUID="85bc0929-de17-4d2d-8b03-1a05d207fb55" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 07 16:04:41 crc kubenswrapper[4716]: I1207 16:04:41.587889 4716 generic.go:334] "Generic (PLEG): container finished" podID="15c001b6-abae-4fb3-ac0e-c23c2e006ff1" containerID="a33654d75b2657f188d9f1be3669412ea7875c46cc4fe8da0fd5f22b7c8d3045" exitCode=0 Dec 07 16:04:41 crc kubenswrapper[4716]: I1207 16:04:41.587999 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"15c001b6-abae-4fb3-ac0e-c23c2e006ff1","Type":"ContainerDied","Data":"a33654d75b2657f188d9f1be3669412ea7875c46cc4fe8da0fd5f22b7c8d3045"} Dec 07 16:04:41 crc kubenswrapper[4716]: I1207 16:04:41.595346 4716 generic.go:334] "Generic (PLEG): container finished" podID="a72aa026-8e89-45c7-86e7-8a5063352df7" containerID="b7a4b01aad1cec88d9bc8a76fce9875c2af3534767fa8f296c5faff7b25d2438" exitCode=0 Dec 07 16:04:41 crc kubenswrapper[4716]: I1207 16:04:41.595509 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8vsbm" event={"ID":"a72aa026-8e89-45c7-86e7-8a5063352df7","Type":"ContainerDied","Data":"b7a4b01aad1cec88d9bc8a76fce9875c2af3534767fa8f296c5faff7b25d2438"} Dec 07 16:04:41 crc kubenswrapper[4716]: I1207 16:04:41.595558 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8vsbm" event={"ID":"a72aa026-8e89-45c7-86e7-8a5063352df7","Type":"ContainerStarted","Data":"662a9f2b50de7fb46858e2cba9008422abac14b326d4a10f70abb53a54567b87"} Dec 07 16:04:41 crc kubenswrapper[4716]: I1207 16:04:41.600216 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hqtd7" event={"ID":"86df58f0-8580-4c66-9667-4f7cc285f3a0","Type":"ContainerStarted","Data":"8e7d933557dd9d1c018fdba795e4eb9afed58d9ec6de915f5e58819fa3c8e0f6"} Dec 07 16:04:41 crc kubenswrapper[4716]: I1207 16:04:41.607587 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-gc4cp" Dec 07 16:04:41 crc kubenswrapper[4716]: I1207 16:04:41.937487 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hwbw9"] Dec 07 16:04:42 crc kubenswrapper[4716]: I1207 16:04:42.485883 4716 patch_prober.go:28] interesting pod/router-default-5444994796-h5l29 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 07 16:04:42 crc kubenswrapper[4716]: [-]has-synced failed: reason withheld Dec 07 16:04:42 crc kubenswrapper[4716]: [+]process-running ok Dec 07 16:04:42 crc kubenswrapper[4716]: healthz check failed Dec 07 16:04:42 crc kubenswrapper[4716]: I1207 16:04:42.486252 4716 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h5l29" podUID="85bc0929-de17-4d2d-8b03-1a05d207fb55" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 07 16:04:42 crc kubenswrapper[4716]: I1207 16:04:42.609495 4716 generic.go:334] "Generic (PLEG): container finished" podID="b9ab4b15-9882-4125-ba5e-1f5f236c5b6b" containerID="58df8766dbfacafd471976e4db1de732a7003e3d811ab9b06c481e1e6ed6f4e9" exitCode=0 Dec 07 16:04:42 crc kubenswrapper[4716]: I1207 16:04:42.609583 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hwbw9" event={"ID":"b9ab4b15-9882-4125-ba5e-1f5f236c5b6b","Type":"ContainerDied","Data":"58df8766dbfacafd471976e4db1de732a7003e3d811ab9b06c481e1e6ed6f4e9"} Dec 07 16:04:42 crc kubenswrapper[4716]: I1207 16:04:42.609681 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hwbw9" event={"ID":"b9ab4b15-9882-4125-ba5e-1f5f236c5b6b","Type":"ContainerStarted","Data":"e2b115a91085b70169015f86c8e98d4db5a5c692bd4ed714ae74d67af96fd42a"} Dec 07 16:04:42 crc kubenswrapper[4716]: I1207 16:04:42.616612 4716 generic.go:334] "Generic (PLEG): container finished" podID="86df58f0-8580-4c66-9667-4f7cc285f3a0" containerID="ab9c2e3cac61dba0189c9c8d56eb34e0396604b46791fb46293ba31372950152" exitCode=0 Dec 07 16:04:42 crc kubenswrapper[4716]: I1207 16:04:42.617291 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hqtd7" event={"ID":"86df58f0-8580-4c66-9667-4f7cc285f3a0","Type":"ContainerDied","Data":"ab9c2e3cac61dba0189c9c8d56eb34e0396604b46791fb46293ba31372950152"} Dec 07 16:04:43 crc kubenswrapper[4716]: I1207 16:04:43.021058 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 07 16:04:43 crc kubenswrapper[4716]: I1207 16:04:43.162855 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/15c001b6-abae-4fb3-ac0e-c23c2e006ff1-kube-api-access\") pod \"15c001b6-abae-4fb3-ac0e-c23c2e006ff1\" (UID: \"15c001b6-abae-4fb3-ac0e-c23c2e006ff1\") " Dec 07 16:04:43 crc kubenswrapper[4716]: I1207 16:04:43.162899 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/15c001b6-abae-4fb3-ac0e-c23c2e006ff1-kubelet-dir\") pod \"15c001b6-abae-4fb3-ac0e-c23c2e006ff1\" (UID: \"15c001b6-abae-4fb3-ac0e-c23c2e006ff1\") " Dec 07 16:04:43 crc kubenswrapper[4716]: I1207 16:04:43.164665 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/15c001b6-abae-4fb3-ac0e-c23c2e006ff1-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "15c001b6-abae-4fb3-ac0e-c23c2e006ff1" (UID: "15c001b6-abae-4fb3-ac0e-c23c2e006ff1"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 16:04:43 crc kubenswrapper[4716]: I1207 16:04:43.193780 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15c001b6-abae-4fb3-ac0e-c23c2e006ff1-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "15c001b6-abae-4fb3-ac0e-c23c2e006ff1" (UID: "15c001b6-abae-4fb3-ac0e-c23c2e006ff1"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:04:43 crc kubenswrapper[4716]: I1207 16:04:43.266260 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/15c001b6-abae-4fb3-ac0e-c23c2e006ff1-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 07 16:04:43 crc kubenswrapper[4716]: I1207 16:04:43.266292 4716 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/15c001b6-abae-4fb3-ac0e-c23c2e006ff1-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 07 16:04:43 crc kubenswrapper[4716]: I1207 16:04:43.486653 4716 patch_prober.go:28] interesting pod/router-default-5444994796-h5l29 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 07 16:04:43 crc kubenswrapper[4716]: [-]has-synced failed: reason withheld Dec 07 16:04:43 crc kubenswrapper[4716]: [+]process-running ok Dec 07 16:04:43 crc kubenswrapper[4716]: healthz check failed Dec 07 16:04:43 crc kubenswrapper[4716]: I1207 16:04:43.486702 4716 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h5l29" podUID="85bc0929-de17-4d2d-8b03-1a05d207fb55" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 07 16:04:43 crc kubenswrapper[4716]: I1207 16:04:43.570118 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 16:04:43 crc kubenswrapper[4716]: I1207 16:04:43.570191 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 16:04:43 crc kubenswrapper[4716]: I1207 16:04:43.574193 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 16:04:43 crc kubenswrapper[4716]: I1207 16:04:43.587823 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 16:04:43 crc kubenswrapper[4716]: I1207 16:04:43.638060 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"15c001b6-abae-4fb3-ac0e-c23c2e006ff1","Type":"ContainerDied","Data":"639f60596f1ed3db30d93722bffd619ab929d7ba4cf1ace04e2a0f1375422e1d"} Dec 07 16:04:43 crc kubenswrapper[4716]: I1207 16:04:43.638146 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 07 16:04:43 crc kubenswrapper[4716]: I1207 16:04:43.639184 4716 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="639f60596f1ed3db30d93722bffd619ab929d7ba4cf1ace04e2a0f1375422e1d" Dec 07 16:04:43 crc kubenswrapper[4716]: I1207 16:04:43.647775 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 07 16:04:43 crc kubenswrapper[4716]: E1207 16:04:43.648710 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15c001b6-abae-4fb3-ac0e-c23c2e006ff1" containerName="pruner" Dec 07 16:04:43 crc kubenswrapper[4716]: I1207 16:04:43.648727 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="15c001b6-abae-4fb3-ac0e-c23c2e006ff1" containerName="pruner" Dec 07 16:04:43 crc kubenswrapper[4716]: I1207 16:04:43.648832 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="15c001b6-abae-4fb3-ac0e-c23c2e006ff1" containerName="pruner" Dec 07 16:04:43 crc kubenswrapper[4716]: I1207 16:04:43.649246 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 07 16:04:43 crc kubenswrapper[4716]: I1207 16:04:43.651650 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 07 16:04:43 crc kubenswrapper[4716]: I1207 16:04:43.652019 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 07 16:04:43 crc kubenswrapper[4716]: I1207 16:04:43.652269 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 07 16:04:43 crc kubenswrapper[4716]: I1207 16:04:43.674982 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:04:43 crc kubenswrapper[4716]: I1207 16:04:43.675117 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:04:43 crc kubenswrapper[4716]: I1207 16:04:43.681321 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 07 16:04:43 crc kubenswrapper[4716]: I1207 16:04:43.697208 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:04:43 crc kubenswrapper[4716]: I1207 16:04:43.705754 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:04:43 crc kubenswrapper[4716]: I1207 16:04:43.777280 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c7641aa3-18ab-4a97-90ea-827a959fa9ff-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"c7641aa3-18ab-4a97-90ea-827a959fa9ff\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 07 16:04:43 crc kubenswrapper[4716]: I1207 16:04:43.777765 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c7641aa3-18ab-4a97-90ea-827a959fa9ff-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"c7641aa3-18ab-4a97-90ea-827a959fa9ff\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 07 16:04:43 crc kubenswrapper[4716]: I1207 16:04:43.785733 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 16:04:43 crc kubenswrapper[4716]: I1207 16:04:43.818782 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 07 16:04:43 crc kubenswrapper[4716]: I1207 16:04:43.879940 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c7641aa3-18ab-4a97-90ea-827a959fa9ff-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"c7641aa3-18ab-4a97-90ea-827a959fa9ff\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 07 16:04:43 crc kubenswrapper[4716]: I1207 16:04:43.880583 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c7641aa3-18ab-4a97-90ea-827a959fa9ff-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"c7641aa3-18ab-4a97-90ea-827a959fa9ff\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 07 16:04:43 crc kubenswrapper[4716]: I1207 16:04:43.880668 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c7641aa3-18ab-4a97-90ea-827a959fa9ff-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"c7641aa3-18ab-4a97-90ea-827a959fa9ff\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 07 16:04:43 crc kubenswrapper[4716]: I1207 16:04:43.905981 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c7641aa3-18ab-4a97-90ea-827a959fa9ff-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"c7641aa3-18ab-4a97-90ea-827a959fa9ff\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 07 16:04:43 crc kubenswrapper[4716]: I1207 16:04:43.983862 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 07 16:04:44 crc kubenswrapper[4716]: I1207 16:04:44.301351 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 07 16:04:44 crc kubenswrapper[4716]: W1207 16:04:44.394788 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-1e25550b04bf22d7c86cc8fdfe28033028626dfca570ad5c8babf4b72da49382 WatchSource:0}: Error finding container 1e25550b04bf22d7c86cc8fdfe28033028626dfca570ad5c8babf4b72da49382: Status 404 returned error can't find the container with id 1e25550b04bf22d7c86cc8fdfe28033028626dfca570ad5c8babf4b72da49382 Dec 07 16:04:44 crc kubenswrapper[4716]: I1207 16:04:44.488539 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-h5l29" Dec 07 16:04:44 crc kubenswrapper[4716]: W1207 16:04:44.495498 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-57bb6cad9b6d6d37e111b15375a543f841a0005d6615b33ddd7295f3c8f0334b WatchSource:0}: Error finding container 57bb6cad9b6d6d37e111b15375a543f841a0005d6615b33ddd7295f3c8f0334b: Status 404 returned error can't find the container with id 57bb6cad9b6d6d37e111b15375a543f841a0005d6615b33ddd7295f3c8f0334b Dec 07 16:04:44 crc kubenswrapper[4716]: I1207 16:04:44.496968 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-h5l29" Dec 07 16:04:44 crc kubenswrapper[4716]: I1207 16:04:44.656636 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"1e25550b04bf22d7c86cc8fdfe28033028626dfca570ad5c8babf4b72da49382"} Dec 07 16:04:44 crc kubenswrapper[4716]: I1207 16:04:44.658451 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"5b4cc27af4383132176a83bef69bf9dd523400016bebe9ef240a823347fac835"} Dec 07 16:04:44 crc kubenswrapper[4716]: I1207 16:04:44.660036 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"c7641aa3-18ab-4a97-90ea-827a959fa9ff","Type":"ContainerStarted","Data":"2847b08292a05b27d020eaa968b93be991dc13f7d2fa434a77453771d886203c"} Dec 07 16:04:44 crc kubenswrapper[4716]: I1207 16:04:44.662501 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"57bb6cad9b6d6d37e111b15375a543f841a0005d6615b33ddd7295f3c8f0334b"} Dec 07 16:04:45 crc kubenswrapper[4716]: I1207 16:04:45.682859 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"2848af6dc9d99fe26104aeae0c74fada597d81da9498fe5eb27ccc25501c43d8"} Dec 07 16:04:45 crc kubenswrapper[4716]: I1207 16:04:45.688515 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"61be60713bb5ba47ffef66b7cb61f97e4d2f4f64ce47f02feb1ef457b21c1e7f"} Dec 07 16:04:45 crc kubenswrapper[4716]: I1207 16:04:45.692590 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"d226072f95ea22a141a54cb51cc38b2867391c39223c4a2e6dab27b06f558541"} Dec 07 16:04:45 crc kubenswrapper[4716]: I1207 16:04:45.692898 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 16:04:46 crc kubenswrapper[4716]: I1207 16:04:46.040381 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" Dec 07 16:04:46 crc kubenswrapper[4716]: I1207 16:04:46.723252 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"c7641aa3-18ab-4a97-90ea-827a959fa9ff","Type":"ContainerStarted","Data":"413a8acf7fcac30f1a349cd249fb19ac7e857363cc854823eb8ee9737589e1b5"} Dec 07 16:04:46 crc kubenswrapper[4716]: I1207 16:04:46.751789 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=3.751755904 podStartE2EDuration="3.751755904s" podCreationTimestamp="2025-12-07 16:04:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:04:46.748446198 +0000 UTC m=+149.438731110" watchObservedRunningTime="2025-12-07 16:04:46.751755904 +0000 UTC m=+149.442040816" Dec 07 16:04:47 crc kubenswrapper[4716]: I1207 16:04:47.300563 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-n4492" Dec 07 16:04:47 crc kubenswrapper[4716]: I1207 16:04:47.736423 4716 generic.go:334] "Generic (PLEG): container finished" podID="c7641aa3-18ab-4a97-90ea-827a959fa9ff" containerID="413a8acf7fcac30f1a349cd249fb19ac7e857363cc854823eb8ee9737589e1b5" exitCode=0 Dec 07 16:04:47 crc kubenswrapper[4716]: I1207 16:04:47.736493 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"c7641aa3-18ab-4a97-90ea-827a959fa9ff","Type":"ContainerDied","Data":"413a8acf7fcac30f1a349cd249fb19ac7e857363cc854823eb8ee9737589e1b5"} Dec 07 16:04:51 crc kubenswrapper[4716]: I1207 16:04:51.416068 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-ndp5v" Dec 07 16:04:51 crc kubenswrapper[4716]: I1207 16:04:51.420024 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-ndp5v" Dec 07 16:04:51 crc kubenswrapper[4716]: I1207 16:04:51.479714 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-b257c" Dec 07 16:04:52 crc kubenswrapper[4716]: I1207 16:04:52.761538 4716 patch_prober.go:28] interesting pod/machine-config-daemon-zcxxp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 16:04:52 crc kubenswrapper[4716]: I1207 16:04:52.761924 4716 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 16:04:52 crc kubenswrapper[4716]: I1207 16:04:52.849808 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 07 16:04:52 crc kubenswrapper[4716]: I1207 16:04:52.867299 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c7641aa3-18ab-4a97-90ea-827a959fa9ff-kube-api-access\") pod \"c7641aa3-18ab-4a97-90ea-827a959fa9ff\" (UID: \"c7641aa3-18ab-4a97-90ea-827a959fa9ff\") " Dec 07 16:04:52 crc kubenswrapper[4716]: I1207 16:04:52.867336 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c7641aa3-18ab-4a97-90ea-827a959fa9ff-kubelet-dir\") pod \"c7641aa3-18ab-4a97-90ea-827a959fa9ff\" (UID: \"c7641aa3-18ab-4a97-90ea-827a959fa9ff\") " Dec 07 16:04:52 crc kubenswrapper[4716]: I1207 16:04:52.867787 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c7641aa3-18ab-4a97-90ea-827a959fa9ff-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "c7641aa3-18ab-4a97-90ea-827a959fa9ff" (UID: "c7641aa3-18ab-4a97-90ea-827a959fa9ff"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 16:04:52 crc kubenswrapper[4716]: I1207 16:04:52.873034 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7641aa3-18ab-4a97-90ea-827a959fa9ff-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "c7641aa3-18ab-4a97-90ea-827a959fa9ff" (UID: "c7641aa3-18ab-4a97-90ea-827a959fa9ff"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:04:52 crc kubenswrapper[4716]: I1207 16:04:52.969193 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c7641aa3-18ab-4a97-90ea-827a959fa9ff-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 07 16:04:52 crc kubenswrapper[4716]: I1207 16:04:52.969231 4716 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c7641aa3-18ab-4a97-90ea-827a959fa9ff-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 07 16:04:53 crc kubenswrapper[4716]: I1207 16:04:53.772853 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"c7641aa3-18ab-4a97-90ea-827a959fa9ff","Type":"ContainerDied","Data":"2847b08292a05b27d020eaa968b93be991dc13f7d2fa434a77453771d886203c"} Dec 07 16:04:53 crc kubenswrapper[4716]: I1207 16:04:53.772937 4716 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2847b08292a05b27d020eaa968b93be991dc13f7d2fa434a77453771d886203c" Dec 07 16:04:53 crc kubenswrapper[4716]: I1207 16:04:53.773014 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 07 16:04:58 crc kubenswrapper[4716]: I1207 16:04:58.439622 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-4d257" Dec 07 16:05:02 crc kubenswrapper[4716]: I1207 16:05:02.286297 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/04b37f6c-457b-4f9a-8733-e88cb9644483-metrics-certs\") pod \"network-metrics-daemon-nx2lh\" (UID: \"04b37f6c-457b-4f9a-8733-e88cb9644483\") " pod="openshift-multus/network-metrics-daemon-nx2lh" Dec 07 16:05:02 crc kubenswrapper[4716]: I1207 16:05:02.295139 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/04b37f6c-457b-4f9a-8733-e88cb9644483-metrics-certs\") pod \"network-metrics-daemon-nx2lh\" (UID: \"04b37f6c-457b-4f9a-8733-e88cb9644483\") " pod="openshift-multus/network-metrics-daemon-nx2lh" Dec 07 16:05:02 crc kubenswrapper[4716]: I1207 16:05:02.405011 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nx2lh" Dec 07 16:05:07 crc kubenswrapper[4716]: E1207 16:05:07.882241 4716 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 07 16:05:07 crc kubenswrapper[4716]: E1207 16:05:07.883194 4716 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fx9hn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-fnkww_openshift-marketplace(12e72f4d-2887-4c37-bf16-403f124d6f68): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 07 16:05:07 crc kubenswrapper[4716]: E1207 16:05:07.885063 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-fnkww" podUID="12e72f4d-2887-4c37-bf16-403f124d6f68" Dec 07 16:05:10 crc kubenswrapper[4716]: E1207 16:05:10.183496 4716 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 07 16:05:10 crc kubenswrapper[4716]: E1207 16:05:10.183646 4716 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wn54r,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-cd798_openshift-marketplace(45c781c9-df95-4b6e-bed3-fc1f22f079a3): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 07 16:05:10 crc kubenswrapper[4716]: E1207 16:05:10.185022 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-cd798" podUID="45c781c9-df95-4b6e-bed3-fc1f22f079a3" Dec 07 16:05:12 crc kubenswrapper[4716]: I1207 16:05:12.252239 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-89v7m" Dec 07 16:05:13 crc kubenswrapper[4716]: E1207 16:05:13.116352 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-cd798" podUID="45c781c9-df95-4b6e-bed3-fc1f22f079a3" Dec 07 16:05:13 crc kubenswrapper[4716]: E1207 16:05:13.117012 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-fnkww" podUID="12e72f4d-2887-4c37-bf16-403f124d6f68" Dec 07 16:05:13 crc kubenswrapper[4716]: E1207 16:05:13.177187 4716 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 07 16:05:13 crc kubenswrapper[4716]: E1207 16:05:13.177382 4716 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qxznt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-hqtd7_openshift-marketplace(86df58f0-8580-4c66-9667-4f7cc285f3a0): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 07 16:05:13 crc kubenswrapper[4716]: E1207 16:05:13.179499 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-hqtd7" podUID="86df58f0-8580-4c66-9667-4f7cc285f3a0" Dec 07 16:05:13 crc kubenswrapper[4716]: E1207 16:05:13.309939 4716 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 07 16:05:13 crc kubenswrapper[4716]: E1207 16:05:13.310441 4716 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9tbdw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-chrks_openshift-marketplace(de1318df-206b-4ad8-98c7-b9c943560e44): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 07 16:05:13 crc kubenswrapper[4716]: E1207 16:05:13.311671 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-chrks" podUID="de1318df-206b-4ad8-98c7-b9c943560e44" Dec 07 16:05:14 crc kubenswrapper[4716]: E1207 16:05:14.421544 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-hqtd7" podUID="86df58f0-8580-4c66-9667-4f7cc285f3a0" Dec 07 16:05:14 crc kubenswrapper[4716]: E1207 16:05:14.421898 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-chrks" podUID="de1318df-206b-4ad8-98c7-b9c943560e44" Dec 07 16:05:14 crc kubenswrapper[4716]: E1207 16:05:14.509006 4716 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 07 16:05:14 crc kubenswrapper[4716]: E1207 16:05:14.509402 4716 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gvh45,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-d92nn_openshift-marketplace(074b6ff7-7e20-4054-811d-34050c07e74e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 07 16:05:14 crc kubenswrapper[4716]: E1207 16:05:14.510467 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-d92nn" podUID="074b6ff7-7e20-4054-811d-34050c07e74e" Dec 07 16:05:14 crc kubenswrapper[4716]: E1207 16:05:14.510691 4716 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 07 16:05:14 crc kubenswrapper[4716]: E1207 16:05:14.510819 4716 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9lbc5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-6wb5h_openshift-marketplace(96a76712-c754-4978-902a-fa07f6c3c634): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 07 16:05:14 crc kubenswrapper[4716]: E1207 16:05:14.512271 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-6wb5h" podUID="96a76712-c754-4978-902a-fa07f6c3c634" Dec 07 16:05:14 crc kubenswrapper[4716]: I1207 16:05:14.819772 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-nx2lh"] Dec 07 16:05:14 crc kubenswrapper[4716]: I1207 16:05:14.884352 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hwbw9" event={"ID":"b9ab4b15-9882-4125-ba5e-1f5f236c5b6b","Type":"ContainerStarted","Data":"b7428c7ac94508590984bc89225dd9ca188a085173594dfec1dd0e3746a90623"} Dec 07 16:05:14 crc kubenswrapper[4716]: I1207 16:05:14.886053 4716 generic.go:334] "Generic (PLEG): container finished" podID="a72aa026-8e89-45c7-86e7-8a5063352df7" containerID="a46985c7bb3495e5b17894566c44c64ef5121e80a63243c39da22efc260b7c2a" exitCode=0 Dec 07 16:05:14 crc kubenswrapper[4716]: I1207 16:05:14.886125 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8vsbm" event={"ID":"a72aa026-8e89-45c7-86e7-8a5063352df7","Type":"ContainerDied","Data":"a46985c7bb3495e5b17894566c44c64ef5121e80a63243c39da22efc260b7c2a"} Dec 07 16:05:14 crc kubenswrapper[4716]: I1207 16:05:14.887385 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-nx2lh" event={"ID":"04b37f6c-457b-4f9a-8733-e88cb9644483","Type":"ContainerStarted","Data":"9a17f52189fd4d309c1443e5a1eb9e01111dce1ea45aac0d3030980edb8dcf54"} Dec 07 16:05:14 crc kubenswrapper[4716]: E1207 16:05:14.888824 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-d92nn" podUID="074b6ff7-7e20-4054-811d-34050c07e74e" Dec 07 16:05:14 crc kubenswrapper[4716]: E1207 16:05:14.889142 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-6wb5h" podUID="96a76712-c754-4978-902a-fa07f6c3c634" Dec 07 16:05:15 crc kubenswrapper[4716]: I1207 16:05:15.897113 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8vsbm" event={"ID":"a72aa026-8e89-45c7-86e7-8a5063352df7","Type":"ContainerStarted","Data":"637bd73718eb950436f699e082b532f8e46f94aa3251c87ca84d979a2bf78255"} Dec 07 16:05:15 crc kubenswrapper[4716]: I1207 16:05:15.900272 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-nx2lh" event={"ID":"04b37f6c-457b-4f9a-8733-e88cb9644483","Type":"ContainerStarted","Data":"310cc7ba329124224fde7a3896d5ffcc3df3287e70b05a195864a7ee7d2ffd2e"} Dec 07 16:05:15 crc kubenswrapper[4716]: I1207 16:05:15.900299 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-nx2lh" event={"ID":"04b37f6c-457b-4f9a-8733-e88cb9644483","Type":"ContainerStarted","Data":"dc77ced3c761949c249178f06c7c9474d57607046c367c11b98c9f33178836f2"} Dec 07 16:05:15 crc kubenswrapper[4716]: I1207 16:05:15.902254 4716 generic.go:334] "Generic (PLEG): container finished" podID="b9ab4b15-9882-4125-ba5e-1f5f236c5b6b" containerID="b7428c7ac94508590984bc89225dd9ca188a085173594dfec1dd0e3746a90623" exitCode=0 Dec 07 16:05:15 crc kubenswrapper[4716]: I1207 16:05:15.902292 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hwbw9" event={"ID":"b9ab4b15-9882-4125-ba5e-1f5f236c5b6b","Type":"ContainerDied","Data":"b7428c7ac94508590984bc89225dd9ca188a085173594dfec1dd0e3746a90623"} Dec 07 16:05:15 crc kubenswrapper[4716]: I1207 16:05:15.919766 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-8vsbm" podStartSLOduration=3.243002831 podStartE2EDuration="36.919740522s" podCreationTimestamp="2025-12-07 16:04:39 +0000 UTC" firstStartedPulling="2025-12-07 16:04:41.597948211 +0000 UTC m=+144.288233123" lastFinishedPulling="2025-12-07 16:05:15.274685902 +0000 UTC m=+177.964970814" observedRunningTime="2025-12-07 16:05:15.917095207 +0000 UTC m=+178.607380129" watchObservedRunningTime="2025-12-07 16:05:15.919740522 +0000 UTC m=+178.610025434" Dec 07 16:05:15 crc kubenswrapper[4716]: I1207 16:05:15.937704 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-nx2lh" podStartSLOduration=156.93768069 podStartE2EDuration="2m36.93768069s" podCreationTimestamp="2025-12-07 16:02:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:05:15.928971219 +0000 UTC m=+178.619256131" watchObservedRunningTime="2025-12-07 16:05:15.93768069 +0000 UTC m=+178.627965612" Dec 07 16:05:16 crc kubenswrapper[4716]: I1207 16:05:16.908816 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hwbw9" event={"ID":"b9ab4b15-9882-4125-ba5e-1f5f236c5b6b","Type":"ContainerStarted","Data":"f018e88f4a0d9680a2947bc0b33060f9051d838316ed125f7c151e897d2f099e"} Dec 07 16:05:16 crc kubenswrapper[4716]: I1207 16:05:16.928909 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-hwbw9" podStartSLOduration=2.079554791 podStartE2EDuration="35.928884746s" podCreationTimestamp="2025-12-07 16:04:41 +0000 UTC" firstStartedPulling="2025-12-07 16:04:42.612028487 +0000 UTC m=+145.302313399" lastFinishedPulling="2025-12-07 16:05:16.461358442 +0000 UTC m=+179.151643354" observedRunningTime="2025-12-07 16:05:16.925687194 +0000 UTC m=+179.615972126" watchObservedRunningTime="2025-12-07 16:05:16.928884746 +0000 UTC m=+179.619169658" Dec 07 16:05:17 crc kubenswrapper[4716]: I1207 16:05:17.852168 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 07 16:05:17 crc kubenswrapper[4716]: E1207 16:05:17.856687 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7641aa3-18ab-4a97-90ea-827a959fa9ff" containerName="pruner" Dec 07 16:05:17 crc kubenswrapper[4716]: I1207 16:05:17.856744 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7641aa3-18ab-4a97-90ea-827a959fa9ff" containerName="pruner" Dec 07 16:05:17 crc kubenswrapper[4716]: I1207 16:05:17.857250 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7641aa3-18ab-4a97-90ea-827a959fa9ff" containerName="pruner" Dec 07 16:05:17 crc kubenswrapper[4716]: I1207 16:05:17.858621 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 07 16:05:17 crc kubenswrapper[4716]: I1207 16:05:17.858737 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 07 16:05:17 crc kubenswrapper[4716]: I1207 16:05:17.863610 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 07 16:05:17 crc kubenswrapper[4716]: I1207 16:05:17.865976 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 07 16:05:17 crc kubenswrapper[4716]: I1207 16:05:17.919381 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/80d3e53e-eb0a-4129-bc62-7b1e445bb06d-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"80d3e53e-eb0a-4129-bc62-7b1e445bb06d\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 07 16:05:17 crc kubenswrapper[4716]: I1207 16:05:17.919750 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/80d3e53e-eb0a-4129-bc62-7b1e445bb06d-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"80d3e53e-eb0a-4129-bc62-7b1e445bb06d\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 07 16:05:18 crc kubenswrapper[4716]: I1207 16:05:18.020994 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/80d3e53e-eb0a-4129-bc62-7b1e445bb06d-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"80d3e53e-eb0a-4129-bc62-7b1e445bb06d\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 07 16:05:18 crc kubenswrapper[4716]: I1207 16:05:18.021066 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/80d3e53e-eb0a-4129-bc62-7b1e445bb06d-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"80d3e53e-eb0a-4129-bc62-7b1e445bb06d\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 07 16:05:18 crc kubenswrapper[4716]: I1207 16:05:18.021182 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/80d3e53e-eb0a-4129-bc62-7b1e445bb06d-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"80d3e53e-eb0a-4129-bc62-7b1e445bb06d\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 07 16:05:18 crc kubenswrapper[4716]: I1207 16:05:18.038154 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/80d3e53e-eb0a-4129-bc62-7b1e445bb06d-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"80d3e53e-eb0a-4129-bc62-7b1e445bb06d\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 07 16:05:18 crc kubenswrapper[4716]: I1207 16:05:18.191350 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 07 16:05:18 crc kubenswrapper[4716]: I1207 16:05:18.598901 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 07 16:05:18 crc kubenswrapper[4716]: I1207 16:05:18.920124 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"80d3e53e-eb0a-4129-bc62-7b1e445bb06d","Type":"ContainerStarted","Data":"c3bd16e1a3adb932345082b228852cbd9de46461c0aa801babad6da4ce1575d5"} Dec 07 16:05:18 crc kubenswrapper[4716]: I1207 16:05:18.920483 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"80d3e53e-eb0a-4129-bc62-7b1e445bb06d","Type":"ContainerStarted","Data":"f56550566aebe76e7b6f22b9de7f6d010e80c0d2cf670d2b0e563d21e1fea0a8"} Dec 07 16:05:18 crc kubenswrapper[4716]: I1207 16:05:18.932725 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=1.9327041459999998 podStartE2EDuration="1.932704146s" podCreationTimestamp="2025-12-07 16:05:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:05:18.931912734 +0000 UTC m=+181.622197646" watchObservedRunningTime="2025-12-07 16:05:18.932704146 +0000 UTC m=+181.622989058" Dec 07 16:05:19 crc kubenswrapper[4716]: I1207 16:05:19.926318 4716 generic.go:334] "Generic (PLEG): container finished" podID="80d3e53e-eb0a-4129-bc62-7b1e445bb06d" containerID="c3bd16e1a3adb932345082b228852cbd9de46461c0aa801babad6da4ce1575d5" exitCode=0 Dec 07 16:05:19 crc kubenswrapper[4716]: I1207 16:05:19.926370 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"80d3e53e-eb0a-4129-bc62-7b1e445bb06d","Type":"ContainerDied","Data":"c3bd16e1a3adb932345082b228852cbd9de46461c0aa801babad6da4ce1575d5"} Dec 07 16:05:20 crc kubenswrapper[4716]: I1207 16:05:20.202707 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-8vsbm" Dec 07 16:05:20 crc kubenswrapper[4716]: I1207 16:05:20.202767 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-8vsbm" Dec 07 16:05:20 crc kubenswrapper[4716]: I1207 16:05:20.383932 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-8vsbm" Dec 07 16:05:20 crc kubenswrapper[4716]: I1207 16:05:20.984858 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-8vsbm" Dec 07 16:05:21 crc kubenswrapper[4716]: I1207 16:05:21.191344 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 07 16:05:21 crc kubenswrapper[4716]: I1207 16:05:21.261763 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/80d3e53e-eb0a-4129-bc62-7b1e445bb06d-kube-api-access\") pod \"80d3e53e-eb0a-4129-bc62-7b1e445bb06d\" (UID: \"80d3e53e-eb0a-4129-bc62-7b1e445bb06d\") " Dec 07 16:05:21 crc kubenswrapper[4716]: I1207 16:05:21.261806 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/80d3e53e-eb0a-4129-bc62-7b1e445bb06d-kubelet-dir\") pod \"80d3e53e-eb0a-4129-bc62-7b1e445bb06d\" (UID: \"80d3e53e-eb0a-4129-bc62-7b1e445bb06d\") " Dec 07 16:05:21 crc kubenswrapper[4716]: I1207 16:05:21.262134 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/80d3e53e-eb0a-4129-bc62-7b1e445bb06d-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "80d3e53e-eb0a-4129-bc62-7b1e445bb06d" (UID: "80d3e53e-eb0a-4129-bc62-7b1e445bb06d"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 16:05:21 crc kubenswrapper[4716]: I1207 16:05:21.272255 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80d3e53e-eb0a-4129-bc62-7b1e445bb06d-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "80d3e53e-eb0a-4129-bc62-7b1e445bb06d" (UID: "80d3e53e-eb0a-4129-bc62-7b1e445bb06d"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:05:21 crc kubenswrapper[4716]: I1207 16:05:21.362916 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/80d3e53e-eb0a-4129-bc62-7b1e445bb06d-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 07 16:05:21 crc kubenswrapper[4716]: I1207 16:05:21.363358 4716 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/80d3e53e-eb0a-4129-bc62-7b1e445bb06d-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 07 16:05:21 crc kubenswrapper[4716]: I1207 16:05:21.409400 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-hwbw9" Dec 07 16:05:21 crc kubenswrapper[4716]: I1207 16:05:21.409626 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-hwbw9" Dec 07 16:05:21 crc kubenswrapper[4716]: I1207 16:05:21.941706 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"80d3e53e-eb0a-4129-bc62-7b1e445bb06d","Type":"ContainerDied","Data":"f56550566aebe76e7b6f22b9de7f6d010e80c0d2cf670d2b0e563d21e1fea0a8"} Dec 07 16:05:21 crc kubenswrapper[4716]: I1207 16:05:21.941776 4716 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f56550566aebe76e7b6f22b9de7f6d010e80c0d2cf670d2b0e563d21e1fea0a8" Dec 07 16:05:21 crc kubenswrapper[4716]: I1207 16:05:21.941820 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 07 16:05:22 crc kubenswrapper[4716]: I1207 16:05:22.109675 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8vsbm"] Dec 07 16:05:22 crc kubenswrapper[4716]: I1207 16:05:22.452999 4716 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-hwbw9" podUID="b9ab4b15-9882-4125-ba5e-1f5f236c5b6b" containerName="registry-server" probeResult="failure" output=< Dec 07 16:05:22 crc kubenswrapper[4716]: timeout: failed to connect service ":50051" within 1s Dec 07 16:05:22 crc kubenswrapper[4716]: > Dec 07 16:05:22 crc kubenswrapper[4716]: I1207 16:05:22.761423 4716 patch_prober.go:28] interesting pod/machine-config-daemon-zcxxp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 16:05:22 crc kubenswrapper[4716]: I1207 16:05:22.761511 4716 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 16:05:22 crc kubenswrapper[4716]: I1207 16:05:22.945210 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-8vsbm" podUID="a72aa026-8e89-45c7-86e7-8a5063352df7" containerName="registry-server" containerID="cri-o://637bd73718eb950436f699e082b532f8e46f94aa3251c87ca84d979a2bf78255" gracePeriod=2 Dec 07 16:05:23 crc kubenswrapper[4716]: I1207 16:05:23.780686 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8vsbm" Dec 07 16:05:23 crc kubenswrapper[4716]: I1207 16:05:23.788841 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 07 16:05:23 crc kubenswrapper[4716]: I1207 16:05:23.898429 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a72aa026-8e89-45c7-86e7-8a5063352df7-catalog-content\") pod \"a72aa026-8e89-45c7-86e7-8a5063352df7\" (UID: \"a72aa026-8e89-45c7-86e7-8a5063352df7\") " Dec 07 16:05:23 crc kubenswrapper[4716]: I1207 16:05:23.898541 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a72aa026-8e89-45c7-86e7-8a5063352df7-utilities\") pod \"a72aa026-8e89-45c7-86e7-8a5063352df7\" (UID: \"a72aa026-8e89-45c7-86e7-8a5063352df7\") " Dec 07 16:05:23 crc kubenswrapper[4716]: I1207 16:05:23.898596 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tz647\" (UniqueName: \"kubernetes.io/projected/a72aa026-8e89-45c7-86e7-8a5063352df7-kube-api-access-tz647\") pod \"a72aa026-8e89-45c7-86e7-8a5063352df7\" (UID: \"a72aa026-8e89-45c7-86e7-8a5063352df7\") " Dec 07 16:05:23 crc kubenswrapper[4716]: I1207 16:05:23.899432 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a72aa026-8e89-45c7-86e7-8a5063352df7-utilities" (OuterVolumeSpecName: "utilities") pod "a72aa026-8e89-45c7-86e7-8a5063352df7" (UID: "a72aa026-8e89-45c7-86e7-8a5063352df7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:05:23 crc kubenswrapper[4716]: I1207 16:05:23.916973 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a72aa026-8e89-45c7-86e7-8a5063352df7-kube-api-access-tz647" (OuterVolumeSpecName: "kube-api-access-tz647") pod "a72aa026-8e89-45c7-86e7-8a5063352df7" (UID: "a72aa026-8e89-45c7-86e7-8a5063352df7"). InnerVolumeSpecName "kube-api-access-tz647". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:05:23 crc kubenswrapper[4716]: I1207 16:05:23.919155 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a72aa026-8e89-45c7-86e7-8a5063352df7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a72aa026-8e89-45c7-86e7-8a5063352df7" (UID: "a72aa026-8e89-45c7-86e7-8a5063352df7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:05:23 crc kubenswrapper[4716]: I1207 16:05:23.950479 4716 generic.go:334] "Generic (PLEG): container finished" podID="a72aa026-8e89-45c7-86e7-8a5063352df7" containerID="637bd73718eb950436f699e082b532f8e46f94aa3251c87ca84d979a2bf78255" exitCode=0 Dec 07 16:05:23 crc kubenswrapper[4716]: I1207 16:05:23.950531 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8vsbm" event={"ID":"a72aa026-8e89-45c7-86e7-8a5063352df7","Type":"ContainerDied","Data":"637bd73718eb950436f699e082b532f8e46f94aa3251c87ca84d979a2bf78255"} Dec 07 16:05:23 crc kubenswrapper[4716]: I1207 16:05:23.950560 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8vsbm" event={"ID":"a72aa026-8e89-45c7-86e7-8a5063352df7","Type":"ContainerDied","Data":"662a9f2b50de7fb46858e2cba9008422abac14b326d4a10f70abb53a54567b87"} Dec 07 16:05:23 crc kubenswrapper[4716]: I1207 16:05:23.950580 4716 scope.go:117] "RemoveContainer" containerID="637bd73718eb950436f699e082b532f8e46f94aa3251c87ca84d979a2bf78255" Dec 07 16:05:23 crc kubenswrapper[4716]: I1207 16:05:23.950532 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8vsbm" Dec 07 16:05:23 crc kubenswrapper[4716]: I1207 16:05:23.976428 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8vsbm"] Dec 07 16:05:23 crc kubenswrapper[4716]: I1207 16:05:23.977913 4716 scope.go:117] "RemoveContainer" containerID="a46985c7bb3495e5b17894566c44c64ef5121e80a63243c39da22efc260b7c2a" Dec 07 16:05:23 crc kubenswrapper[4716]: I1207 16:05:23.979549 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-8vsbm"] Dec 07 16:05:24 crc kubenswrapper[4716]: I1207 16:05:24.000495 4716 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a72aa026-8e89-45c7-86e7-8a5063352df7-utilities\") on node \"crc\" DevicePath \"\"" Dec 07 16:05:24 crc kubenswrapper[4716]: I1207 16:05:24.000578 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tz647\" (UniqueName: \"kubernetes.io/projected/a72aa026-8e89-45c7-86e7-8a5063352df7-kube-api-access-tz647\") on node \"crc\" DevicePath \"\"" Dec 07 16:05:24 crc kubenswrapper[4716]: I1207 16:05:24.000592 4716 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a72aa026-8e89-45c7-86e7-8a5063352df7-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 07 16:05:24 crc kubenswrapper[4716]: I1207 16:05:24.018917 4716 scope.go:117] "RemoveContainer" containerID="b7a4b01aad1cec88d9bc8a76fce9875c2af3534767fa8f296c5faff7b25d2438" Dec 07 16:05:24 crc kubenswrapper[4716]: I1207 16:05:24.040020 4716 scope.go:117] "RemoveContainer" containerID="637bd73718eb950436f699e082b532f8e46f94aa3251c87ca84d979a2bf78255" Dec 07 16:05:24 crc kubenswrapper[4716]: E1207 16:05:24.040615 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"637bd73718eb950436f699e082b532f8e46f94aa3251c87ca84d979a2bf78255\": container with ID starting with 637bd73718eb950436f699e082b532f8e46f94aa3251c87ca84d979a2bf78255 not found: ID does not exist" containerID="637bd73718eb950436f699e082b532f8e46f94aa3251c87ca84d979a2bf78255" Dec 07 16:05:24 crc kubenswrapper[4716]: I1207 16:05:24.040705 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"637bd73718eb950436f699e082b532f8e46f94aa3251c87ca84d979a2bf78255"} err="failed to get container status \"637bd73718eb950436f699e082b532f8e46f94aa3251c87ca84d979a2bf78255\": rpc error: code = NotFound desc = could not find container \"637bd73718eb950436f699e082b532f8e46f94aa3251c87ca84d979a2bf78255\": container with ID starting with 637bd73718eb950436f699e082b532f8e46f94aa3251c87ca84d979a2bf78255 not found: ID does not exist" Dec 07 16:05:24 crc kubenswrapper[4716]: I1207 16:05:24.040791 4716 scope.go:117] "RemoveContainer" containerID="a46985c7bb3495e5b17894566c44c64ef5121e80a63243c39da22efc260b7c2a" Dec 07 16:05:24 crc kubenswrapper[4716]: E1207 16:05:24.041276 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a46985c7bb3495e5b17894566c44c64ef5121e80a63243c39da22efc260b7c2a\": container with ID starting with a46985c7bb3495e5b17894566c44c64ef5121e80a63243c39da22efc260b7c2a not found: ID does not exist" containerID="a46985c7bb3495e5b17894566c44c64ef5121e80a63243c39da22efc260b7c2a" Dec 07 16:05:24 crc kubenswrapper[4716]: I1207 16:05:24.041301 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a46985c7bb3495e5b17894566c44c64ef5121e80a63243c39da22efc260b7c2a"} err="failed to get container status \"a46985c7bb3495e5b17894566c44c64ef5121e80a63243c39da22efc260b7c2a\": rpc error: code = NotFound desc = could not find container \"a46985c7bb3495e5b17894566c44c64ef5121e80a63243c39da22efc260b7c2a\": container with ID starting with a46985c7bb3495e5b17894566c44c64ef5121e80a63243c39da22efc260b7c2a not found: ID does not exist" Dec 07 16:05:24 crc kubenswrapper[4716]: I1207 16:05:24.041322 4716 scope.go:117] "RemoveContainer" containerID="b7a4b01aad1cec88d9bc8a76fce9875c2af3534767fa8f296c5faff7b25d2438" Dec 07 16:05:24 crc kubenswrapper[4716]: E1207 16:05:24.041774 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b7a4b01aad1cec88d9bc8a76fce9875c2af3534767fa8f296c5faff7b25d2438\": container with ID starting with b7a4b01aad1cec88d9bc8a76fce9875c2af3534767fa8f296c5faff7b25d2438 not found: ID does not exist" containerID="b7a4b01aad1cec88d9bc8a76fce9875c2af3534767fa8f296c5faff7b25d2438" Dec 07 16:05:24 crc kubenswrapper[4716]: I1207 16:05:24.041798 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b7a4b01aad1cec88d9bc8a76fce9875c2af3534767fa8f296c5faff7b25d2438"} err="failed to get container status \"b7a4b01aad1cec88d9bc8a76fce9875c2af3534767fa8f296c5faff7b25d2438\": rpc error: code = NotFound desc = could not find container \"b7a4b01aad1cec88d9bc8a76fce9875c2af3534767fa8f296c5faff7b25d2438\": container with ID starting with b7a4b01aad1cec88d9bc8a76fce9875c2af3534767fa8f296c5faff7b25d2438 not found: ID does not exist" Dec 07 16:05:24 crc kubenswrapper[4716]: I1207 16:05:24.238994 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 07 16:05:24 crc kubenswrapper[4716]: E1207 16:05:24.239206 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a72aa026-8e89-45c7-86e7-8a5063352df7" containerName="extract-content" Dec 07 16:05:24 crc kubenswrapper[4716]: I1207 16:05:24.239218 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="a72aa026-8e89-45c7-86e7-8a5063352df7" containerName="extract-content" Dec 07 16:05:24 crc kubenswrapper[4716]: E1207 16:05:24.239226 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80d3e53e-eb0a-4129-bc62-7b1e445bb06d" containerName="pruner" Dec 07 16:05:24 crc kubenswrapper[4716]: I1207 16:05:24.239232 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="80d3e53e-eb0a-4129-bc62-7b1e445bb06d" containerName="pruner" Dec 07 16:05:24 crc kubenswrapper[4716]: E1207 16:05:24.239243 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a72aa026-8e89-45c7-86e7-8a5063352df7" containerName="registry-server" Dec 07 16:05:24 crc kubenswrapper[4716]: I1207 16:05:24.239249 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="a72aa026-8e89-45c7-86e7-8a5063352df7" containerName="registry-server" Dec 07 16:05:24 crc kubenswrapper[4716]: E1207 16:05:24.239260 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a72aa026-8e89-45c7-86e7-8a5063352df7" containerName="extract-utilities" Dec 07 16:05:24 crc kubenswrapper[4716]: I1207 16:05:24.239265 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="a72aa026-8e89-45c7-86e7-8a5063352df7" containerName="extract-utilities" Dec 07 16:05:24 crc kubenswrapper[4716]: I1207 16:05:24.239352 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="80d3e53e-eb0a-4129-bc62-7b1e445bb06d" containerName="pruner" Dec 07 16:05:24 crc kubenswrapper[4716]: I1207 16:05:24.239364 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="a72aa026-8e89-45c7-86e7-8a5063352df7" containerName="registry-server" Dec 07 16:05:24 crc kubenswrapper[4716]: I1207 16:05:24.239728 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 07 16:05:24 crc kubenswrapper[4716]: I1207 16:05:24.243849 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 07 16:05:24 crc kubenswrapper[4716]: I1207 16:05:24.244194 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 07 16:05:24 crc kubenswrapper[4716]: I1207 16:05:24.253367 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 07 16:05:24 crc kubenswrapper[4716]: I1207 16:05:24.405840 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fde20de3-ff39-4cec-b24a-460c6679d188-kube-api-access\") pod \"installer-9-crc\" (UID: \"fde20de3-ff39-4cec-b24a-460c6679d188\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 07 16:05:24 crc kubenswrapper[4716]: I1207 16:05:24.406189 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/fde20de3-ff39-4cec-b24a-460c6679d188-kubelet-dir\") pod \"installer-9-crc\" (UID: \"fde20de3-ff39-4cec-b24a-460c6679d188\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 07 16:05:24 crc kubenswrapper[4716]: I1207 16:05:24.406276 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/fde20de3-ff39-4cec-b24a-460c6679d188-var-lock\") pod \"installer-9-crc\" (UID: \"fde20de3-ff39-4cec-b24a-460c6679d188\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 07 16:05:24 crc kubenswrapper[4716]: I1207 16:05:24.506971 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/fde20de3-ff39-4cec-b24a-460c6679d188-kubelet-dir\") pod \"installer-9-crc\" (UID: \"fde20de3-ff39-4cec-b24a-460c6679d188\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 07 16:05:24 crc kubenswrapper[4716]: I1207 16:05:24.507023 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/fde20de3-ff39-4cec-b24a-460c6679d188-var-lock\") pod \"installer-9-crc\" (UID: \"fde20de3-ff39-4cec-b24a-460c6679d188\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 07 16:05:24 crc kubenswrapper[4716]: I1207 16:05:24.507059 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fde20de3-ff39-4cec-b24a-460c6679d188-kube-api-access\") pod \"installer-9-crc\" (UID: \"fde20de3-ff39-4cec-b24a-460c6679d188\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 07 16:05:24 crc kubenswrapper[4716]: I1207 16:05:24.507139 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/fde20de3-ff39-4cec-b24a-460c6679d188-kubelet-dir\") pod \"installer-9-crc\" (UID: \"fde20de3-ff39-4cec-b24a-460c6679d188\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 07 16:05:24 crc kubenswrapper[4716]: I1207 16:05:24.507191 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/fde20de3-ff39-4cec-b24a-460c6679d188-var-lock\") pod \"installer-9-crc\" (UID: \"fde20de3-ff39-4cec-b24a-460c6679d188\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 07 16:05:24 crc kubenswrapper[4716]: I1207 16:05:24.526161 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fde20de3-ff39-4cec-b24a-460c6679d188-kube-api-access\") pod \"installer-9-crc\" (UID: \"fde20de3-ff39-4cec-b24a-460c6679d188\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 07 16:05:24 crc kubenswrapper[4716]: I1207 16:05:24.554069 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 07 16:05:25 crc kubenswrapper[4716]: I1207 16:05:25.002869 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 07 16:05:25 crc kubenswrapper[4716]: W1207 16:05:25.012836 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podfde20de3_ff39_4cec_b24a_460c6679d188.slice/crio-ffbb0ab65e48f87027f96bb7a3d856c18e0c91a0fbe563f92672cf14aa4b9185 WatchSource:0}: Error finding container ffbb0ab65e48f87027f96bb7a3d856c18e0c91a0fbe563f92672cf14aa4b9185: Status 404 returned error can't find the container with id ffbb0ab65e48f87027f96bb7a3d856c18e0c91a0fbe563f92672cf14aa4b9185 Dec 07 16:05:25 crc kubenswrapper[4716]: I1207 16:05:25.666468 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a72aa026-8e89-45c7-86e7-8a5063352df7" path="/var/lib/kubelet/pods/a72aa026-8e89-45c7-86e7-8a5063352df7/volumes" Dec 07 16:05:25 crc kubenswrapper[4716]: I1207 16:05:25.966992 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"fde20de3-ff39-4cec-b24a-460c6679d188","Type":"ContainerStarted","Data":"d22f07736545a94436104219f2ffe28a7151b2984a8704a3c48be83f55c42499"} Dec 07 16:05:25 crc kubenswrapper[4716]: I1207 16:05:25.967032 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"fde20de3-ff39-4cec-b24a-460c6679d188","Type":"ContainerStarted","Data":"ffbb0ab65e48f87027f96bb7a3d856c18e0c91a0fbe563f92672cf14aa4b9185"} Dec 07 16:05:25 crc kubenswrapper[4716]: I1207 16:05:25.968638 4716 generic.go:334] "Generic (PLEG): container finished" podID="12e72f4d-2887-4c37-bf16-403f124d6f68" containerID="f3f5034db9a73c532d54a8c3ce4db0a417a5c1d66d2b5480d7a35a3417011ddc" exitCode=0 Dec 07 16:05:25 crc kubenswrapper[4716]: I1207 16:05:25.968663 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fnkww" event={"ID":"12e72f4d-2887-4c37-bf16-403f124d6f68","Type":"ContainerDied","Data":"f3f5034db9a73c532d54a8c3ce4db0a417a5c1d66d2b5480d7a35a3417011ddc"} Dec 07 16:05:26 crc kubenswrapper[4716]: I1207 16:05:26.006416 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=2.006400291 podStartE2EDuration="2.006400291s" podCreationTimestamp="2025-12-07 16:05:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:05:25.983891012 +0000 UTC m=+188.674175924" watchObservedRunningTime="2025-12-07 16:05:26.006400291 +0000 UTC m=+188.696685193" Dec 07 16:05:31 crc kubenswrapper[4716]: I1207 16:05:31.446305 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-hwbw9" Dec 07 16:05:31 crc kubenswrapper[4716]: I1207 16:05:31.810784 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-hwbw9" Dec 07 16:05:33 crc kubenswrapper[4716]: I1207 16:05:33.504409 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hwbw9"] Dec 07 16:05:33 crc kubenswrapper[4716]: I1207 16:05:33.504813 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-hwbw9" podUID="b9ab4b15-9882-4125-ba5e-1f5f236c5b6b" containerName="registry-server" containerID="cri-o://f018e88f4a0d9680a2947bc0b33060f9051d838316ed125f7c151e897d2f099e" gracePeriod=2 Dec 07 16:05:33 crc kubenswrapper[4716]: I1207 16:05:33.844365 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hwbw9" Dec 07 16:05:33 crc kubenswrapper[4716]: I1207 16:05:33.949028 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9ab4b15-9882-4125-ba5e-1f5f236c5b6b-utilities\") pod \"b9ab4b15-9882-4125-ba5e-1f5f236c5b6b\" (UID: \"b9ab4b15-9882-4125-ba5e-1f5f236c5b6b\") " Dec 07 16:05:33 crc kubenswrapper[4716]: I1207 16:05:33.949094 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7gd98\" (UniqueName: \"kubernetes.io/projected/b9ab4b15-9882-4125-ba5e-1f5f236c5b6b-kube-api-access-7gd98\") pod \"b9ab4b15-9882-4125-ba5e-1f5f236c5b6b\" (UID: \"b9ab4b15-9882-4125-ba5e-1f5f236c5b6b\") " Dec 07 16:05:33 crc kubenswrapper[4716]: I1207 16:05:33.949237 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9ab4b15-9882-4125-ba5e-1f5f236c5b6b-catalog-content\") pod \"b9ab4b15-9882-4125-ba5e-1f5f236c5b6b\" (UID: \"b9ab4b15-9882-4125-ba5e-1f5f236c5b6b\") " Dec 07 16:05:33 crc kubenswrapper[4716]: I1207 16:05:33.950511 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b9ab4b15-9882-4125-ba5e-1f5f236c5b6b-utilities" (OuterVolumeSpecName: "utilities") pod "b9ab4b15-9882-4125-ba5e-1f5f236c5b6b" (UID: "b9ab4b15-9882-4125-ba5e-1f5f236c5b6b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:05:33 crc kubenswrapper[4716]: I1207 16:05:33.955040 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9ab4b15-9882-4125-ba5e-1f5f236c5b6b-kube-api-access-7gd98" (OuterVolumeSpecName: "kube-api-access-7gd98") pod "b9ab4b15-9882-4125-ba5e-1f5f236c5b6b" (UID: "b9ab4b15-9882-4125-ba5e-1f5f236c5b6b"). InnerVolumeSpecName "kube-api-access-7gd98". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:05:34 crc kubenswrapper[4716]: I1207 16:05:34.021143 4716 generic.go:334] "Generic (PLEG): container finished" podID="b9ab4b15-9882-4125-ba5e-1f5f236c5b6b" containerID="f018e88f4a0d9680a2947bc0b33060f9051d838316ed125f7c151e897d2f099e" exitCode=0 Dec 07 16:05:34 crc kubenswrapper[4716]: I1207 16:05:34.021220 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hwbw9" event={"ID":"b9ab4b15-9882-4125-ba5e-1f5f236c5b6b","Type":"ContainerDied","Data":"f018e88f4a0d9680a2947bc0b33060f9051d838316ed125f7c151e897d2f099e"} Dec 07 16:05:34 crc kubenswrapper[4716]: I1207 16:05:34.021260 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hwbw9" event={"ID":"b9ab4b15-9882-4125-ba5e-1f5f236c5b6b","Type":"ContainerDied","Data":"e2b115a91085b70169015f86c8e98d4db5a5c692bd4ed714ae74d67af96fd42a"} Dec 07 16:05:34 crc kubenswrapper[4716]: I1207 16:05:34.021288 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hwbw9" Dec 07 16:05:34 crc kubenswrapper[4716]: I1207 16:05:34.021291 4716 scope.go:117] "RemoveContainer" containerID="f018e88f4a0d9680a2947bc0b33060f9051d838316ed125f7c151e897d2f099e" Dec 07 16:05:34 crc kubenswrapper[4716]: I1207 16:05:34.024519 4716 generic.go:334] "Generic (PLEG): container finished" podID="de1318df-206b-4ad8-98c7-b9c943560e44" containerID="298cd5090fc2135ec4ce34f00d1298288f512f065677a3aa10997a2d18422682" exitCode=0 Dec 07 16:05:34 crc kubenswrapper[4716]: I1207 16:05:34.024570 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-chrks" event={"ID":"de1318df-206b-4ad8-98c7-b9c943560e44","Type":"ContainerDied","Data":"298cd5090fc2135ec4ce34f00d1298288f512f065677a3aa10997a2d18422682"} Dec 07 16:05:34 crc kubenswrapper[4716]: I1207 16:05:34.046858 4716 generic.go:334] "Generic (PLEG): container finished" podID="45c781c9-df95-4b6e-bed3-fc1f22f079a3" containerID="6470b74cdbdd34df5e88c457a42475515c1a03d0cd8e7303e5dee766cc8686b1" exitCode=0 Dec 07 16:05:34 crc kubenswrapper[4716]: I1207 16:05:34.046999 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cd798" event={"ID":"45c781c9-df95-4b6e-bed3-fc1f22f079a3","Type":"ContainerDied","Data":"6470b74cdbdd34df5e88c457a42475515c1a03d0cd8e7303e5dee766cc8686b1"} Dec 07 16:05:34 crc kubenswrapper[4716]: I1207 16:05:34.052462 4716 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9ab4b15-9882-4125-ba5e-1f5f236c5b6b-utilities\") on node \"crc\" DevicePath \"\"" Dec 07 16:05:34 crc kubenswrapper[4716]: I1207 16:05:34.052524 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7gd98\" (UniqueName: \"kubernetes.io/projected/b9ab4b15-9882-4125-ba5e-1f5f236c5b6b-kube-api-access-7gd98\") on node \"crc\" DevicePath \"\"" Dec 07 16:05:34 crc kubenswrapper[4716]: I1207 16:05:34.060864 4716 generic.go:334] "Generic (PLEG): container finished" podID="96a76712-c754-4978-902a-fa07f6c3c634" containerID="ce336e531e9e49f44be78c018233f05576493bc2003fdf832390d242d3ce2f04" exitCode=0 Dec 07 16:05:34 crc kubenswrapper[4716]: I1207 16:05:34.060944 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6wb5h" event={"ID":"96a76712-c754-4978-902a-fa07f6c3c634","Type":"ContainerDied","Data":"ce336e531e9e49f44be78c018233f05576493bc2003fdf832390d242d3ce2f04"} Dec 07 16:05:34 crc kubenswrapper[4716]: I1207 16:05:34.074650 4716 scope.go:117] "RemoveContainer" containerID="b7428c7ac94508590984bc89225dd9ca188a085173594dfec1dd0e3746a90623" Dec 07 16:05:34 crc kubenswrapper[4716]: I1207 16:05:34.079023 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fnkww" event={"ID":"12e72f4d-2887-4c37-bf16-403f124d6f68","Type":"ContainerStarted","Data":"68b80714dde88883aa18e745470eb64be164168a72184a16f75fabfce0c638db"} Dec 07 16:05:34 crc kubenswrapper[4716]: I1207 16:05:34.082049 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d92nn" event={"ID":"074b6ff7-7e20-4054-811d-34050c07e74e","Type":"ContainerStarted","Data":"f1df7cd6cfed096919086b1925bf53a4f6988a5e169ca593ae033b764f245707"} Dec 07 16:05:34 crc kubenswrapper[4716]: I1207 16:05:34.084788 4716 generic.go:334] "Generic (PLEG): container finished" podID="86df58f0-8580-4c66-9667-4f7cc285f3a0" containerID="3275604dc42563d85c354eea30930dacaef9cd157de6a6344d1812a419e8ca6f" exitCode=0 Dec 07 16:05:34 crc kubenswrapper[4716]: I1207 16:05:34.084851 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hqtd7" event={"ID":"86df58f0-8580-4c66-9667-4f7cc285f3a0","Type":"ContainerDied","Data":"3275604dc42563d85c354eea30930dacaef9cd157de6a6344d1812a419e8ca6f"} Dec 07 16:05:34 crc kubenswrapper[4716]: I1207 16:05:34.087394 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b9ab4b15-9882-4125-ba5e-1f5f236c5b6b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b9ab4b15-9882-4125-ba5e-1f5f236c5b6b" (UID: "b9ab4b15-9882-4125-ba5e-1f5f236c5b6b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:05:34 crc kubenswrapper[4716]: I1207 16:05:34.119635 4716 scope.go:117] "RemoveContainer" containerID="58df8766dbfacafd471976e4db1de732a7003e3d811ab9b06c481e1e6ed6f4e9" Dec 07 16:05:34 crc kubenswrapper[4716]: I1207 16:05:34.144861 4716 scope.go:117] "RemoveContainer" containerID="f018e88f4a0d9680a2947bc0b33060f9051d838316ed125f7c151e897d2f099e" Dec 07 16:05:34 crc kubenswrapper[4716]: E1207 16:05:34.145394 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f018e88f4a0d9680a2947bc0b33060f9051d838316ed125f7c151e897d2f099e\": container with ID starting with f018e88f4a0d9680a2947bc0b33060f9051d838316ed125f7c151e897d2f099e not found: ID does not exist" containerID="f018e88f4a0d9680a2947bc0b33060f9051d838316ed125f7c151e897d2f099e" Dec 07 16:05:34 crc kubenswrapper[4716]: I1207 16:05:34.145607 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f018e88f4a0d9680a2947bc0b33060f9051d838316ed125f7c151e897d2f099e"} err="failed to get container status \"f018e88f4a0d9680a2947bc0b33060f9051d838316ed125f7c151e897d2f099e\": rpc error: code = NotFound desc = could not find container \"f018e88f4a0d9680a2947bc0b33060f9051d838316ed125f7c151e897d2f099e\": container with ID starting with f018e88f4a0d9680a2947bc0b33060f9051d838316ed125f7c151e897d2f099e not found: ID does not exist" Dec 07 16:05:34 crc kubenswrapper[4716]: I1207 16:05:34.145635 4716 scope.go:117] "RemoveContainer" containerID="b7428c7ac94508590984bc89225dd9ca188a085173594dfec1dd0e3746a90623" Dec 07 16:05:34 crc kubenswrapper[4716]: E1207 16:05:34.145965 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b7428c7ac94508590984bc89225dd9ca188a085173594dfec1dd0e3746a90623\": container with ID starting with b7428c7ac94508590984bc89225dd9ca188a085173594dfec1dd0e3746a90623 not found: ID does not exist" containerID="b7428c7ac94508590984bc89225dd9ca188a085173594dfec1dd0e3746a90623" Dec 07 16:05:34 crc kubenswrapper[4716]: I1207 16:05:34.146008 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b7428c7ac94508590984bc89225dd9ca188a085173594dfec1dd0e3746a90623"} err="failed to get container status \"b7428c7ac94508590984bc89225dd9ca188a085173594dfec1dd0e3746a90623\": rpc error: code = NotFound desc = could not find container \"b7428c7ac94508590984bc89225dd9ca188a085173594dfec1dd0e3746a90623\": container with ID starting with b7428c7ac94508590984bc89225dd9ca188a085173594dfec1dd0e3746a90623 not found: ID does not exist" Dec 07 16:05:34 crc kubenswrapper[4716]: I1207 16:05:34.146035 4716 scope.go:117] "RemoveContainer" containerID="58df8766dbfacafd471976e4db1de732a7003e3d811ab9b06c481e1e6ed6f4e9" Dec 07 16:05:34 crc kubenswrapper[4716]: E1207 16:05:34.146646 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58df8766dbfacafd471976e4db1de732a7003e3d811ab9b06c481e1e6ed6f4e9\": container with ID starting with 58df8766dbfacafd471976e4db1de732a7003e3d811ab9b06c481e1e6ed6f4e9 not found: ID does not exist" containerID="58df8766dbfacafd471976e4db1de732a7003e3d811ab9b06c481e1e6ed6f4e9" Dec 07 16:05:34 crc kubenswrapper[4716]: I1207 16:05:34.146896 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58df8766dbfacafd471976e4db1de732a7003e3d811ab9b06c481e1e6ed6f4e9"} err="failed to get container status \"58df8766dbfacafd471976e4db1de732a7003e3d811ab9b06c481e1e6ed6f4e9\": rpc error: code = NotFound desc = could not find container \"58df8766dbfacafd471976e4db1de732a7003e3d811ab9b06c481e1e6ed6f4e9\": container with ID starting with 58df8766dbfacafd471976e4db1de732a7003e3d811ab9b06c481e1e6ed6f4e9 not found: ID does not exist" Dec 07 16:05:34 crc kubenswrapper[4716]: I1207 16:05:34.148261 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-fnkww" podStartSLOduration=2.852778083 podStartE2EDuration="55.148248158s" podCreationTimestamp="2025-12-07 16:04:39 +0000 UTC" firstStartedPulling="2025-12-07 16:04:40.550431661 +0000 UTC m=+143.240716573" lastFinishedPulling="2025-12-07 16:05:32.845901736 +0000 UTC m=+195.536186648" observedRunningTime="2025-12-07 16:05:34.125847083 +0000 UTC m=+196.816131995" watchObservedRunningTime="2025-12-07 16:05:34.148248158 +0000 UTC m=+196.838533060" Dec 07 16:05:34 crc kubenswrapper[4716]: I1207 16:05:34.154093 4716 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9ab4b15-9882-4125-ba5e-1f5f236c5b6b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 07 16:05:34 crc kubenswrapper[4716]: I1207 16:05:34.350605 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hwbw9"] Dec 07 16:05:34 crc kubenswrapper[4716]: I1207 16:05:34.353967 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-hwbw9"] Dec 07 16:05:35 crc kubenswrapper[4716]: I1207 16:05:35.091884 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-chrks" event={"ID":"de1318df-206b-4ad8-98c7-b9c943560e44","Type":"ContainerStarted","Data":"783ff5e0504e8dea0ca2eab1e5e5e0a563316b713aade17bd55fa7318c1dfe97"} Dec 07 16:05:35 crc kubenswrapper[4716]: I1207 16:05:35.093695 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cd798" event={"ID":"45c781c9-df95-4b6e-bed3-fc1f22f079a3","Type":"ContainerStarted","Data":"6b65905e72c05446c18b09994045d7aa88ed1c0439f6a0ae2b7517a3d474faf9"} Dec 07 16:05:35 crc kubenswrapper[4716]: I1207 16:05:35.095317 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6wb5h" event={"ID":"96a76712-c754-4978-902a-fa07f6c3c634","Type":"ContainerStarted","Data":"cd0b7b9f8c0100e26f7e156a9f8a74e588905dac3f00a22748e559d03b2dc613"} Dec 07 16:05:35 crc kubenswrapper[4716]: I1207 16:05:35.096943 4716 generic.go:334] "Generic (PLEG): container finished" podID="074b6ff7-7e20-4054-811d-34050c07e74e" containerID="f1df7cd6cfed096919086b1925bf53a4f6988a5e169ca593ae033b764f245707" exitCode=0 Dec 07 16:05:35 crc kubenswrapper[4716]: I1207 16:05:35.097016 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d92nn" event={"ID":"074b6ff7-7e20-4054-811d-34050c07e74e","Type":"ContainerDied","Data":"f1df7cd6cfed096919086b1925bf53a4f6988a5e169ca593ae033b764f245707"} Dec 07 16:05:35 crc kubenswrapper[4716]: I1207 16:05:35.097052 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d92nn" event={"ID":"074b6ff7-7e20-4054-811d-34050c07e74e","Type":"ContainerStarted","Data":"c0b94112363f26c24a1f79cdb0548597273c2180248bb78e761a44282abcf9c0"} Dec 07 16:05:35 crc kubenswrapper[4716]: I1207 16:05:35.098583 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hqtd7" event={"ID":"86df58f0-8580-4c66-9667-4f7cc285f3a0","Type":"ContainerStarted","Data":"f42afff4ed726d86046eae352ef553422bcf740d40af3eb4cbb9bedea874e41c"} Dec 07 16:05:35 crc kubenswrapper[4716]: I1207 16:05:35.117468 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-chrks" podStartSLOduration=3.115926097 podStartE2EDuration="58.117451289s" podCreationTimestamp="2025-12-07 16:04:37 +0000 UTC" firstStartedPulling="2025-12-07 16:04:39.485538461 +0000 UTC m=+142.175823383" lastFinishedPulling="2025-12-07 16:05:34.487063653 +0000 UTC m=+197.177348575" observedRunningTime="2025-12-07 16:05:35.110112422 +0000 UTC m=+197.800397334" watchObservedRunningTime="2025-12-07 16:05:35.117451289 +0000 UTC m=+197.807736201" Dec 07 16:05:35 crc kubenswrapper[4716]: I1207 16:05:35.140803 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-cd798" podStartSLOduration=3.190065649 podStartE2EDuration="58.140781708s" podCreationTimestamp="2025-12-07 16:04:37 +0000 UTC" firstStartedPulling="2025-12-07 16:04:39.497987539 +0000 UTC m=+142.188272471" lastFinishedPulling="2025-12-07 16:05:34.448703618 +0000 UTC m=+197.138988530" observedRunningTime="2025-12-07 16:05:35.137630944 +0000 UTC m=+197.827915856" watchObservedRunningTime="2025-12-07 16:05:35.140781708 +0000 UTC m=+197.831066620" Dec 07 16:05:35 crc kubenswrapper[4716]: I1207 16:05:35.166352 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-d92nn" podStartSLOduration=3.181487594 podStartE2EDuration="58.166333776s" podCreationTimestamp="2025-12-07 16:04:37 +0000 UTC" firstStartedPulling="2025-12-07 16:04:39.513907619 +0000 UTC m=+142.204192551" lastFinishedPulling="2025-12-07 16:05:34.498753821 +0000 UTC m=+197.189038733" observedRunningTime="2025-12-07 16:05:35.161793033 +0000 UTC m=+197.852077955" watchObservedRunningTime="2025-12-07 16:05:35.166333776 +0000 UTC m=+197.856618688" Dec 07 16:05:35 crc kubenswrapper[4716]: I1207 16:05:35.182709 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-6wb5h" podStartSLOduration=2.020335102 podStartE2EDuration="57.182689067s" podCreationTimestamp="2025-12-07 16:04:38 +0000 UTC" firstStartedPulling="2025-12-07 16:04:39.501920373 +0000 UTC m=+142.192205285" lastFinishedPulling="2025-12-07 16:05:34.664274318 +0000 UTC m=+197.354559250" observedRunningTime="2025-12-07 16:05:35.179342886 +0000 UTC m=+197.869627798" watchObservedRunningTime="2025-12-07 16:05:35.182689067 +0000 UTC m=+197.872973979" Dec 07 16:05:35 crc kubenswrapper[4716]: I1207 16:05:35.205763 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-hqtd7" podStartSLOduration=3.238002014 podStartE2EDuration="55.205742607s" podCreationTimestamp="2025-12-07 16:04:40 +0000 UTC" firstStartedPulling="2025-12-07 16:04:42.623281451 +0000 UTC m=+145.313566363" lastFinishedPulling="2025-12-07 16:05:34.591022044 +0000 UTC m=+197.281306956" observedRunningTime="2025-12-07 16:05:35.203167458 +0000 UTC m=+197.893452370" watchObservedRunningTime="2025-12-07 16:05:35.205742607 +0000 UTC m=+197.896027519" Dec 07 16:05:35 crc kubenswrapper[4716]: I1207 16:05:35.663918 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b9ab4b15-9882-4125-ba5e-1f5f236c5b6b" path="/var/lib/kubelet/pods/b9ab4b15-9882-4125-ba5e-1f5f236c5b6b/volumes" Dec 07 16:05:37 crc kubenswrapper[4716]: I1207 16:05:37.843154 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-cd798" Dec 07 16:05:37 crc kubenswrapper[4716]: I1207 16:05:37.843495 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-cd798" Dec 07 16:05:37 crc kubenswrapper[4716]: I1207 16:05:37.881995 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-cd798" Dec 07 16:05:38 crc kubenswrapper[4716]: I1207 16:05:38.014512 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-d92nn" Dec 07 16:05:38 crc kubenswrapper[4716]: I1207 16:05:38.014574 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-d92nn" Dec 07 16:05:38 crc kubenswrapper[4716]: I1207 16:05:38.058770 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-d92nn" Dec 07 16:05:38 crc kubenswrapper[4716]: I1207 16:05:38.207871 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-chrks" Dec 07 16:05:38 crc kubenswrapper[4716]: I1207 16:05:38.207931 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-chrks" Dec 07 16:05:38 crc kubenswrapper[4716]: I1207 16:05:38.244837 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-chrks" Dec 07 16:05:38 crc kubenswrapper[4716]: I1207 16:05:38.401229 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-6wb5h" Dec 07 16:05:38 crc kubenswrapper[4716]: I1207 16:05:38.401273 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-6wb5h" Dec 07 16:05:38 crc kubenswrapper[4716]: I1207 16:05:38.436133 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-6wb5h" Dec 07 16:05:39 crc kubenswrapper[4716]: I1207 16:05:39.153880 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-chrks" Dec 07 16:05:39 crc kubenswrapper[4716]: I1207 16:05:39.163346 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-6wb5h" Dec 07 16:05:39 crc kubenswrapper[4716]: I1207 16:05:39.782835 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-fnkww" Dec 07 16:05:39 crc kubenswrapper[4716]: I1207 16:05:39.782955 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-fnkww" Dec 07 16:05:39 crc kubenswrapper[4716]: I1207 16:05:39.842382 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-fnkww" Dec 07 16:05:40 crc kubenswrapper[4716]: I1207 16:05:40.179918 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-fnkww" Dec 07 16:05:40 crc kubenswrapper[4716]: I1207 16:05:40.305628 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-chrks"] Dec 07 16:05:41 crc kubenswrapper[4716]: I1207 16:05:41.007795 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-hqtd7" Dec 07 16:05:41 crc kubenswrapper[4716]: I1207 16:05:41.007848 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-hqtd7" Dec 07 16:05:41 crc kubenswrapper[4716]: I1207 16:05:41.046210 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-hqtd7" Dec 07 16:05:41 crc kubenswrapper[4716]: I1207 16:05:41.127895 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-chrks" podUID="de1318df-206b-4ad8-98c7-b9c943560e44" containerName="registry-server" containerID="cri-o://783ff5e0504e8dea0ca2eab1e5e5e0a563316b713aade17bd55fa7318c1dfe97" gracePeriod=2 Dec 07 16:05:41 crc kubenswrapper[4716]: I1207 16:05:41.173995 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-hqtd7" Dec 07 16:05:41 crc kubenswrapper[4716]: I1207 16:05:41.305681 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6wb5h"] Dec 07 16:05:41 crc kubenswrapper[4716]: I1207 16:05:41.305888 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-6wb5h" podUID="96a76712-c754-4978-902a-fa07f6c3c634" containerName="registry-server" containerID="cri-o://cd0b7b9f8c0100e26f7e156a9f8a74e588905dac3f00a22748e559d03b2dc613" gracePeriod=2 Dec 07 16:05:43 crc kubenswrapper[4716]: I1207 16:05:43.140395 4716 generic.go:334] "Generic (PLEG): container finished" podID="de1318df-206b-4ad8-98c7-b9c943560e44" containerID="783ff5e0504e8dea0ca2eab1e5e5e0a563316b713aade17bd55fa7318c1dfe97" exitCode=0 Dec 07 16:05:43 crc kubenswrapper[4716]: I1207 16:05:43.140666 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-chrks" event={"ID":"de1318df-206b-4ad8-98c7-b9c943560e44","Type":"ContainerDied","Data":"783ff5e0504e8dea0ca2eab1e5e5e0a563316b713aade17bd55fa7318c1dfe97"} Dec 07 16:05:43 crc kubenswrapper[4716]: I1207 16:05:43.142390 4716 generic.go:334] "Generic (PLEG): container finished" podID="96a76712-c754-4978-902a-fa07f6c3c634" containerID="cd0b7b9f8c0100e26f7e156a9f8a74e588905dac3f00a22748e559d03b2dc613" exitCode=0 Dec 07 16:05:43 crc kubenswrapper[4716]: I1207 16:05:43.142410 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6wb5h" event={"ID":"96a76712-c754-4978-902a-fa07f6c3c634","Type":"ContainerDied","Data":"cd0b7b9f8c0100e26f7e156a9f8a74e588905dac3f00a22748e559d03b2dc613"} Dec 07 16:05:43 crc kubenswrapper[4716]: I1207 16:05:43.170163 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6wb5h" Dec 07 16:05:43 crc kubenswrapper[4716]: I1207 16:05:43.287670 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96a76712-c754-4978-902a-fa07f6c3c634-catalog-content\") pod \"96a76712-c754-4978-902a-fa07f6c3c634\" (UID: \"96a76712-c754-4978-902a-fa07f6c3c634\") " Dec 07 16:05:43 crc kubenswrapper[4716]: I1207 16:05:43.288052 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96a76712-c754-4978-902a-fa07f6c3c634-utilities\") pod \"96a76712-c754-4978-902a-fa07f6c3c634\" (UID: \"96a76712-c754-4978-902a-fa07f6c3c634\") " Dec 07 16:05:43 crc kubenswrapper[4716]: I1207 16:05:43.288174 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9lbc5\" (UniqueName: \"kubernetes.io/projected/96a76712-c754-4978-902a-fa07f6c3c634-kube-api-access-9lbc5\") pod \"96a76712-c754-4978-902a-fa07f6c3c634\" (UID: \"96a76712-c754-4978-902a-fa07f6c3c634\") " Dec 07 16:05:43 crc kubenswrapper[4716]: I1207 16:05:43.289775 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/96a76712-c754-4978-902a-fa07f6c3c634-utilities" (OuterVolumeSpecName: "utilities") pod "96a76712-c754-4978-902a-fa07f6c3c634" (UID: "96a76712-c754-4978-902a-fa07f6c3c634"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:05:43 crc kubenswrapper[4716]: I1207 16:05:43.293904 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96a76712-c754-4978-902a-fa07f6c3c634-kube-api-access-9lbc5" (OuterVolumeSpecName: "kube-api-access-9lbc5") pod "96a76712-c754-4978-902a-fa07f6c3c634" (UID: "96a76712-c754-4978-902a-fa07f6c3c634"). InnerVolumeSpecName "kube-api-access-9lbc5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:05:43 crc kubenswrapper[4716]: I1207 16:05:43.300669 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-chrks" Dec 07 16:05:43 crc kubenswrapper[4716]: I1207 16:05:43.343637 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/96a76712-c754-4978-902a-fa07f6c3c634-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "96a76712-c754-4978-902a-fa07f6c3c634" (UID: "96a76712-c754-4978-902a-fa07f6c3c634"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:05:43 crc kubenswrapper[4716]: I1207 16:05:43.391056 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9tbdw\" (UniqueName: \"kubernetes.io/projected/de1318df-206b-4ad8-98c7-b9c943560e44-kube-api-access-9tbdw\") pod \"de1318df-206b-4ad8-98c7-b9c943560e44\" (UID: \"de1318df-206b-4ad8-98c7-b9c943560e44\") " Dec 07 16:05:43 crc kubenswrapper[4716]: I1207 16:05:43.391118 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de1318df-206b-4ad8-98c7-b9c943560e44-catalog-content\") pod \"de1318df-206b-4ad8-98c7-b9c943560e44\" (UID: \"de1318df-206b-4ad8-98c7-b9c943560e44\") " Dec 07 16:05:43 crc kubenswrapper[4716]: I1207 16:05:43.391155 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de1318df-206b-4ad8-98c7-b9c943560e44-utilities\") pod \"de1318df-206b-4ad8-98c7-b9c943560e44\" (UID: \"de1318df-206b-4ad8-98c7-b9c943560e44\") " Dec 07 16:05:43 crc kubenswrapper[4716]: I1207 16:05:43.391425 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9lbc5\" (UniqueName: \"kubernetes.io/projected/96a76712-c754-4978-902a-fa07f6c3c634-kube-api-access-9lbc5\") on node \"crc\" DevicePath \"\"" Dec 07 16:05:43 crc kubenswrapper[4716]: I1207 16:05:43.391442 4716 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96a76712-c754-4978-902a-fa07f6c3c634-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 07 16:05:43 crc kubenswrapper[4716]: I1207 16:05:43.391453 4716 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96a76712-c754-4978-902a-fa07f6c3c634-utilities\") on node \"crc\" DevicePath \"\"" Dec 07 16:05:43 crc kubenswrapper[4716]: I1207 16:05:43.391999 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de1318df-206b-4ad8-98c7-b9c943560e44-utilities" (OuterVolumeSpecName: "utilities") pod "de1318df-206b-4ad8-98c7-b9c943560e44" (UID: "de1318df-206b-4ad8-98c7-b9c943560e44"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:05:43 crc kubenswrapper[4716]: I1207 16:05:43.393574 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de1318df-206b-4ad8-98c7-b9c943560e44-kube-api-access-9tbdw" (OuterVolumeSpecName: "kube-api-access-9tbdw") pod "de1318df-206b-4ad8-98c7-b9c943560e44" (UID: "de1318df-206b-4ad8-98c7-b9c943560e44"). InnerVolumeSpecName "kube-api-access-9tbdw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:05:43 crc kubenswrapper[4716]: I1207 16:05:43.444857 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de1318df-206b-4ad8-98c7-b9c943560e44-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "de1318df-206b-4ad8-98c7-b9c943560e44" (UID: "de1318df-206b-4ad8-98c7-b9c943560e44"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:05:43 crc kubenswrapper[4716]: I1207 16:05:43.492393 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9tbdw\" (UniqueName: \"kubernetes.io/projected/de1318df-206b-4ad8-98c7-b9c943560e44-kube-api-access-9tbdw\") on node \"crc\" DevicePath \"\"" Dec 07 16:05:43 crc kubenswrapper[4716]: I1207 16:05:43.492418 4716 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de1318df-206b-4ad8-98c7-b9c943560e44-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 07 16:05:43 crc kubenswrapper[4716]: I1207 16:05:43.492429 4716 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de1318df-206b-4ad8-98c7-b9c943560e44-utilities\") on node \"crc\" DevicePath \"\"" Dec 07 16:05:44 crc kubenswrapper[4716]: I1207 16:05:44.150404 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-chrks" event={"ID":"de1318df-206b-4ad8-98c7-b9c943560e44","Type":"ContainerDied","Data":"34c87a6f423d4af364ae035ff900e68134f28aa0fac938106f075adac4214216"} Dec 07 16:05:44 crc kubenswrapper[4716]: I1207 16:05:44.150450 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-chrks" Dec 07 16:05:44 crc kubenswrapper[4716]: I1207 16:05:44.150492 4716 scope.go:117] "RemoveContainer" containerID="783ff5e0504e8dea0ca2eab1e5e5e0a563316b713aade17bd55fa7318c1dfe97" Dec 07 16:05:44 crc kubenswrapper[4716]: I1207 16:05:44.154877 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6wb5h" event={"ID":"96a76712-c754-4978-902a-fa07f6c3c634","Type":"ContainerDied","Data":"5d68160f8fb287da1d047a59a81beaf4513bf3a9451850aada38fb9c5fc6d6d2"} Dec 07 16:05:44 crc kubenswrapper[4716]: I1207 16:05:44.154961 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6wb5h" Dec 07 16:05:44 crc kubenswrapper[4716]: I1207 16:05:44.177961 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-chrks"] Dec 07 16:05:44 crc kubenswrapper[4716]: I1207 16:05:44.181993 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-chrks"] Dec 07 16:05:44 crc kubenswrapper[4716]: I1207 16:05:44.184142 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6wb5h"] Dec 07 16:05:44 crc kubenswrapper[4716]: I1207 16:05:44.186798 4716 scope.go:117] "RemoveContainer" containerID="298cd5090fc2135ec4ce34f00d1298288f512f065677a3aa10997a2d18422682" Dec 07 16:05:44 crc kubenswrapper[4716]: I1207 16:05:44.186837 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-6wb5h"] Dec 07 16:05:44 crc kubenswrapper[4716]: I1207 16:05:44.208416 4716 scope.go:117] "RemoveContainer" containerID="1fe9014ba1b4e5f8f3b0c1d316017607984f88799209849528531f460b024af7" Dec 07 16:05:44 crc kubenswrapper[4716]: I1207 16:05:44.232751 4716 scope.go:117] "RemoveContainer" containerID="cd0b7b9f8c0100e26f7e156a9f8a74e588905dac3f00a22748e559d03b2dc613" Dec 07 16:05:44 crc kubenswrapper[4716]: I1207 16:05:44.245888 4716 scope.go:117] "RemoveContainer" containerID="ce336e531e9e49f44be78c018233f05576493bc2003fdf832390d242d3ce2f04" Dec 07 16:05:44 crc kubenswrapper[4716]: I1207 16:05:44.273364 4716 scope.go:117] "RemoveContainer" containerID="d772d9a9b3f4a67928c244a2bc14bd2164e8db4fe0f4365713849cebdfa09b53" Dec 07 16:05:45 crc kubenswrapper[4716]: I1207 16:05:45.670418 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96a76712-c754-4978-902a-fa07f6c3c634" path="/var/lib/kubelet/pods/96a76712-c754-4978-902a-fa07f6c3c634/volumes" Dec 07 16:05:45 crc kubenswrapper[4716]: I1207 16:05:45.671814 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de1318df-206b-4ad8-98c7-b9c943560e44" path="/var/lib/kubelet/pods/de1318df-206b-4ad8-98c7-b9c943560e44/volumes" Dec 07 16:05:47 crc kubenswrapper[4716]: I1207 16:05:47.885209 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-cd798" Dec 07 16:05:48 crc kubenswrapper[4716]: I1207 16:05:48.058671 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-d92nn" Dec 07 16:05:50 crc kubenswrapper[4716]: I1207 16:05:50.712839 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-z587m"] Dec 07 16:05:52 crc kubenswrapper[4716]: I1207 16:05:52.760889 4716 patch_prober.go:28] interesting pod/machine-config-daemon-zcxxp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 16:05:52 crc kubenswrapper[4716]: I1207 16:05:52.761250 4716 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 16:05:52 crc kubenswrapper[4716]: I1207 16:05:52.761296 4716 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" Dec 07 16:05:52 crc kubenswrapper[4716]: I1207 16:05:52.761994 4716 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8ccaf6c1e6a978eb978a1c207b6c4c00b23737f78b743f966579aca781e22e0b"} pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 07 16:05:52 crc kubenswrapper[4716]: I1207 16:05:52.762062 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" containerName="machine-config-daemon" containerID="cri-o://8ccaf6c1e6a978eb978a1c207b6c4c00b23737f78b743f966579aca781e22e0b" gracePeriod=600 Dec 07 16:05:53 crc kubenswrapper[4716]: I1207 16:05:53.210861 4716 generic.go:334] "Generic (PLEG): container finished" podID="c15b59eb-565d-4556-a4ce-75afdf159dc8" containerID="8ccaf6c1e6a978eb978a1c207b6c4c00b23737f78b743f966579aca781e22e0b" exitCode=0 Dec 07 16:05:53 crc kubenswrapper[4716]: I1207 16:05:53.210987 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" event={"ID":"c15b59eb-565d-4556-a4ce-75afdf159dc8","Type":"ContainerDied","Data":"8ccaf6c1e6a978eb978a1c207b6c4c00b23737f78b743f966579aca781e22e0b"} Dec 07 16:05:53 crc kubenswrapper[4716]: I1207 16:05:53.211216 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" event={"ID":"c15b59eb-565d-4556-a4ce-75afdf159dc8","Type":"ContainerStarted","Data":"ca669356fa57e43a7c10e3a5b76ace008fedad9bdfdd4c2c358a8731bdaaf42b"} Dec 07 16:06:03 crc kubenswrapper[4716]: I1207 16:06:03.006953 4716 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 07 16:06:03 crc kubenswrapper[4716]: E1207 16:06:03.007689 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96a76712-c754-4978-902a-fa07f6c3c634" containerName="extract-content" Dec 07 16:06:03 crc kubenswrapper[4716]: I1207 16:06:03.007702 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="96a76712-c754-4978-902a-fa07f6c3c634" containerName="extract-content" Dec 07 16:06:03 crc kubenswrapper[4716]: E1207 16:06:03.007713 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9ab4b15-9882-4125-ba5e-1f5f236c5b6b" containerName="registry-server" Dec 07 16:06:03 crc kubenswrapper[4716]: I1207 16:06:03.007719 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9ab4b15-9882-4125-ba5e-1f5f236c5b6b" containerName="registry-server" Dec 07 16:06:03 crc kubenswrapper[4716]: E1207 16:06:03.007726 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de1318df-206b-4ad8-98c7-b9c943560e44" containerName="registry-server" Dec 07 16:06:03 crc kubenswrapper[4716]: I1207 16:06:03.007733 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="de1318df-206b-4ad8-98c7-b9c943560e44" containerName="registry-server" Dec 07 16:06:03 crc kubenswrapper[4716]: E1207 16:06:03.007741 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9ab4b15-9882-4125-ba5e-1f5f236c5b6b" containerName="extract-content" Dec 07 16:06:03 crc kubenswrapper[4716]: I1207 16:06:03.007747 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9ab4b15-9882-4125-ba5e-1f5f236c5b6b" containerName="extract-content" Dec 07 16:06:03 crc kubenswrapper[4716]: E1207 16:06:03.007758 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de1318df-206b-4ad8-98c7-b9c943560e44" containerName="extract-utilities" Dec 07 16:06:03 crc kubenswrapper[4716]: I1207 16:06:03.007763 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="de1318df-206b-4ad8-98c7-b9c943560e44" containerName="extract-utilities" Dec 07 16:06:03 crc kubenswrapper[4716]: E1207 16:06:03.007773 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96a76712-c754-4978-902a-fa07f6c3c634" containerName="registry-server" Dec 07 16:06:03 crc kubenswrapper[4716]: I1207 16:06:03.007779 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="96a76712-c754-4978-902a-fa07f6c3c634" containerName="registry-server" Dec 07 16:06:03 crc kubenswrapper[4716]: E1207 16:06:03.007789 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de1318df-206b-4ad8-98c7-b9c943560e44" containerName="extract-content" Dec 07 16:06:03 crc kubenswrapper[4716]: I1207 16:06:03.007794 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="de1318df-206b-4ad8-98c7-b9c943560e44" containerName="extract-content" Dec 07 16:06:03 crc kubenswrapper[4716]: E1207 16:06:03.007805 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96a76712-c754-4978-902a-fa07f6c3c634" containerName="extract-utilities" Dec 07 16:06:03 crc kubenswrapper[4716]: I1207 16:06:03.007811 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="96a76712-c754-4978-902a-fa07f6c3c634" containerName="extract-utilities" Dec 07 16:06:03 crc kubenswrapper[4716]: E1207 16:06:03.007820 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9ab4b15-9882-4125-ba5e-1f5f236c5b6b" containerName="extract-utilities" Dec 07 16:06:03 crc kubenswrapper[4716]: I1207 16:06:03.007826 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9ab4b15-9882-4125-ba5e-1f5f236c5b6b" containerName="extract-utilities" Dec 07 16:06:03 crc kubenswrapper[4716]: I1207 16:06:03.007911 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="96a76712-c754-4978-902a-fa07f6c3c634" containerName="registry-server" Dec 07 16:06:03 crc kubenswrapper[4716]: I1207 16:06:03.007922 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="de1318df-206b-4ad8-98c7-b9c943560e44" containerName="registry-server" Dec 07 16:06:03 crc kubenswrapper[4716]: I1207 16:06:03.007932 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9ab4b15-9882-4125-ba5e-1f5f236c5b6b" containerName="registry-server" Dec 07 16:06:03 crc kubenswrapper[4716]: I1207 16:06:03.008240 4716 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 07 16:06:03 crc kubenswrapper[4716]: I1207 16:06:03.008419 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 07 16:06:03 crc kubenswrapper[4716]: I1207 16:06:03.008590 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://73f85248b4d68a9beecdacb3a01f433e97a0ffea2247ce92a41ef96855509377" gracePeriod=15 Dec 07 16:06:03 crc kubenswrapper[4716]: I1207 16:06:03.008684 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://27bd4a54f920b2725987c1c7fd2d9fdf16cd08738c246c333a4e9164a3a582a0" gracePeriod=15 Dec 07 16:06:03 crc kubenswrapper[4716]: I1207 16:06:03.008804 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://6276df652efe6ca935e745f24ef20f87a16465475139230c63b9999234a4d12a" gracePeriod=15 Dec 07 16:06:03 crc kubenswrapper[4716]: I1207 16:06:03.008737 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://e9c3b1dd202586ee755ca89fbe12b30bbb12f95a62514f6e4d93cbbdbe55181e" gracePeriod=15 Dec 07 16:06:03 crc kubenswrapper[4716]: I1207 16:06:03.008891 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://dd318647816e4a27849666b0ec4fc2ca4905295661c557dc381e1144f22cf284" gracePeriod=15 Dec 07 16:06:03 crc kubenswrapper[4716]: I1207 16:06:03.011073 4716 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 07 16:06:03 crc kubenswrapper[4716]: E1207 16:06:03.011482 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 07 16:06:03 crc kubenswrapper[4716]: I1207 16:06:03.011506 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 07 16:06:03 crc kubenswrapper[4716]: E1207 16:06:03.011537 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 07 16:06:03 crc kubenswrapper[4716]: I1207 16:06:03.011551 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 07 16:06:03 crc kubenswrapper[4716]: E1207 16:06:03.011573 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 07 16:06:03 crc kubenswrapper[4716]: I1207 16:06:03.011588 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 07 16:06:03 crc kubenswrapper[4716]: E1207 16:06:03.011610 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 07 16:06:03 crc kubenswrapper[4716]: I1207 16:06:03.011624 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 07 16:06:03 crc kubenswrapper[4716]: E1207 16:06:03.011645 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 07 16:06:03 crc kubenswrapper[4716]: I1207 16:06:03.011657 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 07 16:06:03 crc kubenswrapper[4716]: E1207 16:06:03.011672 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 07 16:06:03 crc kubenswrapper[4716]: I1207 16:06:03.011685 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 07 16:06:03 crc kubenswrapper[4716]: I1207 16:06:03.011879 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 07 16:06:03 crc kubenswrapper[4716]: I1207 16:06:03.011898 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 07 16:06:03 crc kubenswrapper[4716]: I1207 16:06:03.011921 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 07 16:06:03 crc kubenswrapper[4716]: I1207 16:06:03.011936 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 07 16:06:03 crc kubenswrapper[4716]: I1207 16:06:03.011957 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 07 16:06:03 crc kubenswrapper[4716]: I1207 16:06:03.011976 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 07 16:06:03 crc kubenswrapper[4716]: E1207 16:06:03.012173 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 07 16:06:03 crc kubenswrapper[4716]: I1207 16:06:03.012187 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 07 16:06:03 crc kubenswrapper[4716]: I1207 16:06:03.042468 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 07 16:06:03 crc kubenswrapper[4716]: I1207 16:06:03.042981 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 07 16:06:03 crc kubenswrapper[4716]: I1207 16:06:03.043099 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 16:06:03 crc kubenswrapper[4716]: I1207 16:06:03.043235 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 16:06:03 crc kubenswrapper[4716]: I1207 16:06:03.043338 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 16:06:03 crc kubenswrapper[4716]: I1207 16:06:03.043434 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 07 16:06:03 crc kubenswrapper[4716]: I1207 16:06:03.043544 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 07 16:06:03 crc kubenswrapper[4716]: I1207 16:06:03.043644 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 07 16:06:03 crc kubenswrapper[4716]: I1207 16:06:03.145180 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 07 16:06:03 crc kubenswrapper[4716]: I1207 16:06:03.145231 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 07 16:06:03 crc kubenswrapper[4716]: I1207 16:06:03.145269 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 16:06:03 crc kubenswrapper[4716]: I1207 16:06:03.145299 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 16:06:03 crc kubenswrapper[4716]: I1207 16:06:03.145322 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 16:06:03 crc kubenswrapper[4716]: I1207 16:06:03.145348 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 07 16:06:03 crc kubenswrapper[4716]: I1207 16:06:03.145378 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 07 16:06:03 crc kubenswrapper[4716]: I1207 16:06:03.145406 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 07 16:06:03 crc kubenswrapper[4716]: I1207 16:06:03.145429 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 16:06:03 crc kubenswrapper[4716]: I1207 16:06:03.145515 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 07 16:06:03 crc kubenswrapper[4716]: I1207 16:06:03.145538 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 07 16:06:03 crc kubenswrapper[4716]: I1207 16:06:03.145568 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 16:06:03 crc kubenswrapper[4716]: I1207 16:06:03.145596 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 16:06:03 crc kubenswrapper[4716]: I1207 16:06:03.145601 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 07 16:06:03 crc kubenswrapper[4716]: I1207 16:06:03.145623 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 07 16:06:03 crc kubenswrapper[4716]: I1207 16:06:03.145347 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 07 16:06:03 crc kubenswrapper[4716]: I1207 16:06:03.267995 4716 generic.go:334] "Generic (PLEG): container finished" podID="fde20de3-ff39-4cec-b24a-460c6679d188" containerID="d22f07736545a94436104219f2ffe28a7151b2984a8704a3c48be83f55c42499" exitCode=0 Dec 07 16:06:03 crc kubenswrapper[4716]: I1207 16:06:03.268138 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"fde20de3-ff39-4cec-b24a-460c6679d188","Type":"ContainerDied","Data":"d22f07736545a94436104219f2ffe28a7151b2984a8704a3c48be83f55c42499"} Dec 07 16:06:03 crc kubenswrapper[4716]: I1207 16:06:03.269746 4716 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.18:6443: connect: connection refused" Dec 07 16:06:03 crc kubenswrapper[4716]: I1207 16:06:03.270448 4716 status_manager.go:851] "Failed to get status for pod" podUID="fde20de3-ff39-4cec-b24a-460c6679d188" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.18:6443: connect: connection refused" Dec 07 16:06:03 crc kubenswrapper[4716]: I1207 16:06:03.271673 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 07 16:06:03 crc kubenswrapper[4716]: I1207 16:06:03.273925 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 07 16:06:03 crc kubenswrapper[4716]: I1207 16:06:03.275279 4716 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="dd318647816e4a27849666b0ec4fc2ca4905295661c557dc381e1144f22cf284" exitCode=0 Dec 07 16:06:03 crc kubenswrapper[4716]: I1207 16:06:03.275323 4716 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="27bd4a54f920b2725987c1c7fd2d9fdf16cd08738c246c333a4e9164a3a582a0" exitCode=0 Dec 07 16:06:03 crc kubenswrapper[4716]: I1207 16:06:03.275343 4716 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="6276df652efe6ca935e745f24ef20f87a16465475139230c63b9999234a4d12a" exitCode=0 Dec 07 16:06:03 crc kubenswrapper[4716]: I1207 16:06:03.275367 4716 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="e9c3b1dd202586ee755ca89fbe12b30bbb12f95a62514f6e4d93cbbdbe55181e" exitCode=2 Dec 07 16:06:03 crc kubenswrapper[4716]: I1207 16:06:03.275400 4716 scope.go:117] "RemoveContainer" containerID="0d31de0326a05216fa10442bdb47151c6e073c97e9a16c34e793028e5e253756" Dec 07 16:06:03 crc kubenswrapper[4716]: E1207 16:06:03.748172 4716 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 38.102.83.18:6443: connect: connection refused" pod="openshift-image-registry/image-registry-697d97f7c8-4d257" volumeName="registry-storage" Dec 07 16:06:04 crc kubenswrapper[4716]: I1207 16:06:04.282610 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 07 16:06:04 crc kubenswrapper[4716]: I1207 16:06:04.556289 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 07 16:06:04 crc kubenswrapper[4716]: I1207 16:06:04.557103 4716 status_manager.go:851] "Failed to get status for pod" podUID="fde20de3-ff39-4cec-b24a-460c6679d188" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.18:6443: connect: connection refused" Dec 07 16:06:04 crc kubenswrapper[4716]: I1207 16:06:04.563577 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/fde20de3-ff39-4cec-b24a-460c6679d188-kubelet-dir\") pod \"fde20de3-ff39-4cec-b24a-460c6679d188\" (UID: \"fde20de3-ff39-4cec-b24a-460c6679d188\") " Dec 07 16:06:04 crc kubenswrapper[4716]: I1207 16:06:04.563633 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fde20de3-ff39-4cec-b24a-460c6679d188-kube-api-access\") pod \"fde20de3-ff39-4cec-b24a-460c6679d188\" (UID: \"fde20de3-ff39-4cec-b24a-460c6679d188\") " Dec 07 16:06:04 crc kubenswrapper[4716]: I1207 16:06:04.563677 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/fde20de3-ff39-4cec-b24a-460c6679d188-var-lock\") pod \"fde20de3-ff39-4cec-b24a-460c6679d188\" (UID: \"fde20de3-ff39-4cec-b24a-460c6679d188\") " Dec 07 16:06:04 crc kubenswrapper[4716]: I1207 16:06:04.563715 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fde20de3-ff39-4cec-b24a-460c6679d188-var-lock" (OuterVolumeSpecName: "var-lock") pod "fde20de3-ff39-4cec-b24a-460c6679d188" (UID: "fde20de3-ff39-4cec-b24a-460c6679d188"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 16:06:04 crc kubenswrapper[4716]: I1207 16:06:04.563753 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fde20de3-ff39-4cec-b24a-460c6679d188-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "fde20de3-ff39-4cec-b24a-460c6679d188" (UID: "fde20de3-ff39-4cec-b24a-460c6679d188"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 16:06:04 crc kubenswrapper[4716]: I1207 16:06:04.564006 4716 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/fde20de3-ff39-4cec-b24a-460c6679d188-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 07 16:06:04 crc kubenswrapper[4716]: I1207 16:06:04.564029 4716 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/fde20de3-ff39-4cec-b24a-460c6679d188-var-lock\") on node \"crc\" DevicePath \"\"" Dec 07 16:06:04 crc kubenswrapper[4716]: I1207 16:06:04.568480 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fde20de3-ff39-4cec-b24a-460c6679d188-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "fde20de3-ff39-4cec-b24a-460c6679d188" (UID: "fde20de3-ff39-4cec-b24a-460c6679d188"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:06:04 crc kubenswrapper[4716]: I1207 16:06:04.665139 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fde20de3-ff39-4cec-b24a-460c6679d188-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 07 16:06:05 crc kubenswrapper[4716]: I1207 16:06:05.289371 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"fde20de3-ff39-4cec-b24a-460c6679d188","Type":"ContainerDied","Data":"ffbb0ab65e48f87027f96bb7a3d856c18e0c91a0fbe563f92672cf14aa4b9185"} Dec 07 16:06:05 crc kubenswrapper[4716]: I1207 16:06:05.289635 4716 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ffbb0ab65e48f87027f96bb7a3d856c18e0c91a0fbe563f92672cf14aa4b9185" Dec 07 16:06:05 crc kubenswrapper[4716]: I1207 16:06:05.289404 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 07 16:06:05 crc kubenswrapper[4716]: I1207 16:06:05.301525 4716 status_manager.go:851] "Failed to get status for pod" podUID="fde20de3-ff39-4cec-b24a-460c6679d188" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.18:6443: connect: connection refused" Dec 07 16:06:05 crc kubenswrapper[4716]: I1207 16:06:05.383744 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 07 16:06:05 crc kubenswrapper[4716]: I1207 16:06:05.384438 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 16:06:05 crc kubenswrapper[4716]: I1207 16:06:05.385106 4716 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.18:6443: connect: connection refused" Dec 07 16:06:05 crc kubenswrapper[4716]: I1207 16:06:05.385651 4716 status_manager.go:851] "Failed to get status for pod" podUID="fde20de3-ff39-4cec-b24a-460c6679d188" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.18:6443: connect: connection refused" Dec 07 16:06:05 crc kubenswrapper[4716]: I1207 16:06:05.476027 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 07 16:06:05 crc kubenswrapper[4716]: I1207 16:06:05.476109 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 07 16:06:05 crc kubenswrapper[4716]: I1207 16:06:05.476169 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 07 16:06:05 crc kubenswrapper[4716]: I1207 16:06:05.476277 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 16:06:05 crc kubenswrapper[4716]: I1207 16:06:05.476340 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 16:06:05 crc kubenswrapper[4716]: I1207 16:06:05.476434 4716 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Dec 07 16:06:05 crc kubenswrapper[4716]: I1207 16:06:05.476478 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 16:06:05 crc kubenswrapper[4716]: I1207 16:06:05.578127 4716 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 07 16:06:05 crc kubenswrapper[4716]: I1207 16:06:05.578174 4716 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 07 16:06:05 crc kubenswrapper[4716]: I1207 16:06:05.667200 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Dec 07 16:06:06 crc kubenswrapper[4716]: I1207 16:06:06.298833 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 07 16:06:06 crc kubenswrapper[4716]: I1207 16:06:06.299830 4716 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="73f85248b4d68a9beecdacb3a01f433e97a0ffea2247ce92a41ef96855509377" exitCode=0 Dec 07 16:06:06 crc kubenswrapper[4716]: I1207 16:06:06.299901 4716 scope.go:117] "RemoveContainer" containerID="dd318647816e4a27849666b0ec4fc2ca4905295661c557dc381e1144f22cf284" Dec 07 16:06:06 crc kubenswrapper[4716]: I1207 16:06:06.299964 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 16:06:06 crc kubenswrapper[4716]: I1207 16:06:06.301217 4716 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.18:6443: connect: connection refused" Dec 07 16:06:06 crc kubenswrapper[4716]: I1207 16:06:06.301590 4716 status_manager.go:851] "Failed to get status for pod" podUID="fde20de3-ff39-4cec-b24a-460c6679d188" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.18:6443: connect: connection refused" Dec 07 16:06:06 crc kubenswrapper[4716]: I1207 16:06:06.303261 4716 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.18:6443: connect: connection refused" Dec 07 16:06:06 crc kubenswrapper[4716]: I1207 16:06:06.303624 4716 status_manager.go:851] "Failed to get status for pod" podUID="fde20de3-ff39-4cec-b24a-460c6679d188" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.18:6443: connect: connection refused" Dec 07 16:06:06 crc kubenswrapper[4716]: I1207 16:06:06.316182 4716 scope.go:117] "RemoveContainer" containerID="27bd4a54f920b2725987c1c7fd2d9fdf16cd08738c246c333a4e9164a3a582a0" Dec 07 16:06:06 crc kubenswrapper[4716]: I1207 16:06:06.333674 4716 scope.go:117] "RemoveContainer" containerID="6276df652efe6ca935e745f24ef20f87a16465475139230c63b9999234a4d12a" Dec 07 16:06:06 crc kubenswrapper[4716]: I1207 16:06:06.348378 4716 scope.go:117] "RemoveContainer" containerID="e9c3b1dd202586ee755ca89fbe12b30bbb12f95a62514f6e4d93cbbdbe55181e" Dec 07 16:06:06 crc kubenswrapper[4716]: I1207 16:06:06.364451 4716 scope.go:117] "RemoveContainer" containerID="73f85248b4d68a9beecdacb3a01f433e97a0ffea2247ce92a41ef96855509377" Dec 07 16:06:06 crc kubenswrapper[4716]: I1207 16:06:06.381710 4716 scope.go:117] "RemoveContainer" containerID="01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a" Dec 07 16:06:06 crc kubenswrapper[4716]: I1207 16:06:06.405095 4716 scope.go:117] "RemoveContainer" containerID="dd318647816e4a27849666b0ec4fc2ca4905295661c557dc381e1144f22cf284" Dec 07 16:06:06 crc kubenswrapper[4716]: E1207 16:06:06.406198 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd318647816e4a27849666b0ec4fc2ca4905295661c557dc381e1144f22cf284\": container with ID starting with dd318647816e4a27849666b0ec4fc2ca4905295661c557dc381e1144f22cf284 not found: ID does not exist" containerID="dd318647816e4a27849666b0ec4fc2ca4905295661c557dc381e1144f22cf284" Dec 07 16:06:06 crc kubenswrapper[4716]: I1207 16:06:06.406229 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd318647816e4a27849666b0ec4fc2ca4905295661c557dc381e1144f22cf284"} err="failed to get container status \"dd318647816e4a27849666b0ec4fc2ca4905295661c557dc381e1144f22cf284\": rpc error: code = NotFound desc = could not find container \"dd318647816e4a27849666b0ec4fc2ca4905295661c557dc381e1144f22cf284\": container with ID starting with dd318647816e4a27849666b0ec4fc2ca4905295661c557dc381e1144f22cf284 not found: ID does not exist" Dec 07 16:06:06 crc kubenswrapper[4716]: I1207 16:06:06.406249 4716 scope.go:117] "RemoveContainer" containerID="27bd4a54f920b2725987c1c7fd2d9fdf16cd08738c246c333a4e9164a3a582a0" Dec 07 16:06:06 crc kubenswrapper[4716]: E1207 16:06:06.406496 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"27bd4a54f920b2725987c1c7fd2d9fdf16cd08738c246c333a4e9164a3a582a0\": container with ID starting with 27bd4a54f920b2725987c1c7fd2d9fdf16cd08738c246c333a4e9164a3a582a0 not found: ID does not exist" containerID="27bd4a54f920b2725987c1c7fd2d9fdf16cd08738c246c333a4e9164a3a582a0" Dec 07 16:06:06 crc kubenswrapper[4716]: I1207 16:06:06.406587 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27bd4a54f920b2725987c1c7fd2d9fdf16cd08738c246c333a4e9164a3a582a0"} err="failed to get container status \"27bd4a54f920b2725987c1c7fd2d9fdf16cd08738c246c333a4e9164a3a582a0\": rpc error: code = NotFound desc = could not find container \"27bd4a54f920b2725987c1c7fd2d9fdf16cd08738c246c333a4e9164a3a582a0\": container with ID starting with 27bd4a54f920b2725987c1c7fd2d9fdf16cd08738c246c333a4e9164a3a582a0 not found: ID does not exist" Dec 07 16:06:06 crc kubenswrapper[4716]: I1207 16:06:06.406706 4716 scope.go:117] "RemoveContainer" containerID="6276df652efe6ca935e745f24ef20f87a16465475139230c63b9999234a4d12a" Dec 07 16:06:06 crc kubenswrapper[4716]: E1207 16:06:06.407062 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6276df652efe6ca935e745f24ef20f87a16465475139230c63b9999234a4d12a\": container with ID starting with 6276df652efe6ca935e745f24ef20f87a16465475139230c63b9999234a4d12a not found: ID does not exist" containerID="6276df652efe6ca935e745f24ef20f87a16465475139230c63b9999234a4d12a" Dec 07 16:06:06 crc kubenswrapper[4716]: I1207 16:06:06.407120 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6276df652efe6ca935e745f24ef20f87a16465475139230c63b9999234a4d12a"} err="failed to get container status \"6276df652efe6ca935e745f24ef20f87a16465475139230c63b9999234a4d12a\": rpc error: code = NotFound desc = could not find container \"6276df652efe6ca935e745f24ef20f87a16465475139230c63b9999234a4d12a\": container with ID starting with 6276df652efe6ca935e745f24ef20f87a16465475139230c63b9999234a4d12a not found: ID does not exist" Dec 07 16:06:06 crc kubenswrapper[4716]: I1207 16:06:06.407147 4716 scope.go:117] "RemoveContainer" containerID="e9c3b1dd202586ee755ca89fbe12b30bbb12f95a62514f6e4d93cbbdbe55181e" Dec 07 16:06:06 crc kubenswrapper[4716]: E1207 16:06:06.407539 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e9c3b1dd202586ee755ca89fbe12b30bbb12f95a62514f6e4d93cbbdbe55181e\": container with ID starting with e9c3b1dd202586ee755ca89fbe12b30bbb12f95a62514f6e4d93cbbdbe55181e not found: ID does not exist" containerID="e9c3b1dd202586ee755ca89fbe12b30bbb12f95a62514f6e4d93cbbdbe55181e" Dec 07 16:06:06 crc kubenswrapper[4716]: I1207 16:06:06.407661 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9c3b1dd202586ee755ca89fbe12b30bbb12f95a62514f6e4d93cbbdbe55181e"} err="failed to get container status \"e9c3b1dd202586ee755ca89fbe12b30bbb12f95a62514f6e4d93cbbdbe55181e\": rpc error: code = NotFound desc = could not find container \"e9c3b1dd202586ee755ca89fbe12b30bbb12f95a62514f6e4d93cbbdbe55181e\": container with ID starting with e9c3b1dd202586ee755ca89fbe12b30bbb12f95a62514f6e4d93cbbdbe55181e not found: ID does not exist" Dec 07 16:06:06 crc kubenswrapper[4716]: I1207 16:06:06.407759 4716 scope.go:117] "RemoveContainer" containerID="73f85248b4d68a9beecdacb3a01f433e97a0ffea2247ce92a41ef96855509377" Dec 07 16:06:06 crc kubenswrapper[4716]: E1207 16:06:06.408620 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"73f85248b4d68a9beecdacb3a01f433e97a0ffea2247ce92a41ef96855509377\": container with ID starting with 73f85248b4d68a9beecdacb3a01f433e97a0ffea2247ce92a41ef96855509377 not found: ID does not exist" containerID="73f85248b4d68a9beecdacb3a01f433e97a0ffea2247ce92a41ef96855509377" Dec 07 16:06:06 crc kubenswrapper[4716]: I1207 16:06:06.408643 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73f85248b4d68a9beecdacb3a01f433e97a0ffea2247ce92a41ef96855509377"} err="failed to get container status \"73f85248b4d68a9beecdacb3a01f433e97a0ffea2247ce92a41ef96855509377\": rpc error: code = NotFound desc = could not find container \"73f85248b4d68a9beecdacb3a01f433e97a0ffea2247ce92a41ef96855509377\": container with ID starting with 73f85248b4d68a9beecdacb3a01f433e97a0ffea2247ce92a41ef96855509377 not found: ID does not exist" Dec 07 16:06:06 crc kubenswrapper[4716]: I1207 16:06:06.408659 4716 scope.go:117] "RemoveContainer" containerID="01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a" Dec 07 16:06:06 crc kubenswrapper[4716]: E1207 16:06:06.408858 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\": container with ID starting with 01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a not found: ID does not exist" containerID="01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a" Dec 07 16:06:06 crc kubenswrapper[4716]: I1207 16:06:06.408887 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a"} err="failed to get container status \"01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\": rpc error: code = NotFound desc = could not find container \"01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a\": container with ID starting with 01eb7fe3e88a2161eaa3da3215a6a76b79bf2759fa69b0186fabac1ab8a32e7a not found: ID does not exist" Dec 07 16:06:07 crc kubenswrapper[4716]: I1207 16:06:07.665187 4716 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.18:6443: connect: connection refused" Dec 07 16:06:07 crc kubenswrapper[4716]: I1207 16:06:07.666110 4716 status_manager.go:851] "Failed to get status for pod" podUID="fde20de3-ff39-4cec-b24a-460c6679d188" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.18:6443: connect: connection refused" Dec 07 16:06:08 crc kubenswrapper[4716]: E1207 16:06:08.057235 4716 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.18:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 07 16:06:08 crc kubenswrapper[4716]: I1207 16:06:08.057993 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 07 16:06:08 crc kubenswrapper[4716]: W1207 16:06:08.084270 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-03f00b48b33fdf70501888d2500a133ad583f064ca5e38f2573cac9fce90c384 WatchSource:0}: Error finding container 03f00b48b33fdf70501888d2500a133ad583f064ca5e38f2573cac9fce90c384: Status 404 returned error can't find the container with id 03f00b48b33fdf70501888d2500a133ad583f064ca5e38f2573cac9fce90c384 Dec 07 16:06:08 crc kubenswrapper[4716]: E1207 16:06:08.087134 4716 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.18:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187efa887c6d541f openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-07 16:06:08.085849119 +0000 UTC m=+230.776134031,LastTimestamp:2025-12-07 16:06:08.085849119 +0000 UTC m=+230.776134031,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 07 16:06:08 crc kubenswrapper[4716]: I1207 16:06:08.319710 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"03f00b48b33fdf70501888d2500a133ad583f064ca5e38f2573cac9fce90c384"} Dec 07 16:06:09 crc kubenswrapper[4716]: I1207 16:06:09.326451 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"a8901ca0b7fa2ead1bf4a7fe6266ec940b74a4b8ecccfa7cf3881c75f7d7cdf6"} Dec 07 16:06:09 crc kubenswrapper[4716]: I1207 16:06:09.327593 4716 status_manager.go:851] "Failed to get status for pod" podUID="fde20de3-ff39-4cec-b24a-460c6679d188" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.18:6443: connect: connection refused" Dec 07 16:06:09 crc kubenswrapper[4716]: E1207 16:06:09.327620 4716 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.18:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 07 16:06:10 crc kubenswrapper[4716]: E1207 16:06:10.333877 4716 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.18:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 07 16:06:10 crc kubenswrapper[4716]: E1207 16:06:10.839096 4716 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:06:10Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:06:10Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:06:10Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-07T16:06:10Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.102.83.18:6443: connect: connection refused" Dec 07 16:06:10 crc kubenswrapper[4716]: E1207 16:06:10.839564 4716 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.18:6443: connect: connection refused" Dec 07 16:06:10 crc kubenswrapper[4716]: E1207 16:06:10.839841 4716 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.18:6443: connect: connection refused" Dec 07 16:06:10 crc kubenswrapper[4716]: E1207 16:06:10.840095 4716 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.18:6443: connect: connection refused" Dec 07 16:06:10 crc kubenswrapper[4716]: E1207 16:06:10.840426 4716 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.18:6443: connect: connection refused" Dec 07 16:06:10 crc kubenswrapper[4716]: E1207 16:06:10.840459 4716 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 07 16:06:11 crc kubenswrapper[4716]: E1207 16:06:11.718184 4716 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.18:6443: connect: connection refused" Dec 07 16:06:11 crc kubenswrapper[4716]: E1207 16:06:11.718864 4716 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.18:6443: connect: connection refused" Dec 07 16:06:11 crc kubenswrapper[4716]: E1207 16:06:11.719690 4716 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.18:6443: connect: connection refused" Dec 07 16:06:11 crc kubenswrapper[4716]: E1207 16:06:11.720887 4716 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.18:6443: connect: connection refused" Dec 07 16:06:11 crc kubenswrapper[4716]: E1207 16:06:11.721414 4716 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.18:6443: connect: connection refused" Dec 07 16:06:11 crc kubenswrapper[4716]: I1207 16:06:11.721477 4716 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Dec 07 16:06:11 crc kubenswrapper[4716]: E1207 16:06:11.722445 4716 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.18:6443: connect: connection refused" interval="200ms" Dec 07 16:06:11 crc kubenswrapper[4716]: E1207 16:06:11.923298 4716 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.18:6443: connect: connection refused" interval="400ms" Dec 07 16:06:12 crc kubenswrapper[4716]: E1207 16:06:12.324569 4716 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.18:6443: connect: connection refused" interval="800ms" Dec 07 16:06:13 crc kubenswrapper[4716]: E1207 16:06:13.126433 4716 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.18:6443: connect: connection refused" interval="1.6s" Dec 07 16:06:13 crc kubenswrapper[4716]: E1207 16:06:13.508443 4716 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.18:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187efa887c6d541f openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-07 16:06:08.085849119 +0000 UTC m=+230.776134031,LastTimestamp:2025-12-07 16:06:08.085849119 +0000 UTC m=+230.776134031,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 07 16:06:14 crc kubenswrapper[4716]: I1207 16:06:14.657150 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 16:06:14 crc kubenswrapper[4716]: I1207 16:06:14.658246 4716 status_manager.go:851] "Failed to get status for pod" podUID="fde20de3-ff39-4cec-b24a-460c6679d188" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.18:6443: connect: connection refused" Dec 07 16:06:14 crc kubenswrapper[4716]: I1207 16:06:14.669666 4716 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="cdee9957-48b9-458f-ba71-e0cd39f80596" Dec 07 16:06:14 crc kubenswrapper[4716]: I1207 16:06:14.669701 4716 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="cdee9957-48b9-458f-ba71-e0cd39f80596" Dec 07 16:06:14 crc kubenswrapper[4716]: E1207 16:06:14.670142 4716 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.18:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 16:06:14 crc kubenswrapper[4716]: I1207 16:06:14.670569 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 16:06:14 crc kubenswrapper[4716]: E1207 16:06:14.727477 4716 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.18:6443: connect: connection refused" interval="3.2s" Dec 07 16:06:15 crc kubenswrapper[4716]: I1207 16:06:15.362394 4716 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="e4da175c430e70a8fab8a61219bd3c2de13d0db9d9772df30748c2f0b02bd0a8" exitCode=0 Dec 07 16:06:15 crc kubenswrapper[4716]: I1207 16:06:15.362475 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"e4da175c430e70a8fab8a61219bd3c2de13d0db9d9772df30748c2f0b02bd0a8"} Dec 07 16:06:15 crc kubenswrapper[4716]: I1207 16:06:15.362531 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"e94e35fea6db35c069bb6413022c40ce0f20387f29e1372142b2c0d5e672eaa3"} Dec 07 16:06:15 crc kubenswrapper[4716]: I1207 16:06:15.362956 4716 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="cdee9957-48b9-458f-ba71-e0cd39f80596" Dec 07 16:06:15 crc kubenswrapper[4716]: I1207 16:06:15.362983 4716 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="cdee9957-48b9-458f-ba71-e0cd39f80596" Dec 07 16:06:15 crc kubenswrapper[4716]: I1207 16:06:15.363468 4716 status_manager.go:851] "Failed to get status for pod" podUID="fde20de3-ff39-4cec-b24a-460c6679d188" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.18:6443: connect: connection refused" Dec 07 16:06:15 crc kubenswrapper[4716]: E1207 16:06:15.363772 4716 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.18:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 16:06:15 crc kubenswrapper[4716]: I1207 16:06:15.735925 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-z587m" podUID="ffff2831-338e-423e-81f2-c8d3b5de7785" containerName="oauth-openshift" containerID="cri-o://70e6048ead4ed66e668777500ea8cf6f6fe15c89bca412f93a7a7cc1823076b3" gracePeriod=15 Dec 07 16:06:16 crc kubenswrapper[4716]: I1207 16:06:16.222513 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-z587m" Dec 07 16:06:16 crc kubenswrapper[4716]: I1207 16:06:16.314366 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ffff2831-338e-423e-81f2-c8d3b5de7785-audit-dir\") pod \"ffff2831-338e-423e-81f2-c8d3b5de7785\" (UID: \"ffff2831-338e-423e-81f2-c8d3b5de7785\") " Dec 07 16:06:16 crc kubenswrapper[4716]: I1207 16:06:16.314691 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nw5sf\" (UniqueName: \"kubernetes.io/projected/ffff2831-338e-423e-81f2-c8d3b5de7785-kube-api-access-nw5sf\") pod \"ffff2831-338e-423e-81f2-c8d3b5de7785\" (UID: \"ffff2831-338e-423e-81f2-c8d3b5de7785\") " Dec 07 16:06:16 crc kubenswrapper[4716]: I1207 16:06:16.314721 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ffff2831-338e-423e-81f2-c8d3b5de7785-audit-policies\") pod \"ffff2831-338e-423e-81f2-c8d3b5de7785\" (UID: \"ffff2831-338e-423e-81f2-c8d3b5de7785\") " Dec 07 16:06:16 crc kubenswrapper[4716]: I1207 16:06:16.314489 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ffff2831-338e-423e-81f2-c8d3b5de7785-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "ffff2831-338e-423e-81f2-c8d3b5de7785" (UID: "ffff2831-338e-423e-81f2-c8d3b5de7785"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 16:06:16 crc kubenswrapper[4716]: I1207 16:06:16.314766 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ffff2831-338e-423e-81f2-c8d3b5de7785-v4-0-config-system-ocp-branding-template\") pod \"ffff2831-338e-423e-81f2-c8d3b5de7785\" (UID: \"ffff2831-338e-423e-81f2-c8d3b5de7785\") " Dec 07 16:06:16 crc kubenswrapper[4716]: I1207 16:06:16.314805 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ffff2831-338e-423e-81f2-c8d3b5de7785-v4-0-config-system-trusted-ca-bundle\") pod \"ffff2831-338e-423e-81f2-c8d3b5de7785\" (UID: \"ffff2831-338e-423e-81f2-c8d3b5de7785\") " Dec 07 16:06:16 crc kubenswrapper[4716]: I1207 16:06:16.314828 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ffff2831-338e-423e-81f2-c8d3b5de7785-v4-0-config-system-session\") pod \"ffff2831-338e-423e-81f2-c8d3b5de7785\" (UID: \"ffff2831-338e-423e-81f2-c8d3b5de7785\") " Dec 07 16:06:16 crc kubenswrapper[4716]: I1207 16:06:16.314850 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ffff2831-338e-423e-81f2-c8d3b5de7785-v4-0-config-user-template-login\") pod \"ffff2831-338e-423e-81f2-c8d3b5de7785\" (UID: \"ffff2831-338e-423e-81f2-c8d3b5de7785\") " Dec 07 16:06:16 crc kubenswrapper[4716]: I1207 16:06:16.314874 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ffff2831-338e-423e-81f2-c8d3b5de7785-v4-0-config-system-cliconfig\") pod \"ffff2831-338e-423e-81f2-c8d3b5de7785\" (UID: \"ffff2831-338e-423e-81f2-c8d3b5de7785\") " Dec 07 16:06:16 crc kubenswrapper[4716]: I1207 16:06:16.314904 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ffff2831-338e-423e-81f2-c8d3b5de7785-v4-0-config-user-idp-0-file-data\") pod \"ffff2831-338e-423e-81f2-c8d3b5de7785\" (UID: \"ffff2831-338e-423e-81f2-c8d3b5de7785\") " Dec 07 16:06:16 crc kubenswrapper[4716]: I1207 16:06:16.314929 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ffff2831-338e-423e-81f2-c8d3b5de7785-v4-0-config-system-service-ca\") pod \"ffff2831-338e-423e-81f2-c8d3b5de7785\" (UID: \"ffff2831-338e-423e-81f2-c8d3b5de7785\") " Dec 07 16:06:16 crc kubenswrapper[4716]: I1207 16:06:16.314953 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ffff2831-338e-423e-81f2-c8d3b5de7785-v4-0-config-system-serving-cert\") pod \"ffff2831-338e-423e-81f2-c8d3b5de7785\" (UID: \"ffff2831-338e-423e-81f2-c8d3b5de7785\") " Dec 07 16:06:16 crc kubenswrapper[4716]: I1207 16:06:16.314981 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ffff2831-338e-423e-81f2-c8d3b5de7785-v4-0-config-system-router-certs\") pod \"ffff2831-338e-423e-81f2-c8d3b5de7785\" (UID: \"ffff2831-338e-423e-81f2-c8d3b5de7785\") " Dec 07 16:06:16 crc kubenswrapper[4716]: I1207 16:06:16.315011 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ffff2831-338e-423e-81f2-c8d3b5de7785-v4-0-config-user-template-provider-selection\") pod \"ffff2831-338e-423e-81f2-c8d3b5de7785\" (UID: \"ffff2831-338e-423e-81f2-c8d3b5de7785\") " Dec 07 16:06:16 crc kubenswrapper[4716]: I1207 16:06:16.315046 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ffff2831-338e-423e-81f2-c8d3b5de7785-v4-0-config-user-template-error\") pod \"ffff2831-338e-423e-81f2-c8d3b5de7785\" (UID: \"ffff2831-338e-423e-81f2-c8d3b5de7785\") " Dec 07 16:06:16 crc kubenswrapper[4716]: I1207 16:06:16.315238 4716 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ffff2831-338e-423e-81f2-c8d3b5de7785-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 07 16:06:16 crc kubenswrapper[4716]: I1207 16:06:16.316036 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ffff2831-338e-423e-81f2-c8d3b5de7785-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "ffff2831-338e-423e-81f2-c8d3b5de7785" (UID: "ffff2831-338e-423e-81f2-c8d3b5de7785"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:06:16 crc kubenswrapper[4716]: I1207 16:06:16.316310 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ffff2831-338e-423e-81f2-c8d3b5de7785-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "ffff2831-338e-423e-81f2-c8d3b5de7785" (UID: "ffff2831-338e-423e-81f2-c8d3b5de7785"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:06:16 crc kubenswrapper[4716]: I1207 16:06:16.316342 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ffff2831-338e-423e-81f2-c8d3b5de7785-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "ffff2831-338e-423e-81f2-c8d3b5de7785" (UID: "ffff2831-338e-423e-81f2-c8d3b5de7785"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:06:16 crc kubenswrapper[4716]: I1207 16:06:16.318190 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ffff2831-338e-423e-81f2-c8d3b5de7785-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "ffff2831-338e-423e-81f2-c8d3b5de7785" (UID: "ffff2831-338e-423e-81f2-c8d3b5de7785"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:06:16 crc kubenswrapper[4716]: I1207 16:06:16.320423 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ffff2831-338e-423e-81f2-c8d3b5de7785-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "ffff2831-338e-423e-81f2-c8d3b5de7785" (UID: "ffff2831-338e-423e-81f2-c8d3b5de7785"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:06:16 crc kubenswrapper[4716]: I1207 16:06:16.320718 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ffff2831-338e-423e-81f2-c8d3b5de7785-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "ffff2831-338e-423e-81f2-c8d3b5de7785" (UID: "ffff2831-338e-423e-81f2-c8d3b5de7785"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:06:16 crc kubenswrapper[4716]: I1207 16:06:16.320908 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ffff2831-338e-423e-81f2-c8d3b5de7785-kube-api-access-nw5sf" (OuterVolumeSpecName: "kube-api-access-nw5sf") pod "ffff2831-338e-423e-81f2-c8d3b5de7785" (UID: "ffff2831-338e-423e-81f2-c8d3b5de7785"). InnerVolumeSpecName "kube-api-access-nw5sf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:06:16 crc kubenswrapper[4716]: I1207 16:06:16.320987 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ffff2831-338e-423e-81f2-c8d3b5de7785-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "ffff2831-338e-423e-81f2-c8d3b5de7785" (UID: "ffff2831-338e-423e-81f2-c8d3b5de7785"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:06:16 crc kubenswrapper[4716]: I1207 16:06:16.321250 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ffff2831-338e-423e-81f2-c8d3b5de7785-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "ffff2831-338e-423e-81f2-c8d3b5de7785" (UID: "ffff2831-338e-423e-81f2-c8d3b5de7785"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:06:16 crc kubenswrapper[4716]: I1207 16:06:16.322028 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ffff2831-338e-423e-81f2-c8d3b5de7785-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "ffff2831-338e-423e-81f2-c8d3b5de7785" (UID: "ffff2831-338e-423e-81f2-c8d3b5de7785"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:06:16 crc kubenswrapper[4716]: I1207 16:06:16.326110 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ffff2831-338e-423e-81f2-c8d3b5de7785-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "ffff2831-338e-423e-81f2-c8d3b5de7785" (UID: "ffff2831-338e-423e-81f2-c8d3b5de7785"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:06:16 crc kubenswrapper[4716]: I1207 16:06:16.335370 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ffff2831-338e-423e-81f2-c8d3b5de7785-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "ffff2831-338e-423e-81f2-c8d3b5de7785" (UID: "ffff2831-338e-423e-81f2-c8d3b5de7785"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:06:16 crc kubenswrapper[4716]: I1207 16:06:16.338752 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ffff2831-338e-423e-81f2-c8d3b5de7785-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "ffff2831-338e-423e-81f2-c8d3b5de7785" (UID: "ffff2831-338e-423e-81f2-c8d3b5de7785"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:06:16 crc kubenswrapper[4716]: I1207 16:06:16.382323 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"f40610ebf151b96dec899f0fb8608ce13d5149ca0668c645b40d66cfca889d30"} Dec 07 16:06:16 crc kubenswrapper[4716]: I1207 16:06:16.382372 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"6e4e7c17d023461a3c88a9b8debf5c36f70881288863cd690113a82efa8f451c"} Dec 07 16:06:16 crc kubenswrapper[4716]: I1207 16:06:16.382386 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"bbdc2fb324313ad45adcfe4a24f7ba29983de78c909e9f2b4354cf3e5b00f7f4"} Dec 07 16:06:16 crc kubenswrapper[4716]: I1207 16:06:16.382396 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"c5159ba5084058f0e7aa15f58a69b4465f88301834c2a8970f80841d12960d29"} Dec 07 16:06:16 crc kubenswrapper[4716]: I1207 16:06:16.383888 4716 generic.go:334] "Generic (PLEG): container finished" podID="ffff2831-338e-423e-81f2-c8d3b5de7785" containerID="70e6048ead4ed66e668777500ea8cf6f6fe15c89bca412f93a7a7cc1823076b3" exitCode=0 Dec 07 16:06:16 crc kubenswrapper[4716]: I1207 16:06:16.383965 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-z587m" event={"ID":"ffff2831-338e-423e-81f2-c8d3b5de7785","Type":"ContainerDied","Data":"70e6048ead4ed66e668777500ea8cf6f6fe15c89bca412f93a7a7cc1823076b3"} Dec 07 16:06:16 crc kubenswrapper[4716]: I1207 16:06:16.383999 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-z587m" event={"ID":"ffff2831-338e-423e-81f2-c8d3b5de7785","Type":"ContainerDied","Data":"659375d5d974bfaecb1f1380f95cb82e35d3ceb12360c50b80b664804fcee30e"} Dec 07 16:06:16 crc kubenswrapper[4716]: I1207 16:06:16.384008 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-z587m" Dec 07 16:06:16 crc kubenswrapper[4716]: I1207 16:06:16.384020 4716 scope.go:117] "RemoveContainer" containerID="70e6048ead4ed66e668777500ea8cf6f6fe15c89bca412f93a7a7cc1823076b3" Dec 07 16:06:16 crc kubenswrapper[4716]: I1207 16:06:16.391510 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 07 16:06:16 crc kubenswrapper[4716]: I1207 16:06:16.391568 4716 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="6a9e9cf048220ca14c996847e9ced9f2367b9331dc6c2836dd023658855cd79d" exitCode=1 Dec 07 16:06:16 crc kubenswrapper[4716]: I1207 16:06:16.391602 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"6a9e9cf048220ca14c996847e9ced9f2367b9331dc6c2836dd023658855cd79d"} Dec 07 16:06:16 crc kubenswrapper[4716]: I1207 16:06:16.392173 4716 scope.go:117] "RemoveContainer" containerID="6a9e9cf048220ca14c996847e9ced9f2367b9331dc6c2836dd023658855cd79d" Dec 07 16:06:16 crc kubenswrapper[4716]: I1207 16:06:16.407298 4716 scope.go:117] "RemoveContainer" containerID="70e6048ead4ed66e668777500ea8cf6f6fe15c89bca412f93a7a7cc1823076b3" Dec 07 16:06:16 crc kubenswrapper[4716]: E1207 16:06:16.409234 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"70e6048ead4ed66e668777500ea8cf6f6fe15c89bca412f93a7a7cc1823076b3\": container with ID starting with 70e6048ead4ed66e668777500ea8cf6f6fe15c89bca412f93a7a7cc1823076b3 not found: ID does not exist" containerID="70e6048ead4ed66e668777500ea8cf6f6fe15c89bca412f93a7a7cc1823076b3" Dec 07 16:06:16 crc kubenswrapper[4716]: I1207 16:06:16.409271 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"70e6048ead4ed66e668777500ea8cf6f6fe15c89bca412f93a7a7cc1823076b3"} err="failed to get container status \"70e6048ead4ed66e668777500ea8cf6f6fe15c89bca412f93a7a7cc1823076b3\": rpc error: code = NotFound desc = could not find container \"70e6048ead4ed66e668777500ea8cf6f6fe15c89bca412f93a7a7cc1823076b3\": container with ID starting with 70e6048ead4ed66e668777500ea8cf6f6fe15c89bca412f93a7a7cc1823076b3 not found: ID does not exist" Dec 07 16:06:16 crc kubenswrapper[4716]: I1207 16:06:16.417493 4716 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ffff2831-338e-423e-81f2-c8d3b5de7785-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 07 16:06:16 crc kubenswrapper[4716]: I1207 16:06:16.417529 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nw5sf\" (UniqueName: \"kubernetes.io/projected/ffff2831-338e-423e-81f2-c8d3b5de7785-kube-api-access-nw5sf\") on node \"crc\" DevicePath \"\"" Dec 07 16:06:16 crc kubenswrapper[4716]: I1207 16:06:16.417544 4716 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ffff2831-338e-423e-81f2-c8d3b5de7785-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 07 16:06:16 crc kubenswrapper[4716]: I1207 16:06:16.417556 4716 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ffff2831-338e-423e-81f2-c8d3b5de7785-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 07 16:06:16 crc kubenswrapper[4716]: I1207 16:06:16.417571 4716 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ffff2831-338e-423e-81f2-c8d3b5de7785-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 16:06:16 crc kubenswrapper[4716]: I1207 16:06:16.417583 4716 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ffff2831-338e-423e-81f2-c8d3b5de7785-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 07 16:06:16 crc kubenswrapper[4716]: I1207 16:06:16.417595 4716 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ffff2831-338e-423e-81f2-c8d3b5de7785-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 07 16:06:16 crc kubenswrapper[4716]: I1207 16:06:16.417607 4716 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ffff2831-338e-423e-81f2-c8d3b5de7785-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 07 16:06:16 crc kubenswrapper[4716]: I1207 16:06:16.417619 4716 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ffff2831-338e-423e-81f2-c8d3b5de7785-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 07 16:06:16 crc kubenswrapper[4716]: I1207 16:06:16.417632 4716 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ffff2831-338e-423e-81f2-c8d3b5de7785-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 07 16:06:16 crc kubenswrapper[4716]: I1207 16:06:16.417644 4716 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ffff2831-338e-423e-81f2-c8d3b5de7785-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 07 16:06:16 crc kubenswrapper[4716]: I1207 16:06:16.417656 4716 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ffff2831-338e-423e-81f2-c8d3b5de7785-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 07 16:06:16 crc kubenswrapper[4716]: I1207 16:06:16.417668 4716 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ffff2831-338e-423e-81f2-c8d3b5de7785-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 07 16:06:17 crc kubenswrapper[4716]: I1207 16:06:17.398938 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"edf99ddea56ea214e1460235dbadc6f14aa501f4de502260967de9e4df843d22"} Dec 07 16:06:17 crc kubenswrapper[4716]: I1207 16:06:17.399192 4716 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="cdee9957-48b9-458f-ba71-e0cd39f80596" Dec 07 16:06:17 crc kubenswrapper[4716]: I1207 16:06:17.399205 4716 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="cdee9957-48b9-458f-ba71-e0cd39f80596" Dec 07 16:06:17 crc kubenswrapper[4716]: I1207 16:06:17.399408 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 16:06:17 crc kubenswrapper[4716]: I1207 16:06:17.401280 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 07 16:06:17 crc kubenswrapper[4716]: I1207 16:06:17.401336 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"c8f23e65fd9a08d0fe6a49f0ca43d10265c21581b7c9f0c06ed02f9f283af93b"} Dec 07 16:06:19 crc kubenswrapper[4716]: I1207 16:06:19.331323 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 07 16:06:19 crc kubenswrapper[4716]: I1207 16:06:19.670986 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 16:06:19 crc kubenswrapper[4716]: I1207 16:06:19.671055 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 16:06:19 crc kubenswrapper[4716]: I1207 16:06:19.679584 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 16:06:22 crc kubenswrapper[4716]: I1207 16:06:22.405742 4716 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 16:06:22 crc kubenswrapper[4716]: I1207 16:06:22.439394 4716 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="cdee9957-48b9-458f-ba71-e0cd39f80596" Dec 07 16:06:22 crc kubenswrapper[4716]: I1207 16:06:22.439421 4716 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="cdee9957-48b9-458f-ba71-e0cd39f80596" Dec 07 16:06:22 crc kubenswrapper[4716]: I1207 16:06:22.443131 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 16:06:22 crc kubenswrapper[4716]: I1207 16:06:22.445709 4716 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="411ac43c-4539-4a80-9e65-39889d068eed" Dec 07 16:06:23 crc kubenswrapper[4716]: I1207 16:06:23.444836 4716 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="cdee9957-48b9-458f-ba71-e0cd39f80596" Dec 07 16:06:23 crc kubenswrapper[4716]: I1207 16:06:23.444878 4716 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="cdee9957-48b9-458f-ba71-e0cd39f80596" Dec 07 16:06:23 crc kubenswrapper[4716]: I1207 16:06:23.485287 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 07 16:06:23 crc kubenswrapper[4716]: I1207 16:06:23.489649 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 07 16:06:27 crc kubenswrapper[4716]: I1207 16:06:27.676966 4716 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="411ac43c-4539-4a80-9e65-39889d068eed" Dec 07 16:06:29 crc kubenswrapper[4716]: I1207 16:06:29.338309 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 07 16:06:30 crc kubenswrapper[4716]: I1207 16:06:30.047317 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 07 16:06:31 crc kubenswrapper[4716]: I1207 16:06:31.980499 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 07 16:06:32 crc kubenswrapper[4716]: I1207 16:06:32.250799 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 07 16:06:33 crc kubenswrapper[4716]: I1207 16:06:33.013720 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 07 16:06:33 crc kubenswrapper[4716]: I1207 16:06:33.357284 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 07 16:06:33 crc kubenswrapper[4716]: I1207 16:06:33.649641 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 07 16:06:33 crc kubenswrapper[4716]: I1207 16:06:33.955142 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 07 16:06:34 crc kubenswrapper[4716]: I1207 16:06:34.174877 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 07 16:06:34 crc kubenswrapper[4716]: I1207 16:06:34.229171 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 07 16:06:35 crc kubenswrapper[4716]: I1207 16:06:35.006830 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 07 16:06:35 crc kubenswrapper[4716]: I1207 16:06:35.111011 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 07 16:06:35 crc kubenswrapper[4716]: I1207 16:06:35.319032 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 07 16:06:35 crc kubenswrapper[4716]: I1207 16:06:35.529423 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 07 16:06:35 crc kubenswrapper[4716]: I1207 16:06:35.688425 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 07 16:06:36 crc kubenswrapper[4716]: I1207 16:06:36.217766 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 07 16:06:36 crc kubenswrapper[4716]: I1207 16:06:36.221380 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 07 16:06:36 crc kubenswrapper[4716]: I1207 16:06:36.253749 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 07 16:06:36 crc kubenswrapper[4716]: I1207 16:06:36.307154 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 07 16:06:36 crc kubenswrapper[4716]: I1207 16:06:36.309220 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 07 16:06:36 crc kubenswrapper[4716]: I1207 16:06:36.361268 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 07 16:06:36 crc kubenswrapper[4716]: I1207 16:06:36.406200 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 07 16:06:36 crc kubenswrapper[4716]: I1207 16:06:36.434845 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 07 16:06:36 crc kubenswrapper[4716]: I1207 16:06:36.545900 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 07 16:06:36 crc kubenswrapper[4716]: I1207 16:06:36.976922 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 07 16:06:37 crc kubenswrapper[4716]: I1207 16:06:37.030614 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 07 16:06:37 crc kubenswrapper[4716]: I1207 16:06:37.041704 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 07 16:06:37 crc kubenswrapper[4716]: I1207 16:06:37.137946 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 07 16:06:37 crc kubenswrapper[4716]: I1207 16:06:37.223508 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 07 16:06:37 crc kubenswrapper[4716]: I1207 16:06:37.280251 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 07 16:06:37 crc kubenswrapper[4716]: I1207 16:06:37.303364 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 07 16:06:37 crc kubenswrapper[4716]: I1207 16:06:37.315129 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 07 16:06:37 crc kubenswrapper[4716]: I1207 16:06:37.344315 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 07 16:06:37 crc kubenswrapper[4716]: I1207 16:06:37.350683 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 07 16:06:37 crc kubenswrapper[4716]: I1207 16:06:37.446682 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 07 16:06:37 crc kubenswrapper[4716]: I1207 16:06:37.447702 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 07 16:06:37 crc kubenswrapper[4716]: I1207 16:06:37.450736 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 07 16:06:37 crc kubenswrapper[4716]: I1207 16:06:37.536833 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 07 16:06:37 crc kubenswrapper[4716]: I1207 16:06:37.605090 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 07 16:06:37 crc kubenswrapper[4716]: I1207 16:06:37.639762 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 07 16:06:37 crc kubenswrapper[4716]: I1207 16:06:37.801331 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 07 16:06:37 crc kubenswrapper[4716]: I1207 16:06:37.808167 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 07 16:06:37 crc kubenswrapper[4716]: I1207 16:06:37.963307 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.040595 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.048384 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.057246 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.095679 4716 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.097348 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.099798 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-authentication/oauth-openshift-558db77b4-z587m"] Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.099867 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-authentication/oauth-openshift-6fb796c88-8hqk6"] Dec 07 16:06:38 crc kubenswrapper[4716]: E1207 16:06:38.100065 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fde20de3-ff39-4cec-b24a-460c6679d188" containerName="installer" Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.100105 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="fde20de3-ff39-4cec-b24a-460c6679d188" containerName="installer" Dec 07 16:06:38 crc kubenswrapper[4716]: E1207 16:06:38.100119 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffff2831-338e-423e-81f2-c8d3b5de7785" containerName="oauth-openshift" Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.100125 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffff2831-338e-423e-81f2-c8d3b5de7785" containerName="oauth-openshift" Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.100239 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffff2831-338e-423e-81f2-c8d3b5de7785" containerName="oauth-openshift" Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.100251 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="fde20de3-ff39-4cec-b24a-460c6679d188" containerName="installer" Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.100590 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-6fb796c88-8hqk6" Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.103494 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.104544 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.105296 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.105526 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.105565 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.106495 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.106312 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.106753 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.107381 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.110962 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.112305 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.113011 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.113171 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.117923 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.119499 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.126330 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.136853 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=16.136834942 podStartE2EDuration="16.136834942s" podCreationTimestamp="2025-12-07 16:06:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:06:38.135543608 +0000 UTC m=+260.825828520" watchObservedRunningTime="2025-12-07 16:06:38.136834942 +0000 UTC m=+260.827119854" Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.284678 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b25484eb-eb98-48f9-a3fa-676622944828-audit-dir\") pod \"oauth-openshift-6fb796c88-8hqk6\" (UID: \"b25484eb-eb98-48f9-a3fa-676622944828\") " pod="openshift-authentication/oauth-openshift-6fb796c88-8hqk6" Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.284767 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/b25484eb-eb98-48f9-a3fa-676622944828-v4-0-config-user-template-error\") pod \"oauth-openshift-6fb796c88-8hqk6\" (UID: \"b25484eb-eb98-48f9-a3fa-676622944828\") " pod="openshift-authentication/oauth-openshift-6fb796c88-8hqk6" Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.284810 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/b25484eb-eb98-48f9-a3fa-676622944828-v4-0-config-system-router-certs\") pod \"oauth-openshift-6fb796c88-8hqk6\" (UID: \"b25484eb-eb98-48f9-a3fa-676622944828\") " pod="openshift-authentication/oauth-openshift-6fb796c88-8hqk6" Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.284850 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/b25484eb-eb98-48f9-a3fa-676622944828-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6fb796c88-8hqk6\" (UID: \"b25484eb-eb98-48f9-a3fa-676622944828\") " pod="openshift-authentication/oauth-openshift-6fb796c88-8hqk6" Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.284939 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/b25484eb-eb98-48f9-a3fa-676622944828-v4-0-config-system-session\") pod \"oauth-openshift-6fb796c88-8hqk6\" (UID: \"b25484eb-eb98-48f9-a3fa-676622944828\") " pod="openshift-authentication/oauth-openshift-6fb796c88-8hqk6" Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.285001 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/b25484eb-eb98-48f9-a3fa-676622944828-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6fb796c88-8hqk6\" (UID: \"b25484eb-eb98-48f9-a3fa-676622944828\") " pod="openshift-authentication/oauth-openshift-6fb796c88-8hqk6" Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.285047 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/b25484eb-eb98-48f9-a3fa-676622944828-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6fb796c88-8hqk6\" (UID: \"b25484eb-eb98-48f9-a3fa-676622944828\") " pod="openshift-authentication/oauth-openshift-6fb796c88-8hqk6" Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.285101 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/b25484eb-eb98-48f9-a3fa-676622944828-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6fb796c88-8hqk6\" (UID: \"b25484eb-eb98-48f9-a3fa-676622944828\") " pod="openshift-authentication/oauth-openshift-6fb796c88-8hqk6" Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.285192 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j948c\" (UniqueName: \"kubernetes.io/projected/b25484eb-eb98-48f9-a3fa-676622944828-kube-api-access-j948c\") pod \"oauth-openshift-6fb796c88-8hqk6\" (UID: \"b25484eb-eb98-48f9-a3fa-676622944828\") " pod="openshift-authentication/oauth-openshift-6fb796c88-8hqk6" Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.285513 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b25484eb-eb98-48f9-a3fa-676622944828-audit-policies\") pod \"oauth-openshift-6fb796c88-8hqk6\" (UID: \"b25484eb-eb98-48f9-a3fa-676622944828\") " pod="openshift-authentication/oauth-openshift-6fb796c88-8hqk6" Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.285614 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b25484eb-eb98-48f9-a3fa-676622944828-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6fb796c88-8hqk6\" (UID: \"b25484eb-eb98-48f9-a3fa-676622944828\") " pod="openshift-authentication/oauth-openshift-6fb796c88-8hqk6" Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.285774 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/b25484eb-eb98-48f9-a3fa-676622944828-v4-0-config-system-service-ca\") pod \"oauth-openshift-6fb796c88-8hqk6\" (UID: \"b25484eb-eb98-48f9-a3fa-676622944828\") " pod="openshift-authentication/oauth-openshift-6fb796c88-8hqk6" Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.286001 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/b25484eb-eb98-48f9-a3fa-676622944828-v4-0-config-user-template-login\") pod \"oauth-openshift-6fb796c88-8hqk6\" (UID: \"b25484eb-eb98-48f9-a3fa-676622944828\") " pod="openshift-authentication/oauth-openshift-6fb796c88-8hqk6" Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.286132 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/b25484eb-eb98-48f9-a3fa-676622944828-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6fb796c88-8hqk6\" (UID: \"b25484eb-eb98-48f9-a3fa-676622944828\") " pod="openshift-authentication/oauth-openshift-6fb796c88-8hqk6" Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.386861 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/b25484eb-eb98-48f9-a3fa-676622944828-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6fb796c88-8hqk6\" (UID: \"b25484eb-eb98-48f9-a3fa-676622944828\") " pod="openshift-authentication/oauth-openshift-6fb796c88-8hqk6" Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.386912 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b25484eb-eb98-48f9-a3fa-676622944828-audit-dir\") pod \"oauth-openshift-6fb796c88-8hqk6\" (UID: \"b25484eb-eb98-48f9-a3fa-676622944828\") " pod="openshift-authentication/oauth-openshift-6fb796c88-8hqk6" Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.386954 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/b25484eb-eb98-48f9-a3fa-676622944828-v4-0-config-user-template-error\") pod \"oauth-openshift-6fb796c88-8hqk6\" (UID: \"b25484eb-eb98-48f9-a3fa-676622944828\") " pod="openshift-authentication/oauth-openshift-6fb796c88-8hqk6" Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.386979 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/b25484eb-eb98-48f9-a3fa-676622944828-v4-0-config-system-router-certs\") pod \"oauth-openshift-6fb796c88-8hqk6\" (UID: \"b25484eb-eb98-48f9-a3fa-676622944828\") " pod="openshift-authentication/oauth-openshift-6fb796c88-8hqk6" Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.387003 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/b25484eb-eb98-48f9-a3fa-676622944828-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6fb796c88-8hqk6\" (UID: \"b25484eb-eb98-48f9-a3fa-676622944828\") " pod="openshift-authentication/oauth-openshift-6fb796c88-8hqk6" Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.387027 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/b25484eb-eb98-48f9-a3fa-676622944828-v4-0-config-system-session\") pod \"oauth-openshift-6fb796c88-8hqk6\" (UID: \"b25484eb-eb98-48f9-a3fa-676622944828\") " pod="openshift-authentication/oauth-openshift-6fb796c88-8hqk6" Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.387049 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/b25484eb-eb98-48f9-a3fa-676622944828-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6fb796c88-8hqk6\" (UID: \"b25484eb-eb98-48f9-a3fa-676622944828\") " pod="openshift-authentication/oauth-openshift-6fb796c88-8hqk6" Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.387073 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/b25484eb-eb98-48f9-a3fa-676622944828-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6fb796c88-8hqk6\" (UID: \"b25484eb-eb98-48f9-a3fa-676622944828\") " pod="openshift-authentication/oauth-openshift-6fb796c88-8hqk6" Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.387097 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b25484eb-eb98-48f9-a3fa-676622944828-audit-dir\") pod \"oauth-openshift-6fb796c88-8hqk6\" (UID: \"b25484eb-eb98-48f9-a3fa-676622944828\") " pod="openshift-authentication/oauth-openshift-6fb796c88-8hqk6" Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.387118 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/b25484eb-eb98-48f9-a3fa-676622944828-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6fb796c88-8hqk6\" (UID: \"b25484eb-eb98-48f9-a3fa-676622944828\") " pod="openshift-authentication/oauth-openshift-6fb796c88-8hqk6" Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.387189 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j948c\" (UniqueName: \"kubernetes.io/projected/b25484eb-eb98-48f9-a3fa-676622944828-kube-api-access-j948c\") pod \"oauth-openshift-6fb796c88-8hqk6\" (UID: \"b25484eb-eb98-48f9-a3fa-676622944828\") " pod="openshift-authentication/oauth-openshift-6fb796c88-8hqk6" Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.387289 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b25484eb-eb98-48f9-a3fa-676622944828-audit-policies\") pod \"oauth-openshift-6fb796c88-8hqk6\" (UID: \"b25484eb-eb98-48f9-a3fa-676622944828\") " pod="openshift-authentication/oauth-openshift-6fb796c88-8hqk6" Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.387313 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b25484eb-eb98-48f9-a3fa-676622944828-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6fb796c88-8hqk6\" (UID: \"b25484eb-eb98-48f9-a3fa-676622944828\") " pod="openshift-authentication/oauth-openshift-6fb796c88-8hqk6" Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.387342 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/b25484eb-eb98-48f9-a3fa-676622944828-v4-0-config-system-service-ca\") pod \"oauth-openshift-6fb796c88-8hqk6\" (UID: \"b25484eb-eb98-48f9-a3fa-676622944828\") " pod="openshift-authentication/oauth-openshift-6fb796c88-8hqk6" Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.387362 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/b25484eb-eb98-48f9-a3fa-676622944828-v4-0-config-user-template-login\") pod \"oauth-openshift-6fb796c88-8hqk6\" (UID: \"b25484eb-eb98-48f9-a3fa-676622944828\") " pod="openshift-authentication/oauth-openshift-6fb796c88-8hqk6" Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.388130 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/b25484eb-eb98-48f9-a3fa-676622944828-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6fb796c88-8hqk6\" (UID: \"b25484eb-eb98-48f9-a3fa-676622944828\") " pod="openshift-authentication/oauth-openshift-6fb796c88-8hqk6" Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.389004 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b25484eb-eb98-48f9-a3fa-676622944828-audit-policies\") pod \"oauth-openshift-6fb796c88-8hqk6\" (UID: \"b25484eb-eb98-48f9-a3fa-676622944828\") " pod="openshift-authentication/oauth-openshift-6fb796c88-8hqk6" Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.389303 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/b25484eb-eb98-48f9-a3fa-676622944828-v4-0-config-system-service-ca\") pod \"oauth-openshift-6fb796c88-8hqk6\" (UID: \"b25484eb-eb98-48f9-a3fa-676622944828\") " pod="openshift-authentication/oauth-openshift-6fb796c88-8hqk6" Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.389601 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b25484eb-eb98-48f9-a3fa-676622944828-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6fb796c88-8hqk6\" (UID: \"b25484eb-eb98-48f9-a3fa-676622944828\") " pod="openshift-authentication/oauth-openshift-6fb796c88-8hqk6" Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.392885 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/b25484eb-eb98-48f9-a3fa-676622944828-v4-0-config-user-template-error\") pod \"oauth-openshift-6fb796c88-8hqk6\" (UID: \"b25484eb-eb98-48f9-a3fa-676622944828\") " pod="openshift-authentication/oauth-openshift-6fb796c88-8hqk6" Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.394065 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/b25484eb-eb98-48f9-a3fa-676622944828-v4-0-config-user-template-login\") pod \"oauth-openshift-6fb796c88-8hqk6\" (UID: \"b25484eb-eb98-48f9-a3fa-676622944828\") " pod="openshift-authentication/oauth-openshift-6fb796c88-8hqk6" Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.394425 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/b25484eb-eb98-48f9-a3fa-676622944828-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6fb796c88-8hqk6\" (UID: \"b25484eb-eb98-48f9-a3fa-676622944828\") " pod="openshift-authentication/oauth-openshift-6fb796c88-8hqk6" Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.394428 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/b25484eb-eb98-48f9-a3fa-676622944828-v4-0-config-system-session\") pod \"oauth-openshift-6fb796c88-8hqk6\" (UID: \"b25484eb-eb98-48f9-a3fa-676622944828\") " pod="openshift-authentication/oauth-openshift-6fb796c88-8hqk6" Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.394435 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/b25484eb-eb98-48f9-a3fa-676622944828-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6fb796c88-8hqk6\" (UID: \"b25484eb-eb98-48f9-a3fa-676622944828\") " pod="openshift-authentication/oauth-openshift-6fb796c88-8hqk6" Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.394651 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/b25484eb-eb98-48f9-a3fa-676622944828-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6fb796c88-8hqk6\" (UID: \"b25484eb-eb98-48f9-a3fa-676622944828\") " pod="openshift-authentication/oauth-openshift-6fb796c88-8hqk6" Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.399388 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/b25484eb-eb98-48f9-a3fa-676622944828-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6fb796c88-8hqk6\" (UID: \"b25484eb-eb98-48f9-a3fa-676622944828\") " pod="openshift-authentication/oauth-openshift-6fb796c88-8hqk6" Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.401484 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/b25484eb-eb98-48f9-a3fa-676622944828-v4-0-config-system-router-certs\") pod \"oauth-openshift-6fb796c88-8hqk6\" (UID: \"b25484eb-eb98-48f9-a3fa-676622944828\") " pod="openshift-authentication/oauth-openshift-6fb796c88-8hqk6" Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.404685 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j948c\" (UniqueName: \"kubernetes.io/projected/b25484eb-eb98-48f9-a3fa-676622944828-kube-api-access-j948c\") pod \"oauth-openshift-6fb796c88-8hqk6\" (UID: \"b25484eb-eb98-48f9-a3fa-676622944828\") " pod="openshift-authentication/oauth-openshift-6fb796c88-8hqk6" Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.420343 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-6fb796c88-8hqk6" Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.551488 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.666472 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.683623 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.759896 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.811684 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 07 16:06:38 crc kubenswrapper[4716]: I1207 16:06:38.958697 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 07 16:06:39 crc kubenswrapper[4716]: I1207 16:06:39.039901 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 07 16:06:39 crc kubenswrapper[4716]: I1207 16:06:39.111321 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 07 16:06:39 crc kubenswrapper[4716]: I1207 16:06:39.191253 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 07 16:06:39 crc kubenswrapper[4716]: I1207 16:06:39.220626 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 07 16:06:39 crc kubenswrapper[4716]: I1207 16:06:39.259150 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 07 16:06:39 crc kubenswrapper[4716]: I1207 16:06:39.266611 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 07 16:06:39 crc kubenswrapper[4716]: I1207 16:06:39.269033 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 07 16:06:39 crc kubenswrapper[4716]: I1207 16:06:39.300628 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 07 16:06:39 crc kubenswrapper[4716]: I1207 16:06:39.309109 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 07 16:06:39 crc kubenswrapper[4716]: I1207 16:06:39.357157 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 07 16:06:39 crc kubenswrapper[4716]: I1207 16:06:39.363012 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 07 16:06:39 crc kubenswrapper[4716]: I1207 16:06:39.371316 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 07 16:06:39 crc kubenswrapper[4716]: I1207 16:06:39.413323 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 07 16:06:39 crc kubenswrapper[4716]: I1207 16:06:39.472476 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 07 16:06:39 crc kubenswrapper[4716]: I1207 16:06:39.521691 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 07 16:06:39 crc kubenswrapper[4716]: I1207 16:06:39.564271 4716 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 07 16:06:39 crc kubenswrapper[4716]: I1207 16:06:39.564271 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 07 16:06:39 crc kubenswrapper[4716]: I1207 16:06:39.602429 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 07 16:06:39 crc kubenswrapper[4716]: I1207 16:06:39.623778 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 07 16:06:39 crc kubenswrapper[4716]: I1207 16:06:39.663845 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ffff2831-338e-423e-81f2-c8d3b5de7785" path="/var/lib/kubelet/pods/ffff2831-338e-423e-81f2-c8d3b5de7785/volumes" Dec 07 16:06:39 crc kubenswrapper[4716]: I1207 16:06:39.743282 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 07 16:06:39 crc kubenswrapper[4716]: I1207 16:06:39.798207 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 07 16:06:39 crc kubenswrapper[4716]: I1207 16:06:39.838336 4716 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 07 16:06:39 crc kubenswrapper[4716]: I1207 16:06:39.880557 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 07 16:06:39 crc kubenswrapper[4716]: I1207 16:06:39.907951 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 07 16:06:39 crc kubenswrapper[4716]: I1207 16:06:39.940434 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 07 16:06:39 crc kubenswrapper[4716]: I1207 16:06:39.954375 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 07 16:06:39 crc kubenswrapper[4716]: I1207 16:06:39.963036 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 07 16:06:39 crc kubenswrapper[4716]: I1207 16:06:39.970536 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 07 16:06:39 crc kubenswrapper[4716]: I1207 16:06:39.989441 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 07 16:06:40 crc kubenswrapper[4716]: I1207 16:06:40.039199 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 07 16:06:40 crc kubenswrapper[4716]: I1207 16:06:40.063718 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 07 16:06:40 crc kubenswrapper[4716]: I1207 16:06:40.081549 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 07 16:06:40 crc kubenswrapper[4716]: I1207 16:06:40.180416 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 07 16:06:40 crc kubenswrapper[4716]: I1207 16:06:40.257876 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 07 16:06:40 crc kubenswrapper[4716]: I1207 16:06:40.283558 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 07 16:06:40 crc kubenswrapper[4716]: I1207 16:06:40.288459 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 07 16:06:40 crc kubenswrapper[4716]: I1207 16:06:40.426969 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 07 16:06:40 crc kubenswrapper[4716]: I1207 16:06:40.451378 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 07 16:06:40 crc kubenswrapper[4716]: I1207 16:06:40.505330 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 07 16:06:40 crc kubenswrapper[4716]: I1207 16:06:40.610558 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 07 16:06:40 crc kubenswrapper[4716]: I1207 16:06:40.741853 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 07 16:06:40 crc kubenswrapper[4716]: I1207 16:06:40.982553 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 07 16:06:41 crc kubenswrapper[4716]: I1207 16:06:41.058438 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 07 16:06:41 crc kubenswrapper[4716]: I1207 16:06:41.106613 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 07 16:06:41 crc kubenswrapper[4716]: I1207 16:06:41.222117 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 07 16:06:41 crc kubenswrapper[4716]: I1207 16:06:41.252347 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 07 16:06:41 crc kubenswrapper[4716]: I1207 16:06:41.262422 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 07 16:06:41 crc kubenswrapper[4716]: I1207 16:06:41.265753 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 07 16:06:41 crc kubenswrapper[4716]: I1207 16:06:41.289669 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 07 16:06:41 crc kubenswrapper[4716]: I1207 16:06:41.341952 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 07 16:06:41 crc kubenswrapper[4716]: I1207 16:06:41.371990 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 07 16:06:41 crc kubenswrapper[4716]: I1207 16:06:41.387210 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 07 16:06:41 crc kubenswrapper[4716]: I1207 16:06:41.410185 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 07 16:06:41 crc kubenswrapper[4716]: I1207 16:06:41.447888 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 07 16:06:41 crc kubenswrapper[4716]: I1207 16:06:41.477250 4716 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 07 16:06:41 crc kubenswrapper[4716]: I1207 16:06:41.496214 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 07 16:06:41 crc kubenswrapper[4716]: I1207 16:06:41.497931 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 07 16:06:41 crc kubenswrapper[4716]: I1207 16:06:41.607201 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 07 16:06:41 crc kubenswrapper[4716]: I1207 16:06:41.614118 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 07 16:06:41 crc kubenswrapper[4716]: I1207 16:06:41.707809 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 07 16:06:41 crc kubenswrapper[4716]: I1207 16:06:41.834283 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 07 16:06:41 crc kubenswrapper[4716]: I1207 16:06:41.847318 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 07 16:06:41 crc kubenswrapper[4716]: I1207 16:06:41.879351 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 07 16:06:41 crc kubenswrapper[4716]: I1207 16:06:41.960122 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 07 16:06:42 crc kubenswrapper[4716]: I1207 16:06:42.053524 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 07 16:06:42 crc kubenswrapper[4716]: I1207 16:06:42.118187 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 07 16:06:42 crc kubenswrapper[4716]: I1207 16:06:42.181673 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 07 16:06:42 crc kubenswrapper[4716]: I1207 16:06:42.268814 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 07 16:06:42 crc kubenswrapper[4716]: I1207 16:06:42.330446 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 07 16:06:42 crc kubenswrapper[4716]: I1207 16:06:42.363596 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 07 16:06:42 crc kubenswrapper[4716]: I1207 16:06:42.478057 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 07 16:06:42 crc kubenswrapper[4716]: I1207 16:06:42.633854 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 07 16:06:42 crc kubenswrapper[4716]: I1207 16:06:42.661315 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 07 16:06:42 crc kubenswrapper[4716]: I1207 16:06:42.704410 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 07 16:06:42 crc kubenswrapper[4716]: I1207 16:06:42.781781 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 07 16:06:42 crc kubenswrapper[4716]: I1207 16:06:42.797912 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 07 16:06:42 crc kubenswrapper[4716]: I1207 16:06:42.824894 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 07 16:06:42 crc kubenswrapper[4716]: I1207 16:06:42.848387 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 07 16:06:42 crc kubenswrapper[4716]: I1207 16:06:42.872611 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 07 16:06:42 crc kubenswrapper[4716]: I1207 16:06:42.917978 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 07 16:06:42 crc kubenswrapper[4716]: I1207 16:06:42.949555 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 07 16:06:43 crc kubenswrapper[4716]: I1207 16:06:43.150240 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 07 16:06:43 crc kubenswrapper[4716]: I1207 16:06:43.307369 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 07 16:06:43 crc kubenswrapper[4716]: I1207 16:06:43.315218 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 07 16:06:43 crc kubenswrapper[4716]: I1207 16:06:43.329815 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 07 16:06:43 crc kubenswrapper[4716]: I1207 16:06:43.367498 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 07 16:06:43 crc kubenswrapper[4716]: I1207 16:06:43.412431 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 07 16:06:43 crc kubenswrapper[4716]: I1207 16:06:43.435566 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 07 16:06:43 crc kubenswrapper[4716]: I1207 16:06:43.446278 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-6fb796c88-8hqk6"] Dec 07 16:06:43 crc kubenswrapper[4716]: I1207 16:06:43.586050 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 07 16:06:43 crc kubenswrapper[4716]: I1207 16:06:43.595154 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 07 16:06:43 crc kubenswrapper[4716]: I1207 16:06:43.616374 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 07 16:06:43 crc kubenswrapper[4716]: I1207 16:06:43.621787 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 07 16:06:43 crc kubenswrapper[4716]: I1207 16:06:43.650491 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 07 16:06:43 crc kubenswrapper[4716]: I1207 16:06:43.679647 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 07 16:06:43 crc kubenswrapper[4716]: I1207 16:06:43.797484 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 07 16:06:43 crc kubenswrapper[4716]: I1207 16:06:43.811541 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 07 16:06:43 crc kubenswrapper[4716]: I1207 16:06:43.820944 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 07 16:06:43 crc kubenswrapper[4716]: I1207 16:06:43.826144 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 07 16:06:43 crc kubenswrapper[4716]: I1207 16:06:43.903008 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 07 16:06:44 crc kubenswrapper[4716]: I1207 16:06:44.000452 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-6fb796c88-8hqk6"] Dec 07 16:06:44 crc kubenswrapper[4716]: I1207 16:06:44.013687 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 07 16:06:44 crc kubenswrapper[4716]: I1207 16:06:44.025050 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 07 16:06:44 crc kubenswrapper[4716]: I1207 16:06:44.048204 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 07 16:06:44 crc kubenswrapper[4716]: I1207 16:06:44.064721 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 07 16:06:44 crc kubenswrapper[4716]: I1207 16:06:44.094212 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 07 16:06:44 crc kubenswrapper[4716]: I1207 16:06:44.114417 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 07 16:06:44 crc kubenswrapper[4716]: I1207 16:06:44.138582 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 07 16:06:44 crc kubenswrapper[4716]: I1207 16:06:44.171002 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 07 16:06:44 crc kubenswrapper[4716]: I1207 16:06:44.229901 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 07 16:06:44 crc kubenswrapper[4716]: I1207 16:06:44.237723 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 07 16:06:44 crc kubenswrapper[4716]: I1207 16:06:44.260522 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 07 16:06:44 crc kubenswrapper[4716]: I1207 16:06:44.280294 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 07 16:06:44 crc kubenswrapper[4716]: I1207 16:06:44.342368 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 07 16:06:44 crc kubenswrapper[4716]: I1207 16:06:44.345427 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 07 16:06:44 crc kubenswrapper[4716]: I1207 16:06:44.361162 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 07 16:06:44 crc kubenswrapper[4716]: I1207 16:06:44.371579 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 07 16:06:44 crc kubenswrapper[4716]: I1207 16:06:44.380132 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 07 16:06:44 crc kubenswrapper[4716]: I1207 16:06:44.407998 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 07 16:06:44 crc kubenswrapper[4716]: I1207 16:06:44.485824 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 07 16:06:44 crc kubenswrapper[4716]: I1207 16:06:44.515503 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 07 16:06:44 crc kubenswrapper[4716]: I1207 16:06:44.516438 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 07 16:06:44 crc kubenswrapper[4716]: I1207 16:06:44.544219 4716 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 07 16:06:44 crc kubenswrapper[4716]: I1207 16:06:44.544565 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://a8901ca0b7fa2ead1bf4a7fe6266ec940b74a4b8ecccfa7cf3881c75f7d7cdf6" gracePeriod=5 Dec 07 16:06:44 crc kubenswrapper[4716]: I1207 16:06:44.549377 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-6fb796c88-8hqk6" event={"ID":"b25484eb-eb98-48f9-a3fa-676622944828","Type":"ContainerStarted","Data":"474d3d8dbc9b6a05c6d25e9cfc646513d586b11b3e80269f9acd6a6b6ee49d9a"} Dec 07 16:06:44 crc kubenswrapper[4716]: I1207 16:06:44.549420 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-6fb796c88-8hqk6" event={"ID":"b25484eb-eb98-48f9-a3fa-676622944828","Type":"ContainerStarted","Data":"4d657a2a782eacf0b1a53be72f1aeba9d2c67920cd8b12c1107384c8a4f6fc24"} Dec 07 16:06:44 crc kubenswrapper[4716]: I1207 16:06:44.549554 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-6fb796c88-8hqk6" Dec 07 16:06:44 crc kubenswrapper[4716]: I1207 16:06:44.567709 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 07 16:06:44 crc kubenswrapper[4716]: I1207 16:06:44.570131 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 07 16:06:44 crc kubenswrapper[4716]: I1207 16:06:44.578567 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-6fb796c88-8hqk6" podStartSLOduration=54.578551254 podStartE2EDuration="54.578551254s" podCreationTimestamp="2025-12-07 16:05:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:06:44.576289403 +0000 UTC m=+267.266574325" watchObservedRunningTime="2025-12-07 16:06:44.578551254 +0000 UTC m=+267.268836166" Dec 07 16:06:44 crc kubenswrapper[4716]: I1207 16:06:44.597775 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 07 16:06:44 crc kubenswrapper[4716]: I1207 16:06:44.649361 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 07 16:06:44 crc kubenswrapper[4716]: I1207 16:06:44.708556 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-6fb796c88-8hqk6" Dec 07 16:06:44 crc kubenswrapper[4716]: I1207 16:06:44.725414 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 07 16:06:44 crc kubenswrapper[4716]: I1207 16:06:44.726713 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 07 16:06:44 crc kubenswrapper[4716]: I1207 16:06:44.753030 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 07 16:06:44 crc kubenswrapper[4716]: I1207 16:06:44.812778 4716 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 07 16:06:44 crc kubenswrapper[4716]: I1207 16:06:44.822884 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 07 16:06:44 crc kubenswrapper[4716]: I1207 16:06:44.911486 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 07 16:06:44 crc kubenswrapper[4716]: I1207 16:06:44.955091 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 07 16:06:45 crc kubenswrapper[4716]: I1207 16:06:45.152774 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 07 16:06:45 crc kubenswrapper[4716]: I1207 16:06:45.227309 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 07 16:06:45 crc kubenswrapper[4716]: I1207 16:06:45.234371 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 07 16:06:45 crc kubenswrapper[4716]: I1207 16:06:45.329170 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 07 16:06:45 crc kubenswrapper[4716]: I1207 16:06:45.596377 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 07 16:06:45 crc kubenswrapper[4716]: I1207 16:06:45.639866 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 07 16:06:45 crc kubenswrapper[4716]: I1207 16:06:45.663734 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 07 16:06:45 crc kubenswrapper[4716]: I1207 16:06:45.667026 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 07 16:06:45 crc kubenswrapper[4716]: I1207 16:06:45.829973 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 07 16:06:45 crc kubenswrapper[4716]: I1207 16:06:45.840627 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 07 16:06:45 crc kubenswrapper[4716]: I1207 16:06:45.852737 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 07 16:06:45 crc kubenswrapper[4716]: I1207 16:06:45.885596 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 07 16:06:45 crc kubenswrapper[4716]: I1207 16:06:45.910924 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 07 16:06:45 crc kubenswrapper[4716]: I1207 16:06:45.987615 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 07 16:06:46 crc kubenswrapper[4716]: I1207 16:06:46.241546 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 07 16:06:46 crc kubenswrapper[4716]: I1207 16:06:46.251721 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 07 16:06:46 crc kubenswrapper[4716]: I1207 16:06:46.448323 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 07 16:06:46 crc kubenswrapper[4716]: I1207 16:06:46.462282 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 07 16:06:46 crc kubenswrapper[4716]: I1207 16:06:46.482952 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 07 16:06:46 crc kubenswrapper[4716]: I1207 16:06:46.526252 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 07 16:06:46 crc kubenswrapper[4716]: I1207 16:06:46.552240 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 07 16:06:46 crc kubenswrapper[4716]: I1207 16:06:46.601330 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 07 16:06:46 crc kubenswrapper[4716]: I1207 16:06:46.749813 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 07 16:06:47 crc kubenswrapper[4716]: I1207 16:06:47.065396 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 07 16:06:47 crc kubenswrapper[4716]: I1207 16:06:47.109749 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 07 16:06:47 crc kubenswrapper[4716]: I1207 16:06:47.119584 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 07 16:06:47 crc kubenswrapper[4716]: I1207 16:06:47.167432 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 07 16:06:47 crc kubenswrapper[4716]: I1207 16:06:47.473583 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 07 16:06:47 crc kubenswrapper[4716]: I1207 16:06:47.531128 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 07 16:06:47 crc kubenswrapper[4716]: I1207 16:06:47.591861 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 07 16:06:47 crc kubenswrapper[4716]: I1207 16:06:47.598316 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 07 16:06:47 crc kubenswrapper[4716]: I1207 16:06:47.719391 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 07 16:06:47 crc kubenswrapper[4716]: I1207 16:06:47.726417 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 07 16:06:47 crc kubenswrapper[4716]: I1207 16:06:47.773998 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 07 16:06:48 crc kubenswrapper[4716]: I1207 16:06:48.020127 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 07 16:06:48 crc kubenswrapper[4716]: I1207 16:06:48.073373 4716 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 07 16:06:48 crc kubenswrapper[4716]: I1207 16:06:48.129303 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 07 16:06:48 crc kubenswrapper[4716]: I1207 16:06:48.196616 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 07 16:06:48 crc kubenswrapper[4716]: I1207 16:06:48.270862 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 07 16:06:48 crc kubenswrapper[4716]: I1207 16:06:48.297795 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 07 16:06:48 crc kubenswrapper[4716]: I1207 16:06:48.482447 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 07 16:06:48 crc kubenswrapper[4716]: I1207 16:06:48.706037 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 07 16:06:48 crc kubenswrapper[4716]: I1207 16:06:48.714801 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 07 16:06:48 crc kubenswrapper[4716]: I1207 16:06:48.814874 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 07 16:06:48 crc kubenswrapper[4716]: I1207 16:06:48.969050 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 07 16:06:49 crc kubenswrapper[4716]: I1207 16:06:49.388167 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 07 16:06:49 crc kubenswrapper[4716]: I1207 16:06:49.487738 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 07 16:06:49 crc kubenswrapper[4716]: I1207 16:06:49.826967 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 07 16:06:49 crc kubenswrapper[4716]: I1207 16:06:49.972888 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 07 16:06:50 crc kubenswrapper[4716]: I1207 16:06:50.129862 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 07 16:06:50 crc kubenswrapper[4716]: I1207 16:06:50.129923 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 07 16:06:50 crc kubenswrapper[4716]: I1207 16:06:50.137834 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 07 16:06:50 crc kubenswrapper[4716]: I1207 16:06:50.235361 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 07 16:06:50 crc kubenswrapper[4716]: I1207 16:06:50.235409 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 07 16:06:50 crc kubenswrapper[4716]: I1207 16:06:50.235440 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 07 16:06:50 crc kubenswrapper[4716]: I1207 16:06:50.235464 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 07 16:06:50 crc kubenswrapper[4716]: I1207 16:06:50.235512 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 07 16:06:50 crc kubenswrapper[4716]: I1207 16:06:50.235789 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 16:06:50 crc kubenswrapper[4716]: I1207 16:06:50.235799 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 16:06:50 crc kubenswrapper[4716]: I1207 16:06:50.235872 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 16:06:50 crc kubenswrapper[4716]: I1207 16:06:50.235846 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 16:06:50 crc kubenswrapper[4716]: I1207 16:06:50.247356 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 16:06:50 crc kubenswrapper[4716]: I1207 16:06:50.337140 4716 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Dec 07 16:06:50 crc kubenswrapper[4716]: I1207 16:06:50.337173 4716 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 07 16:06:50 crc kubenswrapper[4716]: I1207 16:06:50.337185 4716 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Dec 07 16:06:50 crc kubenswrapper[4716]: I1207 16:06:50.337192 4716 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 07 16:06:50 crc kubenswrapper[4716]: I1207 16:06:50.337202 4716 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Dec 07 16:06:50 crc kubenswrapper[4716]: I1207 16:06:50.409332 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 07 16:06:50 crc kubenswrapper[4716]: I1207 16:06:50.589248 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 07 16:06:50 crc kubenswrapper[4716]: I1207 16:06:50.589617 4716 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="a8901ca0b7fa2ead1bf4a7fe6266ec940b74a4b8ecccfa7cf3881c75f7d7cdf6" exitCode=137 Dec 07 16:06:50 crc kubenswrapper[4716]: I1207 16:06:50.589738 4716 scope.go:117] "RemoveContainer" containerID="a8901ca0b7fa2ead1bf4a7fe6266ec940b74a4b8ecccfa7cf3881c75f7d7cdf6" Dec 07 16:06:50 crc kubenswrapper[4716]: I1207 16:06:50.589711 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 07 16:06:50 crc kubenswrapper[4716]: I1207 16:06:50.616201 4716 scope.go:117] "RemoveContainer" containerID="a8901ca0b7fa2ead1bf4a7fe6266ec940b74a4b8ecccfa7cf3881c75f7d7cdf6" Dec 07 16:06:50 crc kubenswrapper[4716]: E1207 16:06:50.616717 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8901ca0b7fa2ead1bf4a7fe6266ec940b74a4b8ecccfa7cf3881c75f7d7cdf6\": container with ID starting with a8901ca0b7fa2ead1bf4a7fe6266ec940b74a4b8ecccfa7cf3881c75f7d7cdf6 not found: ID does not exist" containerID="a8901ca0b7fa2ead1bf4a7fe6266ec940b74a4b8ecccfa7cf3881c75f7d7cdf6" Dec 07 16:06:50 crc kubenswrapper[4716]: I1207 16:06:50.616777 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8901ca0b7fa2ead1bf4a7fe6266ec940b74a4b8ecccfa7cf3881c75f7d7cdf6"} err="failed to get container status \"a8901ca0b7fa2ead1bf4a7fe6266ec940b74a4b8ecccfa7cf3881c75f7d7cdf6\": rpc error: code = NotFound desc = could not find container \"a8901ca0b7fa2ead1bf4a7fe6266ec940b74a4b8ecccfa7cf3881c75f7d7cdf6\": container with ID starting with a8901ca0b7fa2ead1bf4a7fe6266ec940b74a4b8ecccfa7cf3881c75f7d7cdf6 not found: ID does not exist" Dec 07 16:06:50 crc kubenswrapper[4716]: I1207 16:06:50.968423 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 07 16:06:51 crc kubenswrapper[4716]: I1207 16:06:51.236107 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 07 16:06:51 crc kubenswrapper[4716]: I1207 16:06:51.675219 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Dec 07 16:06:54 crc kubenswrapper[4716]: I1207 16:06:54.319429 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-cd798"] Dec 07 16:06:54 crc kubenswrapper[4716]: I1207 16:06:54.320434 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-cd798" podUID="45c781c9-df95-4b6e-bed3-fc1f22f079a3" containerName="registry-server" containerID="cri-o://6b65905e72c05446c18b09994045d7aa88ed1c0439f6a0ae2b7517a3d474faf9" gracePeriod=30 Dec 07 16:06:54 crc kubenswrapper[4716]: I1207 16:06:54.326680 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-d92nn"] Dec 07 16:06:54 crc kubenswrapper[4716]: I1207 16:06:54.326901 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-d92nn" podUID="074b6ff7-7e20-4054-811d-34050c07e74e" containerName="registry-server" containerID="cri-o://c0b94112363f26c24a1f79cdb0548597273c2180248bb78e761a44282abcf9c0" gracePeriod=30 Dec 07 16:06:54 crc kubenswrapper[4716]: I1207 16:06:54.340820 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-v6899"] Dec 07 16:06:54 crc kubenswrapper[4716]: I1207 16:06:54.341236 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-v6899" podUID="7a9cc84d-c02b-466f-9c1c-c1ae60e697ef" containerName="marketplace-operator" containerID="cri-o://4d025fd1d535955cc38cb815b9eaf0aba61650c6a85b46e10fdbf4392568810a" gracePeriod=30 Dec 07 16:06:54 crc kubenswrapper[4716]: I1207 16:06:54.347562 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fnkww"] Dec 07 16:06:54 crc kubenswrapper[4716]: I1207 16:06:54.347803 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-fnkww" podUID="12e72f4d-2887-4c37-bf16-403f124d6f68" containerName="registry-server" containerID="cri-o://68b80714dde88883aa18e745470eb64be164168a72184a16f75fabfce0c638db" gracePeriod=30 Dec 07 16:06:54 crc kubenswrapper[4716]: I1207 16:06:54.360414 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hqtd7"] Dec 07 16:06:54 crc kubenswrapper[4716]: I1207 16:06:54.360661 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-hqtd7" podUID="86df58f0-8580-4c66-9667-4f7cc285f3a0" containerName="registry-server" containerID="cri-o://f42afff4ed726d86046eae352ef553422bcf740d40af3eb4cbb9bedea874e41c" gracePeriod=30 Dec 07 16:06:54 crc kubenswrapper[4716]: I1207 16:06:54.369130 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-dhjm4"] Dec 07 16:06:54 crc kubenswrapper[4716]: E1207 16:06:54.369359 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 07 16:06:54 crc kubenswrapper[4716]: I1207 16:06:54.369369 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 07 16:06:54 crc kubenswrapper[4716]: I1207 16:06:54.369449 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 07 16:06:54 crc kubenswrapper[4716]: I1207 16:06:54.369847 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-dhjm4" Dec 07 16:06:54 crc kubenswrapper[4716]: I1207 16:06:54.381479 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-dhjm4"] Dec 07 16:06:54 crc kubenswrapper[4716]: I1207 16:06:54.490370 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/845614d7-aa72-46de-b358-2e39f0209886-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-dhjm4\" (UID: \"845614d7-aa72-46de-b358-2e39f0209886\") " pod="openshift-marketplace/marketplace-operator-79b997595-dhjm4" Dec 07 16:06:54 crc kubenswrapper[4716]: I1207 16:06:54.490663 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c426m\" (UniqueName: \"kubernetes.io/projected/845614d7-aa72-46de-b358-2e39f0209886-kube-api-access-c426m\") pod \"marketplace-operator-79b997595-dhjm4\" (UID: \"845614d7-aa72-46de-b358-2e39f0209886\") " pod="openshift-marketplace/marketplace-operator-79b997595-dhjm4" Dec 07 16:06:54 crc kubenswrapper[4716]: I1207 16:06:54.490753 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/845614d7-aa72-46de-b358-2e39f0209886-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-dhjm4\" (UID: \"845614d7-aa72-46de-b358-2e39f0209886\") " pod="openshift-marketplace/marketplace-operator-79b997595-dhjm4" Dec 07 16:06:54 crc kubenswrapper[4716]: I1207 16:06:54.591504 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/845614d7-aa72-46de-b358-2e39f0209886-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-dhjm4\" (UID: \"845614d7-aa72-46de-b358-2e39f0209886\") " pod="openshift-marketplace/marketplace-operator-79b997595-dhjm4" Dec 07 16:06:54 crc kubenswrapper[4716]: I1207 16:06:54.591831 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/845614d7-aa72-46de-b358-2e39f0209886-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-dhjm4\" (UID: \"845614d7-aa72-46de-b358-2e39f0209886\") " pod="openshift-marketplace/marketplace-operator-79b997595-dhjm4" Dec 07 16:06:54 crc kubenswrapper[4716]: I1207 16:06:54.591883 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c426m\" (UniqueName: \"kubernetes.io/projected/845614d7-aa72-46de-b358-2e39f0209886-kube-api-access-c426m\") pod \"marketplace-operator-79b997595-dhjm4\" (UID: \"845614d7-aa72-46de-b358-2e39f0209886\") " pod="openshift-marketplace/marketplace-operator-79b997595-dhjm4" Dec 07 16:06:54 crc kubenswrapper[4716]: I1207 16:06:54.598702 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/845614d7-aa72-46de-b358-2e39f0209886-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-dhjm4\" (UID: \"845614d7-aa72-46de-b358-2e39f0209886\") " pod="openshift-marketplace/marketplace-operator-79b997595-dhjm4" Dec 07 16:06:54 crc kubenswrapper[4716]: I1207 16:06:54.604265 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/845614d7-aa72-46de-b358-2e39f0209886-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-dhjm4\" (UID: \"845614d7-aa72-46de-b358-2e39f0209886\") " pod="openshift-marketplace/marketplace-operator-79b997595-dhjm4" Dec 07 16:06:54 crc kubenswrapper[4716]: I1207 16:06:54.614483 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c426m\" (UniqueName: \"kubernetes.io/projected/845614d7-aa72-46de-b358-2e39f0209886-kube-api-access-c426m\") pod \"marketplace-operator-79b997595-dhjm4\" (UID: \"845614d7-aa72-46de-b358-2e39f0209886\") " pod="openshift-marketplace/marketplace-operator-79b997595-dhjm4" Dec 07 16:06:54 crc kubenswrapper[4716]: I1207 16:06:54.621330 4716 generic.go:334] "Generic (PLEG): container finished" podID="7a9cc84d-c02b-466f-9c1c-c1ae60e697ef" containerID="4d025fd1d535955cc38cb815b9eaf0aba61650c6a85b46e10fdbf4392568810a" exitCode=0 Dec 07 16:06:54 crc kubenswrapper[4716]: I1207 16:06:54.621586 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-v6899" event={"ID":"7a9cc84d-c02b-466f-9c1c-c1ae60e697ef","Type":"ContainerDied","Data":"4d025fd1d535955cc38cb815b9eaf0aba61650c6a85b46e10fdbf4392568810a"} Dec 07 16:06:54 crc kubenswrapper[4716]: I1207 16:06:54.625528 4716 generic.go:334] "Generic (PLEG): container finished" podID="45c781c9-df95-4b6e-bed3-fc1f22f079a3" containerID="6b65905e72c05446c18b09994045d7aa88ed1c0439f6a0ae2b7517a3d474faf9" exitCode=0 Dec 07 16:06:54 crc kubenswrapper[4716]: I1207 16:06:54.625622 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cd798" event={"ID":"45c781c9-df95-4b6e-bed3-fc1f22f079a3","Type":"ContainerDied","Data":"6b65905e72c05446c18b09994045d7aa88ed1c0439f6a0ae2b7517a3d474faf9"} Dec 07 16:06:54 crc kubenswrapper[4716]: I1207 16:06:54.628891 4716 generic.go:334] "Generic (PLEG): container finished" podID="12e72f4d-2887-4c37-bf16-403f124d6f68" containerID="68b80714dde88883aa18e745470eb64be164168a72184a16f75fabfce0c638db" exitCode=0 Dec 07 16:06:54 crc kubenswrapper[4716]: I1207 16:06:54.629004 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fnkww" event={"ID":"12e72f4d-2887-4c37-bf16-403f124d6f68","Type":"ContainerDied","Data":"68b80714dde88883aa18e745470eb64be164168a72184a16f75fabfce0c638db"} Dec 07 16:06:54 crc kubenswrapper[4716]: I1207 16:06:54.631583 4716 generic.go:334] "Generic (PLEG): container finished" podID="074b6ff7-7e20-4054-811d-34050c07e74e" containerID="c0b94112363f26c24a1f79cdb0548597273c2180248bb78e761a44282abcf9c0" exitCode=0 Dec 07 16:06:54 crc kubenswrapper[4716]: I1207 16:06:54.631643 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d92nn" event={"ID":"074b6ff7-7e20-4054-811d-34050c07e74e","Type":"ContainerDied","Data":"c0b94112363f26c24a1f79cdb0548597273c2180248bb78e761a44282abcf9c0"} Dec 07 16:06:54 crc kubenswrapper[4716]: I1207 16:06:54.633018 4716 generic.go:334] "Generic (PLEG): container finished" podID="86df58f0-8580-4c66-9667-4f7cc285f3a0" containerID="f42afff4ed726d86046eae352ef553422bcf740d40af3eb4cbb9bedea874e41c" exitCode=0 Dec 07 16:06:54 crc kubenswrapper[4716]: I1207 16:06:54.633040 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hqtd7" event={"ID":"86df58f0-8580-4c66-9667-4f7cc285f3a0","Type":"ContainerDied","Data":"f42afff4ed726d86046eae352ef553422bcf740d40af3eb4cbb9bedea874e41c"} Dec 07 16:06:54 crc kubenswrapper[4716]: I1207 16:06:54.800895 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-dhjm4" Dec 07 16:06:54 crc kubenswrapper[4716]: I1207 16:06:54.816382 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-v6899" Dec 07 16:06:54 crc kubenswrapper[4716]: I1207 16:06:54.823470 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fnkww" Dec 07 16:06:54 crc kubenswrapper[4716]: I1207 16:06:54.833603 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d92nn" Dec 07 16:06:54 crc kubenswrapper[4716]: I1207 16:06:54.843178 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hqtd7" Dec 07 16:06:54 crc kubenswrapper[4716]: I1207 16:06:54.844657 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cd798" Dec 07 16:06:54 crc kubenswrapper[4716]: I1207 16:06:54.996839 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gvh45\" (UniqueName: \"kubernetes.io/projected/074b6ff7-7e20-4054-811d-34050c07e74e-kube-api-access-gvh45\") pod \"074b6ff7-7e20-4054-811d-34050c07e74e\" (UID: \"074b6ff7-7e20-4054-811d-34050c07e74e\") " Dec 07 16:06:54 crc kubenswrapper[4716]: I1207 16:06:54.996889 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45c781c9-df95-4b6e-bed3-fc1f22f079a3-utilities\") pod \"45c781c9-df95-4b6e-bed3-fc1f22f079a3\" (UID: \"45c781c9-df95-4b6e-bed3-fc1f22f079a3\") " Dec 07 16:06:54 crc kubenswrapper[4716]: I1207 16:06:54.996918 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45c781c9-df95-4b6e-bed3-fc1f22f079a3-catalog-content\") pod \"45c781c9-df95-4b6e-bed3-fc1f22f079a3\" (UID: \"45c781c9-df95-4b6e-bed3-fc1f22f079a3\") " Dec 07 16:06:54 crc kubenswrapper[4716]: I1207 16:06:54.996939 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/12e72f4d-2887-4c37-bf16-403f124d6f68-catalog-content\") pod \"12e72f4d-2887-4c37-bf16-403f124d6f68\" (UID: \"12e72f4d-2887-4c37-bf16-403f124d6f68\") " Dec 07 16:06:54 crc kubenswrapper[4716]: I1207 16:06:54.996960 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/074b6ff7-7e20-4054-811d-34050c07e74e-utilities\") pod \"074b6ff7-7e20-4054-811d-34050c07e74e\" (UID: \"074b6ff7-7e20-4054-811d-34050c07e74e\") " Dec 07 16:06:54 crc kubenswrapper[4716]: I1207 16:06:54.997804 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/45c781c9-df95-4b6e-bed3-fc1f22f079a3-utilities" (OuterVolumeSpecName: "utilities") pod "45c781c9-df95-4b6e-bed3-fc1f22f079a3" (UID: "45c781c9-df95-4b6e-bed3-fc1f22f079a3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:06:54 crc kubenswrapper[4716]: I1207 16:06:54.997887 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7a9cc84d-c02b-466f-9c1c-c1ae60e697ef-marketplace-operator-metrics\") pod \"7a9cc84d-c02b-466f-9c1c-c1ae60e697ef\" (UID: \"7a9cc84d-c02b-466f-9c1c-c1ae60e697ef\") " Dec 07 16:06:54 crc kubenswrapper[4716]: I1207 16:06:54.997915 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wn54r\" (UniqueName: \"kubernetes.io/projected/45c781c9-df95-4b6e-bed3-fc1f22f079a3-kube-api-access-wn54r\") pod \"45c781c9-df95-4b6e-bed3-fc1f22f079a3\" (UID: \"45c781c9-df95-4b6e-bed3-fc1f22f079a3\") " Dec 07 16:06:54 crc kubenswrapper[4716]: I1207 16:06:54.998702 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/12e72f4d-2887-4c37-bf16-403f124d6f68-utilities\") pod \"12e72f4d-2887-4c37-bf16-403f124d6f68\" (UID: \"12e72f4d-2887-4c37-bf16-403f124d6f68\") " Dec 07 16:06:54 crc kubenswrapper[4716]: I1207 16:06:54.998730 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dmqh5\" (UniqueName: \"kubernetes.io/projected/7a9cc84d-c02b-466f-9c1c-c1ae60e697ef-kube-api-access-dmqh5\") pod \"7a9cc84d-c02b-466f-9c1c-c1ae60e697ef\" (UID: \"7a9cc84d-c02b-466f-9c1c-c1ae60e697ef\") " Dec 07 16:06:54 crc kubenswrapper[4716]: I1207 16:06:54.998753 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7a9cc84d-c02b-466f-9c1c-c1ae60e697ef-marketplace-trusted-ca\") pod \"7a9cc84d-c02b-466f-9c1c-c1ae60e697ef\" (UID: \"7a9cc84d-c02b-466f-9c1c-c1ae60e697ef\") " Dec 07 16:06:54 crc kubenswrapper[4716]: I1207 16:06:54.998785 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86df58f0-8580-4c66-9667-4f7cc285f3a0-catalog-content\") pod \"86df58f0-8580-4c66-9667-4f7cc285f3a0\" (UID: \"86df58f0-8580-4c66-9667-4f7cc285f3a0\") " Dec 07 16:06:54 crc kubenswrapper[4716]: I1207 16:06:54.998813 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/074b6ff7-7e20-4054-811d-34050c07e74e-catalog-content\") pod \"074b6ff7-7e20-4054-811d-34050c07e74e\" (UID: \"074b6ff7-7e20-4054-811d-34050c07e74e\") " Dec 07 16:06:54 crc kubenswrapper[4716]: I1207 16:06:54.998863 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qxznt\" (UniqueName: \"kubernetes.io/projected/86df58f0-8580-4c66-9667-4f7cc285f3a0-kube-api-access-qxznt\") pod \"86df58f0-8580-4c66-9667-4f7cc285f3a0\" (UID: \"86df58f0-8580-4c66-9667-4f7cc285f3a0\") " Dec 07 16:06:55 crc kubenswrapper[4716]: I1207 16:06:54.998141 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/074b6ff7-7e20-4054-811d-34050c07e74e-utilities" (OuterVolumeSpecName: "utilities") pod "074b6ff7-7e20-4054-811d-34050c07e74e" (UID: "074b6ff7-7e20-4054-811d-34050c07e74e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:06:55 crc kubenswrapper[4716]: I1207 16:06:54.999682 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/12e72f4d-2887-4c37-bf16-403f124d6f68-utilities" (OuterVolumeSpecName: "utilities") pod "12e72f4d-2887-4c37-bf16-403f124d6f68" (UID: "12e72f4d-2887-4c37-bf16-403f124d6f68"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:06:55 crc kubenswrapper[4716]: I1207 16:06:54.999742 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/86df58f0-8580-4c66-9667-4f7cc285f3a0-utilities" (OuterVolumeSpecName: "utilities") pod "86df58f0-8580-4c66-9667-4f7cc285f3a0" (UID: "86df58f0-8580-4c66-9667-4f7cc285f3a0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:06:55 crc kubenswrapper[4716]: I1207 16:06:55.000711 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a9cc84d-c02b-466f-9c1c-c1ae60e697ef-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "7a9cc84d-c02b-466f-9c1c-c1ae60e697ef" (UID: "7a9cc84d-c02b-466f-9c1c-c1ae60e697ef"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:06:55 crc kubenswrapper[4716]: I1207 16:06:55.000961 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45c781c9-df95-4b6e-bed3-fc1f22f079a3-kube-api-access-wn54r" (OuterVolumeSpecName: "kube-api-access-wn54r") pod "45c781c9-df95-4b6e-bed3-fc1f22f079a3" (UID: "45c781c9-df95-4b6e-bed3-fc1f22f079a3"). InnerVolumeSpecName "kube-api-access-wn54r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:06:55 crc kubenswrapper[4716]: I1207 16:06:54.998895 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86df58f0-8580-4c66-9667-4f7cc285f3a0-utilities\") pod \"86df58f0-8580-4c66-9667-4f7cc285f3a0\" (UID: \"86df58f0-8580-4c66-9667-4f7cc285f3a0\") " Dec 07 16:06:55 crc kubenswrapper[4716]: I1207 16:06:55.005026 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fx9hn\" (UniqueName: \"kubernetes.io/projected/12e72f4d-2887-4c37-bf16-403f124d6f68-kube-api-access-fx9hn\") pod \"12e72f4d-2887-4c37-bf16-403f124d6f68\" (UID: \"12e72f4d-2887-4c37-bf16-403f124d6f68\") " Dec 07 16:06:55 crc kubenswrapper[4716]: I1207 16:06:55.001587 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a9cc84d-c02b-466f-9c1c-c1ae60e697ef-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "7a9cc84d-c02b-466f-9c1c-c1ae60e697ef" (UID: "7a9cc84d-c02b-466f-9c1c-c1ae60e697ef"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:06:55 crc kubenswrapper[4716]: I1207 16:06:55.002709 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a9cc84d-c02b-466f-9c1c-c1ae60e697ef-kube-api-access-dmqh5" (OuterVolumeSpecName: "kube-api-access-dmqh5") pod "7a9cc84d-c02b-466f-9c1c-c1ae60e697ef" (UID: "7a9cc84d-c02b-466f-9c1c-c1ae60e697ef"). InnerVolumeSpecName "kube-api-access-dmqh5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:06:55 crc kubenswrapper[4716]: I1207 16:06:55.003115 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86df58f0-8580-4c66-9667-4f7cc285f3a0-kube-api-access-qxznt" (OuterVolumeSpecName: "kube-api-access-qxznt") pod "86df58f0-8580-4c66-9667-4f7cc285f3a0" (UID: "86df58f0-8580-4c66-9667-4f7cc285f3a0"). InnerVolumeSpecName "kube-api-access-qxznt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:06:55 crc kubenswrapper[4716]: I1207 16:06:55.005721 4716 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/074b6ff7-7e20-4054-811d-34050c07e74e-utilities\") on node \"crc\" DevicePath \"\"" Dec 07 16:06:55 crc kubenswrapper[4716]: I1207 16:06:55.005747 4716 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7a9cc84d-c02b-466f-9c1c-c1ae60e697ef-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 07 16:06:55 crc kubenswrapper[4716]: I1207 16:06:55.005759 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wn54r\" (UniqueName: \"kubernetes.io/projected/45c781c9-df95-4b6e-bed3-fc1f22f079a3-kube-api-access-wn54r\") on node \"crc\" DevicePath \"\"" Dec 07 16:06:55 crc kubenswrapper[4716]: I1207 16:06:55.005768 4716 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/12e72f4d-2887-4c37-bf16-403f124d6f68-utilities\") on node \"crc\" DevicePath \"\"" Dec 07 16:06:55 crc kubenswrapper[4716]: I1207 16:06:55.005777 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dmqh5\" (UniqueName: \"kubernetes.io/projected/7a9cc84d-c02b-466f-9c1c-c1ae60e697ef-kube-api-access-dmqh5\") on node \"crc\" DevicePath \"\"" Dec 07 16:06:55 crc kubenswrapper[4716]: I1207 16:06:55.005786 4716 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7a9cc84d-c02b-466f-9c1c-c1ae60e697ef-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 07 16:06:55 crc kubenswrapper[4716]: I1207 16:06:55.005795 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qxznt\" (UniqueName: \"kubernetes.io/projected/86df58f0-8580-4c66-9667-4f7cc285f3a0-kube-api-access-qxznt\") on node \"crc\" DevicePath \"\"" Dec 07 16:06:55 crc kubenswrapper[4716]: I1207 16:06:55.005835 4716 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86df58f0-8580-4c66-9667-4f7cc285f3a0-utilities\") on node \"crc\" DevicePath \"\"" Dec 07 16:06:55 crc kubenswrapper[4716]: I1207 16:06:55.005852 4716 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45c781c9-df95-4b6e-bed3-fc1f22f079a3-utilities\") on node \"crc\" DevicePath \"\"" Dec 07 16:06:55 crc kubenswrapper[4716]: I1207 16:06:55.009270 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12e72f4d-2887-4c37-bf16-403f124d6f68-kube-api-access-fx9hn" (OuterVolumeSpecName: "kube-api-access-fx9hn") pod "12e72f4d-2887-4c37-bf16-403f124d6f68" (UID: "12e72f4d-2887-4c37-bf16-403f124d6f68"). InnerVolumeSpecName "kube-api-access-fx9hn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:06:55 crc kubenswrapper[4716]: I1207 16:06:55.011193 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/074b6ff7-7e20-4054-811d-34050c07e74e-kube-api-access-gvh45" (OuterVolumeSpecName: "kube-api-access-gvh45") pod "074b6ff7-7e20-4054-811d-34050c07e74e" (UID: "074b6ff7-7e20-4054-811d-34050c07e74e"). InnerVolumeSpecName "kube-api-access-gvh45". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:06:55 crc kubenswrapper[4716]: I1207 16:06:55.026823 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/12e72f4d-2887-4c37-bf16-403f124d6f68-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "12e72f4d-2887-4c37-bf16-403f124d6f68" (UID: "12e72f4d-2887-4c37-bf16-403f124d6f68"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:06:55 crc kubenswrapper[4716]: I1207 16:06:55.053044 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/45c781c9-df95-4b6e-bed3-fc1f22f079a3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "45c781c9-df95-4b6e-bed3-fc1f22f079a3" (UID: "45c781c9-df95-4b6e-bed3-fc1f22f079a3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:06:55 crc kubenswrapper[4716]: I1207 16:06:55.064673 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/074b6ff7-7e20-4054-811d-34050c07e74e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "074b6ff7-7e20-4054-811d-34050c07e74e" (UID: "074b6ff7-7e20-4054-811d-34050c07e74e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:06:55 crc kubenswrapper[4716]: I1207 16:06:55.107170 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fx9hn\" (UniqueName: \"kubernetes.io/projected/12e72f4d-2887-4c37-bf16-403f124d6f68-kube-api-access-fx9hn\") on node \"crc\" DevicePath \"\"" Dec 07 16:06:55 crc kubenswrapper[4716]: I1207 16:06:55.107212 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gvh45\" (UniqueName: \"kubernetes.io/projected/074b6ff7-7e20-4054-811d-34050c07e74e-kube-api-access-gvh45\") on node \"crc\" DevicePath \"\"" Dec 07 16:06:55 crc kubenswrapper[4716]: I1207 16:06:55.107224 4716 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45c781c9-df95-4b6e-bed3-fc1f22f079a3-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 07 16:06:55 crc kubenswrapper[4716]: I1207 16:06:55.107257 4716 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/12e72f4d-2887-4c37-bf16-403f124d6f68-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 07 16:06:55 crc kubenswrapper[4716]: I1207 16:06:55.107266 4716 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/074b6ff7-7e20-4054-811d-34050c07e74e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 07 16:06:55 crc kubenswrapper[4716]: I1207 16:06:55.122138 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/86df58f0-8580-4c66-9667-4f7cc285f3a0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "86df58f0-8580-4c66-9667-4f7cc285f3a0" (UID: "86df58f0-8580-4c66-9667-4f7cc285f3a0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:06:55 crc kubenswrapper[4716]: I1207 16:06:55.208061 4716 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86df58f0-8580-4c66-9667-4f7cc285f3a0-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 07 16:06:55 crc kubenswrapper[4716]: I1207 16:06:55.212531 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-dhjm4"] Dec 07 16:06:55 crc kubenswrapper[4716]: I1207 16:06:55.639649 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hqtd7" event={"ID":"86df58f0-8580-4c66-9667-4f7cc285f3a0","Type":"ContainerDied","Data":"8e7d933557dd9d1c018fdba795e4eb9afed58d9ec6de915f5e58819fa3c8e0f6"} Dec 07 16:06:55 crc kubenswrapper[4716]: I1207 16:06:55.639714 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hqtd7" Dec 07 16:06:55 crc kubenswrapper[4716]: I1207 16:06:55.639964 4716 scope.go:117] "RemoveContainer" containerID="f42afff4ed726d86046eae352ef553422bcf740d40af3eb4cbb9bedea874e41c" Dec 07 16:06:55 crc kubenswrapper[4716]: I1207 16:06:55.641770 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-v6899" Dec 07 16:06:55 crc kubenswrapper[4716]: I1207 16:06:55.641789 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-v6899" event={"ID":"7a9cc84d-c02b-466f-9c1c-c1ae60e697ef","Type":"ContainerDied","Data":"beab624532a9706a8e95993e9414eb9c16bb7e0e9dd3a1c7c1d54e439d0fb3c4"} Dec 07 16:06:55 crc kubenswrapper[4716]: I1207 16:06:55.643100 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-dhjm4" event={"ID":"845614d7-aa72-46de-b358-2e39f0209886","Type":"ContainerStarted","Data":"fd8b8fa1467eb8c13d8a96095c5bb78da8675126d7d9e09d7fbe207313d607f4"} Dec 07 16:06:55 crc kubenswrapper[4716]: I1207 16:06:55.643134 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-dhjm4" event={"ID":"845614d7-aa72-46de-b358-2e39f0209886","Type":"ContainerStarted","Data":"b021afb4d156dc2ca38edc5d201df39c803cbc1b259aac76b2d85de09514ca53"} Dec 07 16:06:55 crc kubenswrapper[4716]: I1207 16:06:55.643699 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-dhjm4" Dec 07 16:06:55 crc kubenswrapper[4716]: I1207 16:06:55.644940 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cd798" event={"ID":"45c781c9-df95-4b6e-bed3-fc1f22f079a3","Type":"ContainerDied","Data":"50fe4fc8b66ae902be0b4d4206384da12fa977147ca08eb352aec9efbb98cdbb"} Dec 07 16:06:55 crc kubenswrapper[4716]: I1207 16:06:55.644994 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cd798" Dec 07 16:06:55 crc kubenswrapper[4716]: I1207 16:06:55.648717 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-dhjm4" Dec 07 16:06:55 crc kubenswrapper[4716]: I1207 16:06:55.650170 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fnkww" event={"ID":"12e72f4d-2887-4c37-bf16-403f124d6f68","Type":"ContainerDied","Data":"54ab04ae5b53e4cd7e6ad02374b0fe0ae54d32e9f91897ef9129dc7d43995839"} Dec 07 16:06:55 crc kubenswrapper[4716]: I1207 16:06:55.650234 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fnkww" Dec 07 16:06:55 crc kubenswrapper[4716]: I1207 16:06:55.655432 4716 scope.go:117] "RemoveContainer" containerID="3275604dc42563d85c354eea30930dacaef9cd157de6a6344d1812a419e8ca6f" Dec 07 16:06:55 crc kubenswrapper[4716]: I1207 16:06:55.662045 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d92nn" Dec 07 16:06:55 crc kubenswrapper[4716]: I1207 16:06:55.663909 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-dhjm4" podStartSLOduration=1.663894078 podStartE2EDuration="1.663894078s" podCreationTimestamp="2025-12-07 16:06:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:06:55.661882643 +0000 UTC m=+278.352167555" watchObservedRunningTime="2025-12-07 16:06:55.663894078 +0000 UTC m=+278.354178990" Dec 07 16:06:55 crc kubenswrapper[4716]: I1207 16:06:55.668476 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d92nn" event={"ID":"074b6ff7-7e20-4054-811d-34050c07e74e","Type":"ContainerDied","Data":"b4c6a8dad58efff23ae8852394652eb42a704a02d7de36cb77fc9ab189dd5221"} Dec 07 16:06:55 crc kubenswrapper[4716]: I1207 16:06:55.691099 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hqtd7"] Dec 07 16:06:55 crc kubenswrapper[4716]: I1207 16:06:55.695021 4716 scope.go:117] "RemoveContainer" containerID="ab9c2e3cac61dba0189c9c8d56eb34e0396604b46791fb46293ba31372950152" Dec 07 16:06:55 crc kubenswrapper[4716]: I1207 16:06:55.699911 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-hqtd7"] Dec 07 16:06:55 crc kubenswrapper[4716]: I1207 16:06:55.709639 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-v6899"] Dec 07 16:06:55 crc kubenswrapper[4716]: I1207 16:06:55.714425 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-v6899"] Dec 07 16:06:55 crc kubenswrapper[4716]: I1207 16:06:55.734428 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fnkww"] Dec 07 16:06:55 crc kubenswrapper[4716]: I1207 16:06:55.736520 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-fnkww"] Dec 07 16:06:55 crc kubenswrapper[4716]: I1207 16:06:55.739310 4716 scope.go:117] "RemoveContainer" containerID="4d025fd1d535955cc38cb815b9eaf0aba61650c6a85b46e10fdbf4392568810a" Dec 07 16:06:55 crc kubenswrapper[4716]: I1207 16:06:55.747059 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-cd798"] Dec 07 16:06:55 crc kubenswrapper[4716]: I1207 16:06:55.761625 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-cd798"] Dec 07 16:06:55 crc kubenswrapper[4716]: I1207 16:06:55.762578 4716 scope.go:117] "RemoveContainer" containerID="6b65905e72c05446c18b09994045d7aa88ed1c0439f6a0ae2b7517a3d474faf9" Dec 07 16:06:55 crc kubenswrapper[4716]: I1207 16:06:55.765499 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-d92nn"] Dec 07 16:06:55 crc kubenswrapper[4716]: I1207 16:06:55.779168 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-d92nn"] Dec 07 16:06:55 crc kubenswrapper[4716]: I1207 16:06:55.780219 4716 scope.go:117] "RemoveContainer" containerID="6470b74cdbdd34df5e88c457a42475515c1a03d0cd8e7303e5dee766cc8686b1" Dec 07 16:06:55 crc kubenswrapper[4716]: I1207 16:06:55.799968 4716 scope.go:117] "RemoveContainer" containerID="1063c57a5e4100898b9507ac1f578155291d67907813fbb7c8def6e42f86febc" Dec 07 16:06:55 crc kubenswrapper[4716]: I1207 16:06:55.815328 4716 scope.go:117] "RemoveContainer" containerID="68b80714dde88883aa18e745470eb64be164168a72184a16f75fabfce0c638db" Dec 07 16:06:55 crc kubenswrapper[4716]: I1207 16:06:55.831101 4716 scope.go:117] "RemoveContainer" containerID="f3f5034db9a73c532d54a8c3ce4db0a417a5c1d66d2b5480d7a35a3417011ddc" Dec 07 16:06:55 crc kubenswrapper[4716]: I1207 16:06:55.843587 4716 scope.go:117] "RemoveContainer" containerID="091f16f5b3fa9f7922dac579894eb8f866cf7bed963db6281ba83385ce94aa70" Dec 07 16:06:55 crc kubenswrapper[4716]: I1207 16:06:55.854820 4716 scope.go:117] "RemoveContainer" containerID="c0b94112363f26c24a1f79cdb0548597273c2180248bb78e761a44282abcf9c0" Dec 07 16:06:55 crc kubenswrapper[4716]: I1207 16:06:55.869297 4716 scope.go:117] "RemoveContainer" containerID="f1df7cd6cfed096919086b1925bf53a4f6988a5e169ca593ae033b764f245707" Dec 07 16:06:55 crc kubenswrapper[4716]: I1207 16:06:55.884651 4716 scope.go:117] "RemoveContainer" containerID="fdb1818e7e1be2170714671f582dd8cfac64141be6c0bc6c67d635a72e4e3aff" Dec 07 16:06:57 crc kubenswrapper[4716]: I1207 16:06:57.664168 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="074b6ff7-7e20-4054-811d-34050c07e74e" path="/var/lib/kubelet/pods/074b6ff7-7e20-4054-811d-34050c07e74e/volumes" Dec 07 16:06:57 crc kubenswrapper[4716]: I1207 16:06:57.664947 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12e72f4d-2887-4c37-bf16-403f124d6f68" path="/var/lib/kubelet/pods/12e72f4d-2887-4c37-bf16-403f124d6f68/volumes" Dec 07 16:06:57 crc kubenswrapper[4716]: I1207 16:06:57.665722 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45c781c9-df95-4b6e-bed3-fc1f22f079a3" path="/var/lib/kubelet/pods/45c781c9-df95-4b6e-bed3-fc1f22f079a3/volumes" Dec 07 16:06:57 crc kubenswrapper[4716]: I1207 16:06:57.667121 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a9cc84d-c02b-466f-9c1c-c1ae60e697ef" path="/var/lib/kubelet/pods/7a9cc84d-c02b-466f-9c1c-c1ae60e697ef/volumes" Dec 07 16:06:57 crc kubenswrapper[4716]: I1207 16:06:57.667721 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="86df58f0-8580-4c66-9667-4f7cc285f3a0" path="/var/lib/kubelet/pods/86df58f0-8580-4c66-9667-4f7cc285f3a0/volumes" Dec 07 16:07:17 crc kubenswrapper[4716]: I1207 16:07:17.454042 4716 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Dec 07 16:07:18 crc kubenswrapper[4716]: I1207 16:07:18.959946 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-tgn9w"] Dec 07 16:07:18 crc kubenswrapper[4716]: E1207 16:07:18.960427 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12e72f4d-2887-4c37-bf16-403f124d6f68" containerName="registry-server" Dec 07 16:07:18 crc kubenswrapper[4716]: I1207 16:07:18.960439 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="12e72f4d-2887-4c37-bf16-403f124d6f68" containerName="registry-server" Dec 07 16:07:18 crc kubenswrapper[4716]: E1207 16:07:18.960450 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86df58f0-8580-4c66-9667-4f7cc285f3a0" containerName="extract-utilities" Dec 07 16:07:18 crc kubenswrapper[4716]: I1207 16:07:18.960456 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="86df58f0-8580-4c66-9667-4f7cc285f3a0" containerName="extract-utilities" Dec 07 16:07:18 crc kubenswrapper[4716]: E1207 16:07:18.960462 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="074b6ff7-7e20-4054-811d-34050c07e74e" containerName="extract-content" Dec 07 16:07:18 crc kubenswrapper[4716]: I1207 16:07:18.960468 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="074b6ff7-7e20-4054-811d-34050c07e74e" containerName="extract-content" Dec 07 16:07:18 crc kubenswrapper[4716]: E1207 16:07:18.960477 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86df58f0-8580-4c66-9667-4f7cc285f3a0" containerName="extract-content" Dec 07 16:07:18 crc kubenswrapper[4716]: I1207 16:07:18.960486 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="86df58f0-8580-4c66-9667-4f7cc285f3a0" containerName="extract-content" Dec 07 16:07:18 crc kubenswrapper[4716]: E1207 16:07:18.960497 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12e72f4d-2887-4c37-bf16-403f124d6f68" containerName="extract-utilities" Dec 07 16:07:18 crc kubenswrapper[4716]: I1207 16:07:18.960504 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="12e72f4d-2887-4c37-bf16-403f124d6f68" containerName="extract-utilities" Dec 07 16:07:18 crc kubenswrapper[4716]: E1207 16:07:18.960513 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86df58f0-8580-4c66-9667-4f7cc285f3a0" containerName="registry-server" Dec 07 16:07:18 crc kubenswrapper[4716]: I1207 16:07:18.960518 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="86df58f0-8580-4c66-9667-4f7cc285f3a0" containerName="registry-server" Dec 07 16:07:18 crc kubenswrapper[4716]: E1207 16:07:18.960531 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="074b6ff7-7e20-4054-811d-34050c07e74e" containerName="registry-server" Dec 07 16:07:18 crc kubenswrapper[4716]: I1207 16:07:18.960536 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="074b6ff7-7e20-4054-811d-34050c07e74e" containerName="registry-server" Dec 07 16:07:18 crc kubenswrapper[4716]: E1207 16:07:18.960543 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45c781c9-df95-4b6e-bed3-fc1f22f079a3" containerName="extract-content" Dec 07 16:07:18 crc kubenswrapper[4716]: I1207 16:07:18.960549 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="45c781c9-df95-4b6e-bed3-fc1f22f079a3" containerName="extract-content" Dec 07 16:07:18 crc kubenswrapper[4716]: E1207 16:07:18.960555 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a9cc84d-c02b-466f-9c1c-c1ae60e697ef" containerName="marketplace-operator" Dec 07 16:07:18 crc kubenswrapper[4716]: I1207 16:07:18.960560 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a9cc84d-c02b-466f-9c1c-c1ae60e697ef" containerName="marketplace-operator" Dec 07 16:07:18 crc kubenswrapper[4716]: E1207 16:07:18.960569 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45c781c9-df95-4b6e-bed3-fc1f22f079a3" containerName="extract-utilities" Dec 07 16:07:18 crc kubenswrapper[4716]: I1207 16:07:18.960576 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="45c781c9-df95-4b6e-bed3-fc1f22f079a3" containerName="extract-utilities" Dec 07 16:07:18 crc kubenswrapper[4716]: E1207 16:07:18.960587 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12e72f4d-2887-4c37-bf16-403f124d6f68" containerName="extract-content" Dec 07 16:07:18 crc kubenswrapper[4716]: I1207 16:07:18.960595 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="12e72f4d-2887-4c37-bf16-403f124d6f68" containerName="extract-content" Dec 07 16:07:18 crc kubenswrapper[4716]: E1207 16:07:18.960605 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="074b6ff7-7e20-4054-811d-34050c07e74e" containerName="extract-utilities" Dec 07 16:07:18 crc kubenswrapper[4716]: I1207 16:07:18.960612 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="074b6ff7-7e20-4054-811d-34050c07e74e" containerName="extract-utilities" Dec 07 16:07:18 crc kubenswrapper[4716]: E1207 16:07:18.960621 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45c781c9-df95-4b6e-bed3-fc1f22f079a3" containerName="registry-server" Dec 07 16:07:18 crc kubenswrapper[4716]: I1207 16:07:18.960628 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="45c781c9-df95-4b6e-bed3-fc1f22f079a3" containerName="registry-server" Dec 07 16:07:18 crc kubenswrapper[4716]: I1207 16:07:18.960732 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="12e72f4d-2887-4c37-bf16-403f124d6f68" containerName="registry-server" Dec 07 16:07:18 crc kubenswrapper[4716]: I1207 16:07:18.960745 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="074b6ff7-7e20-4054-811d-34050c07e74e" containerName="registry-server" Dec 07 16:07:18 crc kubenswrapper[4716]: I1207 16:07:18.960752 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a9cc84d-c02b-466f-9c1c-c1ae60e697ef" containerName="marketplace-operator" Dec 07 16:07:18 crc kubenswrapper[4716]: I1207 16:07:18.960761 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="45c781c9-df95-4b6e-bed3-fc1f22f079a3" containerName="registry-server" Dec 07 16:07:18 crc kubenswrapper[4716]: I1207 16:07:18.960771 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="86df58f0-8580-4c66-9667-4f7cc285f3a0" containerName="registry-server" Dec 07 16:07:18 crc kubenswrapper[4716]: I1207 16:07:18.961503 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tgn9w" Dec 07 16:07:18 crc kubenswrapper[4716]: I1207 16:07:18.965109 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 07 16:07:18 crc kubenswrapper[4716]: I1207 16:07:18.971789 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tgn9w"] Dec 07 16:07:19 crc kubenswrapper[4716]: I1207 16:07:19.081960 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc342361-554a-41c0-a512-06e046ff03eb-utilities\") pod \"certified-operators-tgn9w\" (UID: \"fc342361-554a-41c0-a512-06e046ff03eb\") " pod="openshift-marketplace/certified-operators-tgn9w" Dec 07 16:07:19 crc kubenswrapper[4716]: I1207 16:07:19.082048 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wt4jk\" (UniqueName: \"kubernetes.io/projected/fc342361-554a-41c0-a512-06e046ff03eb-kube-api-access-wt4jk\") pod \"certified-operators-tgn9w\" (UID: \"fc342361-554a-41c0-a512-06e046ff03eb\") " pod="openshift-marketplace/certified-operators-tgn9w" Dec 07 16:07:19 crc kubenswrapper[4716]: I1207 16:07:19.082111 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc342361-554a-41c0-a512-06e046ff03eb-catalog-content\") pod \"certified-operators-tgn9w\" (UID: \"fc342361-554a-41c0-a512-06e046ff03eb\") " pod="openshift-marketplace/certified-operators-tgn9w" Dec 07 16:07:19 crc kubenswrapper[4716]: I1207 16:07:19.147061 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-f7pqt"] Dec 07 16:07:19 crc kubenswrapper[4716]: I1207 16:07:19.148209 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f7pqt" Dec 07 16:07:19 crc kubenswrapper[4716]: I1207 16:07:19.150160 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 07 16:07:19 crc kubenswrapper[4716]: I1207 16:07:19.162518 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-f7pqt"] Dec 07 16:07:19 crc kubenswrapper[4716]: I1207 16:07:19.184467 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc342361-554a-41c0-a512-06e046ff03eb-utilities\") pod \"certified-operators-tgn9w\" (UID: \"fc342361-554a-41c0-a512-06e046ff03eb\") " pod="openshift-marketplace/certified-operators-tgn9w" Dec 07 16:07:19 crc kubenswrapper[4716]: I1207 16:07:19.184515 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wt4jk\" (UniqueName: \"kubernetes.io/projected/fc342361-554a-41c0-a512-06e046ff03eb-kube-api-access-wt4jk\") pod \"certified-operators-tgn9w\" (UID: \"fc342361-554a-41c0-a512-06e046ff03eb\") " pod="openshift-marketplace/certified-operators-tgn9w" Dec 07 16:07:19 crc kubenswrapper[4716]: I1207 16:07:19.184551 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc342361-554a-41c0-a512-06e046ff03eb-catalog-content\") pod \"certified-operators-tgn9w\" (UID: \"fc342361-554a-41c0-a512-06e046ff03eb\") " pod="openshift-marketplace/certified-operators-tgn9w" Dec 07 16:07:19 crc kubenswrapper[4716]: I1207 16:07:19.185027 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc342361-554a-41c0-a512-06e046ff03eb-utilities\") pod \"certified-operators-tgn9w\" (UID: \"fc342361-554a-41c0-a512-06e046ff03eb\") " pod="openshift-marketplace/certified-operators-tgn9w" Dec 07 16:07:19 crc kubenswrapper[4716]: I1207 16:07:19.186597 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc342361-554a-41c0-a512-06e046ff03eb-catalog-content\") pod \"certified-operators-tgn9w\" (UID: \"fc342361-554a-41c0-a512-06e046ff03eb\") " pod="openshift-marketplace/certified-operators-tgn9w" Dec 07 16:07:19 crc kubenswrapper[4716]: I1207 16:07:19.205901 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wt4jk\" (UniqueName: \"kubernetes.io/projected/fc342361-554a-41c0-a512-06e046ff03eb-kube-api-access-wt4jk\") pod \"certified-operators-tgn9w\" (UID: \"fc342361-554a-41c0-a512-06e046ff03eb\") " pod="openshift-marketplace/certified-operators-tgn9w" Dec 07 16:07:19 crc kubenswrapper[4716]: I1207 16:07:19.283297 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tgn9w" Dec 07 16:07:19 crc kubenswrapper[4716]: I1207 16:07:19.285539 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6v6mm\" (UniqueName: \"kubernetes.io/projected/0f030783-1303-4815-8ac1-fbecf8e036fe-kube-api-access-6v6mm\") pod \"redhat-marketplace-f7pqt\" (UID: \"0f030783-1303-4815-8ac1-fbecf8e036fe\") " pod="openshift-marketplace/redhat-marketplace-f7pqt" Dec 07 16:07:19 crc kubenswrapper[4716]: I1207 16:07:19.285591 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f030783-1303-4815-8ac1-fbecf8e036fe-utilities\") pod \"redhat-marketplace-f7pqt\" (UID: \"0f030783-1303-4815-8ac1-fbecf8e036fe\") " pod="openshift-marketplace/redhat-marketplace-f7pqt" Dec 07 16:07:19 crc kubenswrapper[4716]: I1207 16:07:19.285610 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f030783-1303-4815-8ac1-fbecf8e036fe-catalog-content\") pod \"redhat-marketplace-f7pqt\" (UID: \"0f030783-1303-4815-8ac1-fbecf8e036fe\") " pod="openshift-marketplace/redhat-marketplace-f7pqt" Dec 07 16:07:19 crc kubenswrapper[4716]: I1207 16:07:19.386231 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f030783-1303-4815-8ac1-fbecf8e036fe-utilities\") pod \"redhat-marketplace-f7pqt\" (UID: \"0f030783-1303-4815-8ac1-fbecf8e036fe\") " pod="openshift-marketplace/redhat-marketplace-f7pqt" Dec 07 16:07:19 crc kubenswrapper[4716]: I1207 16:07:19.386454 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f030783-1303-4815-8ac1-fbecf8e036fe-catalog-content\") pod \"redhat-marketplace-f7pqt\" (UID: \"0f030783-1303-4815-8ac1-fbecf8e036fe\") " pod="openshift-marketplace/redhat-marketplace-f7pqt" Dec 07 16:07:19 crc kubenswrapper[4716]: I1207 16:07:19.386521 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6v6mm\" (UniqueName: \"kubernetes.io/projected/0f030783-1303-4815-8ac1-fbecf8e036fe-kube-api-access-6v6mm\") pod \"redhat-marketplace-f7pqt\" (UID: \"0f030783-1303-4815-8ac1-fbecf8e036fe\") " pod="openshift-marketplace/redhat-marketplace-f7pqt" Dec 07 16:07:19 crc kubenswrapper[4716]: I1207 16:07:19.386909 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f030783-1303-4815-8ac1-fbecf8e036fe-utilities\") pod \"redhat-marketplace-f7pqt\" (UID: \"0f030783-1303-4815-8ac1-fbecf8e036fe\") " pod="openshift-marketplace/redhat-marketplace-f7pqt" Dec 07 16:07:19 crc kubenswrapper[4716]: I1207 16:07:19.387110 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f030783-1303-4815-8ac1-fbecf8e036fe-catalog-content\") pod \"redhat-marketplace-f7pqt\" (UID: \"0f030783-1303-4815-8ac1-fbecf8e036fe\") " pod="openshift-marketplace/redhat-marketplace-f7pqt" Dec 07 16:07:19 crc kubenswrapper[4716]: I1207 16:07:19.409182 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6v6mm\" (UniqueName: \"kubernetes.io/projected/0f030783-1303-4815-8ac1-fbecf8e036fe-kube-api-access-6v6mm\") pod \"redhat-marketplace-f7pqt\" (UID: \"0f030783-1303-4815-8ac1-fbecf8e036fe\") " pod="openshift-marketplace/redhat-marketplace-f7pqt" Dec 07 16:07:19 crc kubenswrapper[4716]: I1207 16:07:19.466292 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f7pqt" Dec 07 16:07:19 crc kubenswrapper[4716]: I1207 16:07:19.474711 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tgn9w"] Dec 07 16:07:19 crc kubenswrapper[4716]: I1207 16:07:19.666042 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-f7pqt"] Dec 07 16:07:19 crc kubenswrapper[4716]: W1207 16:07:19.671287 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0f030783_1303_4815_8ac1_fbecf8e036fe.slice/crio-5b53d3c67b862b8e7b0a341d2b7ee2d702c2c3a5e182522bc1e6a8f6ec3b1574 WatchSource:0}: Error finding container 5b53d3c67b862b8e7b0a341d2b7ee2d702c2c3a5e182522bc1e6a8f6ec3b1574: Status 404 returned error can't find the container with id 5b53d3c67b862b8e7b0a341d2b7ee2d702c2c3a5e182522bc1e6a8f6ec3b1574 Dec 07 16:07:19 crc kubenswrapper[4716]: I1207 16:07:19.783016 4716 generic.go:334] "Generic (PLEG): container finished" podID="fc342361-554a-41c0-a512-06e046ff03eb" containerID="44944f9fdc896ee51f61e61fc4564829eb978c49a220861d24903cf22fafc2f4" exitCode=0 Dec 07 16:07:19 crc kubenswrapper[4716]: I1207 16:07:19.783108 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tgn9w" event={"ID":"fc342361-554a-41c0-a512-06e046ff03eb","Type":"ContainerDied","Data":"44944f9fdc896ee51f61e61fc4564829eb978c49a220861d24903cf22fafc2f4"} Dec 07 16:07:19 crc kubenswrapper[4716]: I1207 16:07:19.783144 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tgn9w" event={"ID":"fc342361-554a-41c0-a512-06e046ff03eb","Type":"ContainerStarted","Data":"348ebb34833a2bd68977db86f0378030a4fe0947abe4a701ffcad54c09ecfa61"} Dec 07 16:07:19 crc kubenswrapper[4716]: I1207 16:07:19.785707 4716 generic.go:334] "Generic (PLEG): container finished" podID="0f030783-1303-4815-8ac1-fbecf8e036fe" containerID="952bce1ab564fa0fc418b3cb1f16f0dafaacd023a91e3735cf01aa6b4b25f92d" exitCode=0 Dec 07 16:07:19 crc kubenswrapper[4716]: I1207 16:07:19.785758 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f7pqt" event={"ID":"0f030783-1303-4815-8ac1-fbecf8e036fe","Type":"ContainerDied","Data":"952bce1ab564fa0fc418b3cb1f16f0dafaacd023a91e3735cf01aa6b4b25f92d"} Dec 07 16:07:19 crc kubenswrapper[4716]: I1207 16:07:19.785787 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f7pqt" event={"ID":"0f030783-1303-4815-8ac1-fbecf8e036fe","Type":"ContainerStarted","Data":"5b53d3c67b862b8e7b0a341d2b7ee2d702c2c3a5e182522bc1e6a8f6ec3b1574"} Dec 07 16:07:21 crc kubenswrapper[4716]: I1207 16:07:21.549358 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-nkjl4"] Dec 07 16:07:21 crc kubenswrapper[4716]: I1207 16:07:21.551126 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nkjl4" Dec 07 16:07:21 crc kubenswrapper[4716]: I1207 16:07:21.552995 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 07 16:07:21 crc kubenswrapper[4716]: I1207 16:07:21.564774 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nkjl4"] Dec 07 16:07:21 crc kubenswrapper[4716]: I1207 16:07:21.712621 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af948d72-6e6c-42d9-9d0b-1821d7f47176-catalog-content\") pod \"redhat-operators-nkjl4\" (UID: \"af948d72-6e6c-42d9-9d0b-1821d7f47176\") " pod="openshift-marketplace/redhat-operators-nkjl4" Dec 07 16:07:21 crc kubenswrapper[4716]: I1207 16:07:21.712686 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af948d72-6e6c-42d9-9d0b-1821d7f47176-utilities\") pod \"redhat-operators-nkjl4\" (UID: \"af948d72-6e6c-42d9-9d0b-1821d7f47176\") " pod="openshift-marketplace/redhat-operators-nkjl4" Dec 07 16:07:21 crc kubenswrapper[4716]: I1207 16:07:21.712725 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vm6gc\" (UniqueName: \"kubernetes.io/projected/af948d72-6e6c-42d9-9d0b-1821d7f47176-kube-api-access-vm6gc\") pod \"redhat-operators-nkjl4\" (UID: \"af948d72-6e6c-42d9-9d0b-1821d7f47176\") " pod="openshift-marketplace/redhat-operators-nkjl4" Dec 07 16:07:21 crc kubenswrapper[4716]: I1207 16:07:21.749666 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vmnk2"] Dec 07 16:07:21 crc kubenswrapper[4716]: I1207 16:07:21.751213 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vmnk2" Dec 07 16:07:21 crc kubenswrapper[4716]: I1207 16:07:21.755707 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 07 16:07:21 crc kubenswrapper[4716]: I1207 16:07:21.760398 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vmnk2"] Dec 07 16:07:21 crc kubenswrapper[4716]: I1207 16:07:21.809373 4716 generic.go:334] "Generic (PLEG): container finished" podID="fc342361-554a-41c0-a512-06e046ff03eb" containerID="1afcbaf51dadcbe854f6b66fca5ea85801b4fc2a33ae7e86a3d56d884785b284" exitCode=0 Dec 07 16:07:21 crc kubenswrapper[4716]: I1207 16:07:21.809868 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tgn9w" event={"ID":"fc342361-554a-41c0-a512-06e046ff03eb","Type":"ContainerDied","Data":"1afcbaf51dadcbe854f6b66fca5ea85801b4fc2a33ae7e86a3d56d884785b284"} Dec 07 16:07:21 crc kubenswrapper[4716]: I1207 16:07:21.813303 4716 generic.go:334] "Generic (PLEG): container finished" podID="0f030783-1303-4815-8ac1-fbecf8e036fe" containerID="d2d6b1e7f2ca356a2d5719cfd6c316fb9f4258620f4ff285ee6fb133ccb60f01" exitCode=0 Dec 07 16:07:21 crc kubenswrapper[4716]: I1207 16:07:21.813338 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f7pqt" event={"ID":"0f030783-1303-4815-8ac1-fbecf8e036fe","Type":"ContainerDied","Data":"d2d6b1e7f2ca356a2d5719cfd6c316fb9f4258620f4ff285ee6fb133ccb60f01"} Dec 07 16:07:21 crc kubenswrapper[4716]: I1207 16:07:21.813507 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vm6gc\" (UniqueName: \"kubernetes.io/projected/af948d72-6e6c-42d9-9d0b-1821d7f47176-kube-api-access-vm6gc\") pod \"redhat-operators-nkjl4\" (UID: \"af948d72-6e6c-42d9-9d0b-1821d7f47176\") " pod="openshift-marketplace/redhat-operators-nkjl4" Dec 07 16:07:21 crc kubenswrapper[4716]: I1207 16:07:21.813571 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af948d72-6e6c-42d9-9d0b-1821d7f47176-catalog-content\") pod \"redhat-operators-nkjl4\" (UID: \"af948d72-6e6c-42d9-9d0b-1821d7f47176\") " pod="openshift-marketplace/redhat-operators-nkjl4" Dec 07 16:07:21 crc kubenswrapper[4716]: I1207 16:07:21.813625 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af948d72-6e6c-42d9-9d0b-1821d7f47176-utilities\") pod \"redhat-operators-nkjl4\" (UID: \"af948d72-6e6c-42d9-9d0b-1821d7f47176\") " pod="openshift-marketplace/redhat-operators-nkjl4" Dec 07 16:07:21 crc kubenswrapper[4716]: I1207 16:07:21.814663 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af948d72-6e6c-42d9-9d0b-1821d7f47176-catalog-content\") pod \"redhat-operators-nkjl4\" (UID: \"af948d72-6e6c-42d9-9d0b-1821d7f47176\") " pod="openshift-marketplace/redhat-operators-nkjl4" Dec 07 16:07:21 crc kubenswrapper[4716]: I1207 16:07:21.822476 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af948d72-6e6c-42d9-9d0b-1821d7f47176-utilities\") pod \"redhat-operators-nkjl4\" (UID: \"af948d72-6e6c-42d9-9d0b-1821d7f47176\") " pod="openshift-marketplace/redhat-operators-nkjl4" Dec 07 16:07:21 crc kubenswrapper[4716]: I1207 16:07:21.841512 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vm6gc\" (UniqueName: \"kubernetes.io/projected/af948d72-6e6c-42d9-9d0b-1821d7f47176-kube-api-access-vm6gc\") pod \"redhat-operators-nkjl4\" (UID: \"af948d72-6e6c-42d9-9d0b-1821d7f47176\") " pod="openshift-marketplace/redhat-operators-nkjl4" Dec 07 16:07:21 crc kubenswrapper[4716]: I1207 16:07:21.865121 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nkjl4" Dec 07 16:07:21 crc kubenswrapper[4716]: I1207 16:07:21.916375 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33db3f8a-1f66-44c4-b563-2f07dbc8267e-catalog-content\") pod \"community-operators-vmnk2\" (UID: \"33db3f8a-1f66-44c4-b563-2f07dbc8267e\") " pod="openshift-marketplace/community-operators-vmnk2" Dec 07 16:07:21 crc kubenswrapper[4716]: I1207 16:07:21.916777 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33db3f8a-1f66-44c4-b563-2f07dbc8267e-utilities\") pod \"community-operators-vmnk2\" (UID: \"33db3f8a-1f66-44c4-b563-2f07dbc8267e\") " pod="openshift-marketplace/community-operators-vmnk2" Dec 07 16:07:21 crc kubenswrapper[4716]: I1207 16:07:21.916842 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99qxt\" (UniqueName: \"kubernetes.io/projected/33db3f8a-1f66-44c4-b563-2f07dbc8267e-kube-api-access-99qxt\") pod \"community-operators-vmnk2\" (UID: \"33db3f8a-1f66-44c4-b563-2f07dbc8267e\") " pod="openshift-marketplace/community-operators-vmnk2" Dec 07 16:07:22 crc kubenswrapper[4716]: I1207 16:07:22.017907 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33db3f8a-1f66-44c4-b563-2f07dbc8267e-utilities\") pod \"community-operators-vmnk2\" (UID: \"33db3f8a-1f66-44c4-b563-2f07dbc8267e\") " pod="openshift-marketplace/community-operators-vmnk2" Dec 07 16:07:22 crc kubenswrapper[4716]: I1207 16:07:22.017981 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99qxt\" (UniqueName: \"kubernetes.io/projected/33db3f8a-1f66-44c4-b563-2f07dbc8267e-kube-api-access-99qxt\") pod \"community-operators-vmnk2\" (UID: \"33db3f8a-1f66-44c4-b563-2f07dbc8267e\") " pod="openshift-marketplace/community-operators-vmnk2" Dec 07 16:07:22 crc kubenswrapper[4716]: I1207 16:07:22.018020 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33db3f8a-1f66-44c4-b563-2f07dbc8267e-catalog-content\") pod \"community-operators-vmnk2\" (UID: \"33db3f8a-1f66-44c4-b563-2f07dbc8267e\") " pod="openshift-marketplace/community-operators-vmnk2" Dec 07 16:07:22 crc kubenswrapper[4716]: I1207 16:07:22.018591 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33db3f8a-1f66-44c4-b563-2f07dbc8267e-catalog-content\") pod \"community-operators-vmnk2\" (UID: \"33db3f8a-1f66-44c4-b563-2f07dbc8267e\") " pod="openshift-marketplace/community-operators-vmnk2" Dec 07 16:07:22 crc kubenswrapper[4716]: I1207 16:07:22.018651 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33db3f8a-1f66-44c4-b563-2f07dbc8267e-utilities\") pod \"community-operators-vmnk2\" (UID: \"33db3f8a-1f66-44c4-b563-2f07dbc8267e\") " pod="openshift-marketplace/community-operators-vmnk2" Dec 07 16:07:22 crc kubenswrapper[4716]: I1207 16:07:22.033977 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99qxt\" (UniqueName: \"kubernetes.io/projected/33db3f8a-1f66-44c4-b563-2f07dbc8267e-kube-api-access-99qxt\") pod \"community-operators-vmnk2\" (UID: \"33db3f8a-1f66-44c4-b563-2f07dbc8267e\") " pod="openshift-marketplace/community-operators-vmnk2" Dec 07 16:07:22 crc kubenswrapper[4716]: I1207 16:07:22.062832 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nkjl4"] Dec 07 16:07:22 crc kubenswrapper[4716]: W1207 16:07:22.065802 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaf948d72_6e6c_42d9_9d0b_1821d7f47176.slice/crio-5d9cd206bc50417f4e6509de3a6b68921fd8266880e558f8b2bde0883bc98e5f WatchSource:0}: Error finding container 5d9cd206bc50417f4e6509de3a6b68921fd8266880e558f8b2bde0883bc98e5f: Status 404 returned error can't find the container with id 5d9cd206bc50417f4e6509de3a6b68921fd8266880e558f8b2bde0883bc98e5f Dec 07 16:07:22 crc kubenswrapper[4716]: I1207 16:07:22.066442 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vmnk2" Dec 07 16:07:22 crc kubenswrapper[4716]: I1207 16:07:22.237956 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vmnk2"] Dec 07 16:07:22 crc kubenswrapper[4716]: W1207 16:07:22.267978 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod33db3f8a_1f66_44c4_b563_2f07dbc8267e.slice/crio-1fbb01dce7e8c688466c358ca5f05cbc5cfad3c6e70e758be8d7d1de9b9945af WatchSource:0}: Error finding container 1fbb01dce7e8c688466c358ca5f05cbc5cfad3c6e70e758be8d7d1de9b9945af: Status 404 returned error can't find the container with id 1fbb01dce7e8c688466c358ca5f05cbc5cfad3c6e70e758be8d7d1de9b9945af Dec 07 16:07:22 crc kubenswrapper[4716]: I1207 16:07:22.820549 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f7pqt" event={"ID":"0f030783-1303-4815-8ac1-fbecf8e036fe","Type":"ContainerStarted","Data":"eb9276609e92ff0cdfaf88c36795b7db404e8b341e8909de1a93376911295230"} Dec 07 16:07:22 crc kubenswrapper[4716]: I1207 16:07:22.821971 4716 generic.go:334] "Generic (PLEG): container finished" podID="33db3f8a-1f66-44c4-b563-2f07dbc8267e" containerID="9c4711b4d9dc2be74a629ca9af5f0696210c7a7f085aea2b5bf63fe8383d1452" exitCode=0 Dec 07 16:07:22 crc kubenswrapper[4716]: I1207 16:07:22.822028 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vmnk2" event={"ID":"33db3f8a-1f66-44c4-b563-2f07dbc8267e","Type":"ContainerDied","Data":"9c4711b4d9dc2be74a629ca9af5f0696210c7a7f085aea2b5bf63fe8383d1452"} Dec 07 16:07:22 crc kubenswrapper[4716]: I1207 16:07:22.822131 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vmnk2" event={"ID":"33db3f8a-1f66-44c4-b563-2f07dbc8267e","Type":"ContainerStarted","Data":"1fbb01dce7e8c688466c358ca5f05cbc5cfad3c6e70e758be8d7d1de9b9945af"} Dec 07 16:07:22 crc kubenswrapper[4716]: I1207 16:07:22.823242 4716 generic.go:334] "Generic (PLEG): container finished" podID="af948d72-6e6c-42d9-9d0b-1821d7f47176" containerID="2bee975ae0f46bf82edd14b4a8d21811a68a89138ca918522bbf201adc7bb2e4" exitCode=0 Dec 07 16:07:22 crc kubenswrapper[4716]: I1207 16:07:22.823858 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nkjl4" event={"ID":"af948d72-6e6c-42d9-9d0b-1821d7f47176","Type":"ContainerDied","Data":"2bee975ae0f46bf82edd14b4a8d21811a68a89138ca918522bbf201adc7bb2e4"} Dec 07 16:07:22 crc kubenswrapper[4716]: I1207 16:07:22.824001 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nkjl4" event={"ID":"af948d72-6e6c-42d9-9d0b-1821d7f47176","Type":"ContainerStarted","Data":"5d9cd206bc50417f4e6509de3a6b68921fd8266880e558f8b2bde0883bc98e5f"} Dec 07 16:07:22 crc kubenswrapper[4716]: I1207 16:07:22.826252 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tgn9w" event={"ID":"fc342361-554a-41c0-a512-06e046ff03eb","Type":"ContainerStarted","Data":"e7741478222b60dd7018806c507ee000d60830699cfafbf5fe6fb5e39f502754"} Dec 07 16:07:22 crc kubenswrapper[4716]: I1207 16:07:22.840598 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-f7pqt" podStartSLOduration=1.389783761 podStartE2EDuration="3.840580362s" podCreationTimestamp="2025-12-07 16:07:19 +0000 UTC" firstStartedPulling="2025-12-07 16:07:19.786968373 +0000 UTC m=+302.477253285" lastFinishedPulling="2025-12-07 16:07:22.237764974 +0000 UTC m=+304.928049886" observedRunningTime="2025-12-07 16:07:22.838326861 +0000 UTC m=+305.528611773" watchObservedRunningTime="2025-12-07 16:07:22.840580362 +0000 UTC m=+305.530865274" Dec 07 16:07:22 crc kubenswrapper[4716]: I1207 16:07:22.886032 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-tgn9w" podStartSLOduration=2.472633268 podStartE2EDuration="4.886014138s" podCreationTimestamp="2025-12-07 16:07:18 +0000 UTC" firstStartedPulling="2025-12-07 16:07:19.786545342 +0000 UTC m=+302.476830254" lastFinishedPulling="2025-12-07 16:07:22.199926212 +0000 UTC m=+304.890211124" observedRunningTime="2025-12-07 16:07:22.882482593 +0000 UTC m=+305.572767495" watchObservedRunningTime="2025-12-07 16:07:22.886014138 +0000 UTC m=+305.576299040" Dec 07 16:07:24 crc kubenswrapper[4716]: I1207 16:07:24.836730 4716 generic.go:334] "Generic (PLEG): container finished" podID="af948d72-6e6c-42d9-9d0b-1821d7f47176" containerID="076a4095d17a11f72b6e7f25166ee28139a8a3a231693ccf99f8a230525b9fce" exitCode=0 Dec 07 16:07:24 crc kubenswrapper[4716]: I1207 16:07:24.836867 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nkjl4" event={"ID":"af948d72-6e6c-42d9-9d0b-1821d7f47176","Type":"ContainerDied","Data":"076a4095d17a11f72b6e7f25166ee28139a8a3a231693ccf99f8a230525b9fce"} Dec 07 16:07:24 crc kubenswrapper[4716]: I1207 16:07:24.839220 4716 generic.go:334] "Generic (PLEG): container finished" podID="33db3f8a-1f66-44c4-b563-2f07dbc8267e" containerID="b79b9534dddbecfe28f21e1e60d5ea4c43da2349b57c67dc1619e3c5481c487b" exitCode=0 Dec 07 16:07:24 crc kubenswrapper[4716]: I1207 16:07:24.839263 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vmnk2" event={"ID":"33db3f8a-1f66-44c4-b563-2f07dbc8267e","Type":"ContainerDied","Data":"b79b9534dddbecfe28f21e1e60d5ea4c43da2349b57c67dc1619e3c5481c487b"} Dec 07 16:07:25 crc kubenswrapper[4716]: I1207 16:07:25.853434 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vmnk2" event={"ID":"33db3f8a-1f66-44c4-b563-2f07dbc8267e","Type":"ContainerStarted","Data":"75b92240e273d6fc1bb27991ab044c7c8f33535ab9c885fff56635ab4a776e1f"} Dec 07 16:07:25 crc kubenswrapper[4716]: I1207 16:07:25.856266 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nkjl4" event={"ID":"af948d72-6e6c-42d9-9d0b-1821d7f47176","Type":"ContainerStarted","Data":"cee75999bb98f39d8c22bc7215aff6657666e35e76669dffbaf711701f3c0748"} Dec 07 16:07:25 crc kubenswrapper[4716]: I1207 16:07:25.870699 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vmnk2" podStartSLOduration=2.285771006 podStartE2EDuration="4.870677316s" podCreationTimestamp="2025-12-07 16:07:21 +0000 UTC" firstStartedPulling="2025-12-07 16:07:22.823881341 +0000 UTC m=+305.514166243" lastFinishedPulling="2025-12-07 16:07:25.408787641 +0000 UTC m=+308.099072553" observedRunningTime="2025-12-07 16:07:25.870301936 +0000 UTC m=+308.560586858" watchObservedRunningTime="2025-12-07 16:07:25.870677316 +0000 UTC m=+308.560962228" Dec 07 16:07:25 crc kubenswrapper[4716]: I1207 16:07:25.888029 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-nkjl4" podStartSLOduration=2.346241478 podStartE2EDuration="4.888014994s" podCreationTimestamp="2025-12-07 16:07:21 +0000 UTC" firstStartedPulling="2025-12-07 16:07:22.824743334 +0000 UTC m=+305.515028236" lastFinishedPulling="2025-12-07 16:07:25.36651681 +0000 UTC m=+308.056801752" observedRunningTime="2025-12-07 16:07:25.885775533 +0000 UTC m=+308.576060445" watchObservedRunningTime="2025-12-07 16:07:25.888014994 +0000 UTC m=+308.578299906" Dec 07 16:07:29 crc kubenswrapper[4716]: I1207 16:07:29.284378 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-tgn9w" Dec 07 16:07:29 crc kubenswrapper[4716]: I1207 16:07:29.284720 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-tgn9w" Dec 07 16:07:29 crc kubenswrapper[4716]: I1207 16:07:29.326650 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-tgn9w" Dec 07 16:07:29 crc kubenswrapper[4716]: I1207 16:07:29.468414 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-f7pqt" Dec 07 16:07:29 crc kubenswrapper[4716]: I1207 16:07:29.468457 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-f7pqt" Dec 07 16:07:29 crc kubenswrapper[4716]: I1207 16:07:29.504704 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-f7pqt" Dec 07 16:07:29 crc kubenswrapper[4716]: I1207 16:07:29.915971 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-tgn9w" Dec 07 16:07:29 crc kubenswrapper[4716]: I1207 16:07:29.917889 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-f7pqt" Dec 07 16:07:31 crc kubenswrapper[4716]: I1207 16:07:31.060300 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-qjcjv"] Dec 07 16:07:31 crc kubenswrapper[4716]: I1207 16:07:31.060704 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-qjcjv" podUID="3cef2105-ae00-4224-bf5d-246393f3caa6" containerName="controller-manager" containerID="cri-o://612a6e8ee10acdd7b730f5775b4e6a6ac3fb0b8ebd6af77189069e871b78a6fd" gracePeriod=30 Dec 07 16:07:31 crc kubenswrapper[4716]: I1207 16:07:31.166767 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-4d2qh"] Dec 07 16:07:31 crc kubenswrapper[4716]: I1207 16:07:31.166972 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4d2qh" podUID="97d9c075-80c3-4dcb-aebc-649eec930413" containerName="route-controller-manager" containerID="cri-o://d69ed019410a6b4f19b787ca716186162802145b814e2f0be53510d9806175c6" gracePeriod=30 Dec 07 16:07:31 crc kubenswrapper[4716]: I1207 16:07:31.317288 4716 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-4d2qh container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.21:8443/healthz\": dial tcp 10.217.0.21:8443: connect: connection refused" start-of-body= Dec 07 16:07:31 crc kubenswrapper[4716]: I1207 16:07:31.317360 4716 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4d2qh" podUID="97d9c075-80c3-4dcb-aebc-649eec930413" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.21:8443/healthz\": dial tcp 10.217.0.21:8443: connect: connection refused" Dec 07 16:07:31 crc kubenswrapper[4716]: I1207 16:07:31.483818 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-qjcjv" Dec 07 16:07:31 crc kubenswrapper[4716]: I1207 16:07:31.641686 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3cef2105-ae00-4224-bf5d-246393f3caa6-client-ca\") pod \"3cef2105-ae00-4224-bf5d-246393f3caa6\" (UID: \"3cef2105-ae00-4224-bf5d-246393f3caa6\") " Dec 07 16:07:31 crc kubenswrapper[4716]: I1207 16:07:31.641979 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3cef2105-ae00-4224-bf5d-246393f3caa6-serving-cert\") pod \"3cef2105-ae00-4224-bf5d-246393f3caa6\" (UID: \"3cef2105-ae00-4224-bf5d-246393f3caa6\") " Dec 07 16:07:31 crc kubenswrapper[4716]: I1207 16:07:31.642015 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8rhn6\" (UniqueName: \"kubernetes.io/projected/3cef2105-ae00-4224-bf5d-246393f3caa6-kube-api-access-8rhn6\") pod \"3cef2105-ae00-4224-bf5d-246393f3caa6\" (UID: \"3cef2105-ae00-4224-bf5d-246393f3caa6\") " Dec 07 16:07:31 crc kubenswrapper[4716]: I1207 16:07:31.642037 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3cef2105-ae00-4224-bf5d-246393f3caa6-proxy-ca-bundles\") pod \"3cef2105-ae00-4224-bf5d-246393f3caa6\" (UID: \"3cef2105-ae00-4224-bf5d-246393f3caa6\") " Dec 07 16:07:31 crc kubenswrapper[4716]: I1207 16:07:31.642137 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3cef2105-ae00-4224-bf5d-246393f3caa6-config\") pod \"3cef2105-ae00-4224-bf5d-246393f3caa6\" (UID: \"3cef2105-ae00-4224-bf5d-246393f3caa6\") " Dec 07 16:07:31 crc kubenswrapper[4716]: I1207 16:07:31.643104 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cef2105-ae00-4224-bf5d-246393f3caa6-client-ca" (OuterVolumeSpecName: "client-ca") pod "3cef2105-ae00-4224-bf5d-246393f3caa6" (UID: "3cef2105-ae00-4224-bf5d-246393f3caa6"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:07:31 crc kubenswrapper[4716]: I1207 16:07:31.643141 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cef2105-ae00-4224-bf5d-246393f3caa6-config" (OuterVolumeSpecName: "config") pod "3cef2105-ae00-4224-bf5d-246393f3caa6" (UID: "3cef2105-ae00-4224-bf5d-246393f3caa6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:07:31 crc kubenswrapper[4716]: I1207 16:07:31.645880 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cef2105-ae00-4224-bf5d-246393f3caa6-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "3cef2105-ae00-4224-bf5d-246393f3caa6" (UID: "3cef2105-ae00-4224-bf5d-246393f3caa6"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:07:31 crc kubenswrapper[4716]: I1207 16:07:31.649657 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cef2105-ae00-4224-bf5d-246393f3caa6-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "3cef2105-ae00-4224-bf5d-246393f3caa6" (UID: "3cef2105-ae00-4224-bf5d-246393f3caa6"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:07:31 crc kubenswrapper[4716]: I1207 16:07:31.661843 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cef2105-ae00-4224-bf5d-246393f3caa6-kube-api-access-8rhn6" (OuterVolumeSpecName: "kube-api-access-8rhn6") pod "3cef2105-ae00-4224-bf5d-246393f3caa6" (UID: "3cef2105-ae00-4224-bf5d-246393f3caa6"). InnerVolumeSpecName "kube-api-access-8rhn6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:07:31 crc kubenswrapper[4716]: I1207 16:07:31.743315 4716 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3cef2105-ae00-4224-bf5d-246393f3caa6-client-ca\") on node \"crc\" DevicePath \"\"" Dec 07 16:07:31 crc kubenswrapper[4716]: I1207 16:07:31.743359 4716 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3cef2105-ae00-4224-bf5d-246393f3caa6-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 07 16:07:31 crc kubenswrapper[4716]: I1207 16:07:31.743374 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8rhn6\" (UniqueName: \"kubernetes.io/projected/3cef2105-ae00-4224-bf5d-246393f3caa6-kube-api-access-8rhn6\") on node \"crc\" DevicePath \"\"" Dec 07 16:07:31 crc kubenswrapper[4716]: I1207 16:07:31.743389 4716 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3cef2105-ae00-4224-bf5d-246393f3caa6-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 07 16:07:31 crc kubenswrapper[4716]: I1207 16:07:31.743403 4716 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3cef2105-ae00-4224-bf5d-246393f3caa6-config\") on node \"crc\" DevicePath \"\"" Dec 07 16:07:31 crc kubenswrapper[4716]: I1207 16:07:31.748753 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4d2qh" Dec 07 16:07:31 crc kubenswrapper[4716]: I1207 16:07:31.843822 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/97d9c075-80c3-4dcb-aebc-649eec930413-client-ca\") pod \"97d9c075-80c3-4dcb-aebc-649eec930413\" (UID: \"97d9c075-80c3-4dcb-aebc-649eec930413\") " Dec 07 16:07:31 crc kubenswrapper[4716]: I1207 16:07:31.843864 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qlvv6\" (UniqueName: \"kubernetes.io/projected/97d9c075-80c3-4dcb-aebc-649eec930413-kube-api-access-qlvv6\") pod \"97d9c075-80c3-4dcb-aebc-649eec930413\" (UID: \"97d9c075-80c3-4dcb-aebc-649eec930413\") " Dec 07 16:07:31 crc kubenswrapper[4716]: I1207 16:07:31.843888 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/97d9c075-80c3-4dcb-aebc-649eec930413-serving-cert\") pod \"97d9c075-80c3-4dcb-aebc-649eec930413\" (UID: \"97d9c075-80c3-4dcb-aebc-649eec930413\") " Dec 07 16:07:31 crc kubenswrapper[4716]: I1207 16:07:31.843907 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97d9c075-80c3-4dcb-aebc-649eec930413-config\") pod \"97d9c075-80c3-4dcb-aebc-649eec930413\" (UID: \"97d9c075-80c3-4dcb-aebc-649eec930413\") " Dec 07 16:07:31 crc kubenswrapper[4716]: I1207 16:07:31.844845 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/97d9c075-80c3-4dcb-aebc-649eec930413-client-ca" (OuterVolumeSpecName: "client-ca") pod "97d9c075-80c3-4dcb-aebc-649eec930413" (UID: "97d9c075-80c3-4dcb-aebc-649eec930413"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:07:31 crc kubenswrapper[4716]: I1207 16:07:31.844880 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/97d9c075-80c3-4dcb-aebc-649eec930413-config" (OuterVolumeSpecName: "config") pod "97d9c075-80c3-4dcb-aebc-649eec930413" (UID: "97d9c075-80c3-4dcb-aebc-649eec930413"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:07:31 crc kubenswrapper[4716]: I1207 16:07:31.847525 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97d9c075-80c3-4dcb-aebc-649eec930413-kube-api-access-qlvv6" (OuterVolumeSpecName: "kube-api-access-qlvv6") pod "97d9c075-80c3-4dcb-aebc-649eec930413" (UID: "97d9c075-80c3-4dcb-aebc-649eec930413"). InnerVolumeSpecName "kube-api-access-qlvv6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:07:31 crc kubenswrapper[4716]: I1207 16:07:31.848419 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97d9c075-80c3-4dcb-aebc-649eec930413-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "97d9c075-80c3-4dcb-aebc-649eec930413" (UID: "97d9c075-80c3-4dcb-aebc-649eec930413"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:07:31 crc kubenswrapper[4716]: I1207 16:07:31.865494 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-nkjl4" Dec 07 16:07:31 crc kubenswrapper[4716]: I1207 16:07:31.865536 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-nkjl4" Dec 07 16:07:31 crc kubenswrapper[4716]: I1207 16:07:31.887184 4716 generic.go:334] "Generic (PLEG): container finished" podID="3cef2105-ae00-4224-bf5d-246393f3caa6" containerID="612a6e8ee10acdd7b730f5775b4e6a6ac3fb0b8ebd6af77189069e871b78a6fd" exitCode=0 Dec 07 16:07:31 crc kubenswrapper[4716]: I1207 16:07:31.887251 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-qjcjv" event={"ID":"3cef2105-ae00-4224-bf5d-246393f3caa6","Type":"ContainerDied","Data":"612a6e8ee10acdd7b730f5775b4e6a6ac3fb0b8ebd6af77189069e871b78a6fd"} Dec 07 16:07:31 crc kubenswrapper[4716]: I1207 16:07:31.887283 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-qjcjv" event={"ID":"3cef2105-ae00-4224-bf5d-246393f3caa6","Type":"ContainerDied","Data":"829cc2178638ad9f3b726631e977079615f192b96abda8d78a8fa8cccf2b9f5b"} Dec 07 16:07:31 crc kubenswrapper[4716]: I1207 16:07:31.887301 4716 scope.go:117] "RemoveContainer" containerID="612a6e8ee10acdd7b730f5775b4e6a6ac3fb0b8ebd6af77189069e871b78a6fd" Dec 07 16:07:31 crc kubenswrapper[4716]: I1207 16:07:31.887393 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-qjcjv" Dec 07 16:07:31 crc kubenswrapper[4716]: I1207 16:07:31.891235 4716 generic.go:334] "Generic (PLEG): container finished" podID="97d9c075-80c3-4dcb-aebc-649eec930413" containerID="d69ed019410a6b4f19b787ca716186162802145b814e2f0be53510d9806175c6" exitCode=0 Dec 07 16:07:31 crc kubenswrapper[4716]: I1207 16:07:31.891268 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4d2qh" event={"ID":"97d9c075-80c3-4dcb-aebc-649eec930413","Type":"ContainerDied","Data":"d69ed019410a6b4f19b787ca716186162802145b814e2f0be53510d9806175c6"} Dec 07 16:07:31 crc kubenswrapper[4716]: I1207 16:07:31.891292 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4d2qh" event={"ID":"97d9c075-80c3-4dcb-aebc-649eec930413","Type":"ContainerDied","Data":"9db3e06592b07e08f8a1d6f4cbb7ad969e1f1bce6cbc0de81cf9f32d702d66c6"} Dec 07 16:07:31 crc kubenswrapper[4716]: I1207 16:07:31.891352 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4d2qh" Dec 07 16:07:31 crc kubenswrapper[4716]: I1207 16:07:31.902921 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-qjcjv"] Dec 07 16:07:31 crc kubenswrapper[4716]: I1207 16:07:31.909750 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-qjcjv"] Dec 07 16:07:31 crc kubenswrapper[4716]: I1207 16:07:31.910985 4716 scope.go:117] "RemoveContainer" containerID="612a6e8ee10acdd7b730f5775b4e6a6ac3fb0b8ebd6af77189069e871b78a6fd" Dec 07 16:07:31 crc kubenswrapper[4716]: E1207 16:07:31.912692 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"612a6e8ee10acdd7b730f5775b4e6a6ac3fb0b8ebd6af77189069e871b78a6fd\": container with ID starting with 612a6e8ee10acdd7b730f5775b4e6a6ac3fb0b8ebd6af77189069e871b78a6fd not found: ID does not exist" containerID="612a6e8ee10acdd7b730f5775b4e6a6ac3fb0b8ebd6af77189069e871b78a6fd" Dec 07 16:07:31 crc kubenswrapper[4716]: I1207 16:07:31.912732 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"612a6e8ee10acdd7b730f5775b4e6a6ac3fb0b8ebd6af77189069e871b78a6fd"} err="failed to get container status \"612a6e8ee10acdd7b730f5775b4e6a6ac3fb0b8ebd6af77189069e871b78a6fd\": rpc error: code = NotFound desc = could not find container \"612a6e8ee10acdd7b730f5775b4e6a6ac3fb0b8ebd6af77189069e871b78a6fd\": container with ID starting with 612a6e8ee10acdd7b730f5775b4e6a6ac3fb0b8ebd6af77189069e871b78a6fd not found: ID does not exist" Dec 07 16:07:31 crc kubenswrapper[4716]: I1207 16:07:31.912757 4716 scope.go:117] "RemoveContainer" containerID="d69ed019410a6b4f19b787ca716186162802145b814e2f0be53510d9806175c6" Dec 07 16:07:31 crc kubenswrapper[4716]: I1207 16:07:31.921234 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-nkjl4" Dec 07 16:07:31 crc kubenswrapper[4716]: I1207 16:07:31.930302 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-4d2qh"] Dec 07 16:07:31 crc kubenswrapper[4716]: I1207 16:07:31.934613 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-4d2qh"] Dec 07 16:07:31 crc kubenswrapper[4716]: I1207 16:07:31.938891 4716 scope.go:117] "RemoveContainer" containerID="d69ed019410a6b4f19b787ca716186162802145b814e2f0be53510d9806175c6" Dec 07 16:07:31 crc kubenswrapper[4716]: E1207 16:07:31.940252 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d69ed019410a6b4f19b787ca716186162802145b814e2f0be53510d9806175c6\": container with ID starting with d69ed019410a6b4f19b787ca716186162802145b814e2f0be53510d9806175c6 not found: ID does not exist" containerID="d69ed019410a6b4f19b787ca716186162802145b814e2f0be53510d9806175c6" Dec 07 16:07:31 crc kubenswrapper[4716]: I1207 16:07:31.940300 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d69ed019410a6b4f19b787ca716186162802145b814e2f0be53510d9806175c6"} err="failed to get container status \"d69ed019410a6b4f19b787ca716186162802145b814e2f0be53510d9806175c6\": rpc error: code = NotFound desc = could not find container \"d69ed019410a6b4f19b787ca716186162802145b814e2f0be53510d9806175c6\": container with ID starting with d69ed019410a6b4f19b787ca716186162802145b814e2f0be53510d9806175c6 not found: ID does not exist" Dec 07 16:07:31 crc kubenswrapper[4716]: I1207 16:07:31.945594 4716 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/97d9c075-80c3-4dcb-aebc-649eec930413-client-ca\") on node \"crc\" DevicePath \"\"" Dec 07 16:07:31 crc kubenswrapper[4716]: I1207 16:07:31.945618 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qlvv6\" (UniqueName: \"kubernetes.io/projected/97d9c075-80c3-4dcb-aebc-649eec930413-kube-api-access-qlvv6\") on node \"crc\" DevicePath \"\"" Dec 07 16:07:31 crc kubenswrapper[4716]: I1207 16:07:31.945664 4716 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/97d9c075-80c3-4dcb-aebc-649eec930413-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 07 16:07:31 crc kubenswrapper[4716]: I1207 16:07:31.945677 4716 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97d9c075-80c3-4dcb-aebc-649eec930413-config\") on node \"crc\" DevicePath \"\"" Dec 07 16:07:31 crc kubenswrapper[4716]: I1207 16:07:31.956691 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-nkjl4" Dec 07 16:07:32 crc kubenswrapper[4716]: I1207 16:07:32.067147 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vmnk2" Dec 07 16:07:32 crc kubenswrapper[4716]: I1207 16:07:32.067201 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vmnk2" Dec 07 16:07:32 crc kubenswrapper[4716]: I1207 16:07:32.114965 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vmnk2" Dec 07 16:07:32 crc kubenswrapper[4716]: I1207 16:07:32.945812 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vmnk2" Dec 07 16:07:33 crc kubenswrapper[4716]: I1207 16:07:33.045252 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7c9b544d8-h8g9h"] Dec 07 16:07:33 crc kubenswrapper[4716]: E1207 16:07:33.045756 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cef2105-ae00-4224-bf5d-246393f3caa6" containerName="controller-manager" Dec 07 16:07:33 crc kubenswrapper[4716]: I1207 16:07:33.045791 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cef2105-ae00-4224-bf5d-246393f3caa6" containerName="controller-manager" Dec 07 16:07:33 crc kubenswrapper[4716]: E1207 16:07:33.045832 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97d9c075-80c3-4dcb-aebc-649eec930413" containerName="route-controller-manager" Dec 07 16:07:33 crc kubenswrapper[4716]: I1207 16:07:33.045846 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="97d9c075-80c3-4dcb-aebc-649eec930413" containerName="route-controller-manager" Dec 07 16:07:33 crc kubenswrapper[4716]: I1207 16:07:33.046025 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="97d9c075-80c3-4dcb-aebc-649eec930413" containerName="route-controller-manager" Dec 07 16:07:33 crc kubenswrapper[4716]: I1207 16:07:33.046064 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="3cef2105-ae00-4224-bf5d-246393f3caa6" containerName="controller-manager" Dec 07 16:07:33 crc kubenswrapper[4716]: I1207 16:07:33.046850 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7c9b544d8-h8g9h" Dec 07 16:07:33 crc kubenswrapper[4716]: I1207 16:07:33.049712 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6c7c66dbd8-dwgtq"] Dec 07 16:07:33 crc kubenswrapper[4716]: I1207 16:07:33.050760 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6c7c66dbd8-dwgtq" Dec 07 16:07:33 crc kubenswrapper[4716]: I1207 16:07:33.052695 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 07 16:07:33 crc kubenswrapper[4716]: I1207 16:07:33.052898 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 07 16:07:33 crc kubenswrapper[4716]: I1207 16:07:33.053059 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 07 16:07:33 crc kubenswrapper[4716]: I1207 16:07:33.053325 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 07 16:07:33 crc kubenswrapper[4716]: I1207 16:07:33.053483 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 07 16:07:33 crc kubenswrapper[4716]: I1207 16:07:33.053569 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 07 16:07:33 crc kubenswrapper[4716]: I1207 16:07:33.053675 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 07 16:07:33 crc kubenswrapper[4716]: I1207 16:07:33.053712 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 07 16:07:33 crc kubenswrapper[4716]: I1207 16:07:33.053766 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 07 16:07:33 crc kubenswrapper[4716]: I1207 16:07:33.054218 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 07 16:07:33 crc kubenswrapper[4716]: I1207 16:07:33.054433 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 07 16:07:33 crc kubenswrapper[4716]: I1207 16:07:33.057409 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 07 16:07:33 crc kubenswrapper[4716]: I1207 16:07:33.062603 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 07 16:07:33 crc kubenswrapper[4716]: I1207 16:07:33.063632 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6c7c66dbd8-dwgtq"] Dec 07 16:07:33 crc kubenswrapper[4716]: I1207 16:07:33.067425 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7c9b544d8-h8g9h"] Dec 07 16:07:33 crc kubenswrapper[4716]: I1207 16:07:33.158807 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6acc423b-537c-4fc2-a098-78376fdc812c-serving-cert\") pod \"controller-manager-7c9b544d8-h8g9h\" (UID: \"6acc423b-537c-4fc2-a098-78376fdc812c\") " pod="openshift-controller-manager/controller-manager-7c9b544d8-h8g9h" Dec 07 16:07:33 crc kubenswrapper[4716]: I1207 16:07:33.158854 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ba7ae3c6-4861-4602-bba3-d72858464cb1-client-ca\") pod \"route-controller-manager-6c7c66dbd8-dwgtq\" (UID: \"ba7ae3c6-4861-4602-bba3-d72858464cb1\") " pod="openshift-route-controller-manager/route-controller-manager-6c7c66dbd8-dwgtq" Dec 07 16:07:33 crc kubenswrapper[4716]: I1207 16:07:33.158873 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6acc423b-537c-4fc2-a098-78376fdc812c-proxy-ca-bundles\") pod \"controller-manager-7c9b544d8-h8g9h\" (UID: \"6acc423b-537c-4fc2-a098-78376fdc812c\") " pod="openshift-controller-manager/controller-manager-7c9b544d8-h8g9h" Dec 07 16:07:33 crc kubenswrapper[4716]: I1207 16:07:33.158890 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jn7ql\" (UniqueName: \"kubernetes.io/projected/6acc423b-537c-4fc2-a098-78376fdc812c-kube-api-access-jn7ql\") pod \"controller-manager-7c9b544d8-h8g9h\" (UID: \"6acc423b-537c-4fc2-a098-78376fdc812c\") " pod="openshift-controller-manager/controller-manager-7c9b544d8-h8g9h" Dec 07 16:07:33 crc kubenswrapper[4716]: I1207 16:07:33.159229 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ba7ae3c6-4861-4602-bba3-d72858464cb1-serving-cert\") pod \"route-controller-manager-6c7c66dbd8-dwgtq\" (UID: \"ba7ae3c6-4861-4602-bba3-d72858464cb1\") " pod="openshift-route-controller-manager/route-controller-manager-6c7c66dbd8-dwgtq" Dec 07 16:07:33 crc kubenswrapper[4716]: I1207 16:07:33.159315 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6acc423b-537c-4fc2-a098-78376fdc812c-client-ca\") pod \"controller-manager-7c9b544d8-h8g9h\" (UID: \"6acc423b-537c-4fc2-a098-78376fdc812c\") " pod="openshift-controller-manager/controller-manager-7c9b544d8-h8g9h" Dec 07 16:07:33 crc kubenswrapper[4716]: I1207 16:07:33.159458 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dxp46\" (UniqueName: \"kubernetes.io/projected/ba7ae3c6-4861-4602-bba3-d72858464cb1-kube-api-access-dxp46\") pod \"route-controller-manager-6c7c66dbd8-dwgtq\" (UID: \"ba7ae3c6-4861-4602-bba3-d72858464cb1\") " pod="openshift-route-controller-manager/route-controller-manager-6c7c66dbd8-dwgtq" Dec 07 16:07:33 crc kubenswrapper[4716]: I1207 16:07:33.159513 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba7ae3c6-4861-4602-bba3-d72858464cb1-config\") pod \"route-controller-manager-6c7c66dbd8-dwgtq\" (UID: \"ba7ae3c6-4861-4602-bba3-d72858464cb1\") " pod="openshift-route-controller-manager/route-controller-manager-6c7c66dbd8-dwgtq" Dec 07 16:07:33 crc kubenswrapper[4716]: I1207 16:07:33.159637 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6acc423b-537c-4fc2-a098-78376fdc812c-config\") pod \"controller-manager-7c9b544d8-h8g9h\" (UID: \"6acc423b-537c-4fc2-a098-78376fdc812c\") " pod="openshift-controller-manager/controller-manager-7c9b544d8-h8g9h" Dec 07 16:07:33 crc kubenswrapper[4716]: I1207 16:07:33.260945 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ba7ae3c6-4861-4602-bba3-d72858464cb1-serving-cert\") pod \"route-controller-manager-6c7c66dbd8-dwgtq\" (UID: \"ba7ae3c6-4861-4602-bba3-d72858464cb1\") " pod="openshift-route-controller-manager/route-controller-manager-6c7c66dbd8-dwgtq" Dec 07 16:07:33 crc kubenswrapper[4716]: I1207 16:07:33.261003 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6acc423b-537c-4fc2-a098-78376fdc812c-client-ca\") pod \"controller-manager-7c9b544d8-h8g9h\" (UID: \"6acc423b-537c-4fc2-a098-78376fdc812c\") " pod="openshift-controller-manager/controller-manager-7c9b544d8-h8g9h" Dec 07 16:07:33 crc kubenswrapper[4716]: I1207 16:07:33.261156 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dxp46\" (UniqueName: \"kubernetes.io/projected/ba7ae3c6-4861-4602-bba3-d72858464cb1-kube-api-access-dxp46\") pod \"route-controller-manager-6c7c66dbd8-dwgtq\" (UID: \"ba7ae3c6-4861-4602-bba3-d72858464cb1\") " pod="openshift-route-controller-manager/route-controller-manager-6c7c66dbd8-dwgtq" Dec 07 16:07:33 crc kubenswrapper[4716]: I1207 16:07:33.261199 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba7ae3c6-4861-4602-bba3-d72858464cb1-config\") pod \"route-controller-manager-6c7c66dbd8-dwgtq\" (UID: \"ba7ae3c6-4861-4602-bba3-d72858464cb1\") " pod="openshift-route-controller-manager/route-controller-manager-6c7c66dbd8-dwgtq" Dec 07 16:07:33 crc kubenswrapper[4716]: I1207 16:07:33.261251 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6acc423b-537c-4fc2-a098-78376fdc812c-config\") pod \"controller-manager-7c9b544d8-h8g9h\" (UID: \"6acc423b-537c-4fc2-a098-78376fdc812c\") " pod="openshift-controller-manager/controller-manager-7c9b544d8-h8g9h" Dec 07 16:07:33 crc kubenswrapper[4716]: I1207 16:07:33.261284 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6acc423b-537c-4fc2-a098-78376fdc812c-serving-cert\") pod \"controller-manager-7c9b544d8-h8g9h\" (UID: \"6acc423b-537c-4fc2-a098-78376fdc812c\") " pod="openshift-controller-manager/controller-manager-7c9b544d8-h8g9h" Dec 07 16:07:33 crc kubenswrapper[4716]: I1207 16:07:33.261304 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ba7ae3c6-4861-4602-bba3-d72858464cb1-client-ca\") pod \"route-controller-manager-6c7c66dbd8-dwgtq\" (UID: \"ba7ae3c6-4861-4602-bba3-d72858464cb1\") " pod="openshift-route-controller-manager/route-controller-manager-6c7c66dbd8-dwgtq" Dec 07 16:07:33 crc kubenswrapper[4716]: I1207 16:07:33.261327 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6acc423b-537c-4fc2-a098-78376fdc812c-proxy-ca-bundles\") pod \"controller-manager-7c9b544d8-h8g9h\" (UID: \"6acc423b-537c-4fc2-a098-78376fdc812c\") " pod="openshift-controller-manager/controller-manager-7c9b544d8-h8g9h" Dec 07 16:07:33 crc kubenswrapper[4716]: I1207 16:07:33.261349 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jn7ql\" (UniqueName: \"kubernetes.io/projected/6acc423b-537c-4fc2-a098-78376fdc812c-kube-api-access-jn7ql\") pod \"controller-manager-7c9b544d8-h8g9h\" (UID: \"6acc423b-537c-4fc2-a098-78376fdc812c\") " pod="openshift-controller-manager/controller-manager-7c9b544d8-h8g9h" Dec 07 16:07:33 crc kubenswrapper[4716]: I1207 16:07:33.262015 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6acc423b-537c-4fc2-a098-78376fdc812c-client-ca\") pod \"controller-manager-7c9b544d8-h8g9h\" (UID: \"6acc423b-537c-4fc2-a098-78376fdc812c\") " pod="openshift-controller-manager/controller-manager-7c9b544d8-h8g9h" Dec 07 16:07:33 crc kubenswrapper[4716]: I1207 16:07:33.262635 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ba7ae3c6-4861-4602-bba3-d72858464cb1-client-ca\") pod \"route-controller-manager-6c7c66dbd8-dwgtq\" (UID: \"ba7ae3c6-4861-4602-bba3-d72858464cb1\") " pod="openshift-route-controller-manager/route-controller-manager-6c7c66dbd8-dwgtq" Dec 07 16:07:33 crc kubenswrapper[4716]: I1207 16:07:33.263666 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6acc423b-537c-4fc2-a098-78376fdc812c-proxy-ca-bundles\") pod \"controller-manager-7c9b544d8-h8g9h\" (UID: \"6acc423b-537c-4fc2-a098-78376fdc812c\") " pod="openshift-controller-manager/controller-manager-7c9b544d8-h8g9h" Dec 07 16:07:33 crc kubenswrapper[4716]: I1207 16:07:33.263686 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba7ae3c6-4861-4602-bba3-d72858464cb1-config\") pod \"route-controller-manager-6c7c66dbd8-dwgtq\" (UID: \"ba7ae3c6-4861-4602-bba3-d72858464cb1\") " pod="openshift-route-controller-manager/route-controller-manager-6c7c66dbd8-dwgtq" Dec 07 16:07:33 crc kubenswrapper[4716]: I1207 16:07:33.264177 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6acc423b-537c-4fc2-a098-78376fdc812c-config\") pod \"controller-manager-7c9b544d8-h8g9h\" (UID: \"6acc423b-537c-4fc2-a098-78376fdc812c\") " pod="openshift-controller-manager/controller-manager-7c9b544d8-h8g9h" Dec 07 16:07:33 crc kubenswrapper[4716]: I1207 16:07:33.269638 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ba7ae3c6-4861-4602-bba3-d72858464cb1-serving-cert\") pod \"route-controller-manager-6c7c66dbd8-dwgtq\" (UID: \"ba7ae3c6-4861-4602-bba3-d72858464cb1\") " pod="openshift-route-controller-manager/route-controller-manager-6c7c66dbd8-dwgtq" Dec 07 16:07:33 crc kubenswrapper[4716]: I1207 16:07:33.270265 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6acc423b-537c-4fc2-a098-78376fdc812c-serving-cert\") pod \"controller-manager-7c9b544d8-h8g9h\" (UID: \"6acc423b-537c-4fc2-a098-78376fdc812c\") " pod="openshift-controller-manager/controller-manager-7c9b544d8-h8g9h" Dec 07 16:07:33 crc kubenswrapper[4716]: I1207 16:07:33.278821 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jn7ql\" (UniqueName: \"kubernetes.io/projected/6acc423b-537c-4fc2-a098-78376fdc812c-kube-api-access-jn7ql\") pod \"controller-manager-7c9b544d8-h8g9h\" (UID: \"6acc423b-537c-4fc2-a098-78376fdc812c\") " pod="openshift-controller-manager/controller-manager-7c9b544d8-h8g9h" Dec 07 16:07:33 crc kubenswrapper[4716]: I1207 16:07:33.280260 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dxp46\" (UniqueName: \"kubernetes.io/projected/ba7ae3c6-4861-4602-bba3-d72858464cb1-kube-api-access-dxp46\") pod \"route-controller-manager-6c7c66dbd8-dwgtq\" (UID: \"ba7ae3c6-4861-4602-bba3-d72858464cb1\") " pod="openshift-route-controller-manager/route-controller-manager-6c7c66dbd8-dwgtq" Dec 07 16:07:33 crc kubenswrapper[4716]: I1207 16:07:33.373024 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7c9b544d8-h8g9h" Dec 07 16:07:33 crc kubenswrapper[4716]: I1207 16:07:33.391448 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6c7c66dbd8-dwgtq" Dec 07 16:07:33 crc kubenswrapper[4716]: I1207 16:07:33.668427 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cef2105-ae00-4224-bf5d-246393f3caa6" path="/var/lib/kubelet/pods/3cef2105-ae00-4224-bf5d-246393f3caa6/volumes" Dec 07 16:07:33 crc kubenswrapper[4716]: I1207 16:07:33.669790 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="97d9c075-80c3-4dcb-aebc-649eec930413" path="/var/lib/kubelet/pods/97d9c075-80c3-4dcb-aebc-649eec930413/volumes" Dec 07 16:07:33 crc kubenswrapper[4716]: I1207 16:07:33.670347 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7c9b544d8-h8g9h"] Dec 07 16:07:33 crc kubenswrapper[4716]: I1207 16:07:33.790424 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7c9b544d8-h8g9h"] Dec 07 16:07:33 crc kubenswrapper[4716]: I1207 16:07:33.804195 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6c7c66dbd8-dwgtq"] Dec 07 16:07:33 crc kubenswrapper[4716]: I1207 16:07:33.832936 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6c7c66dbd8-dwgtq"] Dec 07 16:07:33 crc kubenswrapper[4716]: W1207 16:07:33.838502 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podba7ae3c6_4861_4602_bba3_d72858464cb1.slice/crio-79f65626ca28c3e24c5f9240ad2a08d394691dfbf784f63ed9fb2214896586ce WatchSource:0}: Error finding container 79f65626ca28c3e24c5f9240ad2a08d394691dfbf784f63ed9fb2214896586ce: Status 404 returned error can't find the container with id 79f65626ca28c3e24c5f9240ad2a08d394691dfbf784f63ed9fb2214896586ce Dec 07 16:07:33 crc kubenswrapper[4716]: I1207 16:07:33.902367 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6c7c66dbd8-dwgtq" event={"ID":"ba7ae3c6-4861-4602-bba3-d72858464cb1","Type":"ContainerStarted","Data":"79f65626ca28c3e24c5f9240ad2a08d394691dfbf784f63ed9fb2214896586ce"} Dec 07 16:07:33 crc kubenswrapper[4716]: I1207 16:07:33.903570 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7c9b544d8-h8g9h" event={"ID":"6acc423b-537c-4fc2-a098-78376fdc812c","Type":"ContainerStarted","Data":"39d596a9d385e830993a5ef097e1edb70c139b7052c87907f46beb52f948a32c"} Dec 07 16:07:35 crc kubenswrapper[4716]: I1207 16:07:35.913862 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7c9b544d8-h8g9h" event={"ID":"6acc423b-537c-4fc2-a098-78376fdc812c","Type":"ContainerStarted","Data":"56af6622941b689ea303da653b8532c81628a968eb1c273a55acfdcfcf8cc881"} Dec 07 16:07:35 crc kubenswrapper[4716]: I1207 16:07:35.914517 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7c9b544d8-h8g9h" Dec 07 16:07:35 crc kubenswrapper[4716]: I1207 16:07:35.913961 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-7c9b544d8-h8g9h" podUID="6acc423b-537c-4fc2-a098-78376fdc812c" containerName="controller-manager" containerID="cri-o://56af6622941b689ea303da653b8532c81628a968eb1c273a55acfdcfcf8cc881" gracePeriod=30 Dec 07 16:07:35 crc kubenswrapper[4716]: I1207 16:07:35.914873 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6c7c66dbd8-dwgtq" event={"ID":"ba7ae3c6-4861-4602-bba3-d72858464cb1","Type":"ContainerStarted","Data":"5d9ccfb70b20aed39972a147694629c066608698614a5c95dcf9c4ad946245a3"} Dec 07 16:07:35 crc kubenswrapper[4716]: I1207 16:07:35.914953 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6c7c66dbd8-dwgtq" podUID="ba7ae3c6-4861-4602-bba3-d72858464cb1" containerName="route-controller-manager" containerID="cri-o://5d9ccfb70b20aed39972a147694629c066608698614a5c95dcf9c4ad946245a3" gracePeriod=30 Dec 07 16:07:35 crc kubenswrapper[4716]: I1207 16:07:35.915246 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6c7c66dbd8-dwgtq" Dec 07 16:07:35 crc kubenswrapper[4716]: I1207 16:07:35.919357 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7c9b544d8-h8g9h" Dec 07 16:07:35 crc kubenswrapper[4716]: I1207 16:07:35.924055 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6c7c66dbd8-dwgtq" Dec 07 16:07:35 crc kubenswrapper[4716]: I1207 16:07:35.938616 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7c9b544d8-h8g9h" podStartSLOduration=4.938593891 podStartE2EDuration="4.938593891s" podCreationTimestamp="2025-12-07 16:07:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:07:35.935051104 +0000 UTC m=+318.625336036" watchObservedRunningTime="2025-12-07 16:07:35.938593891 +0000 UTC m=+318.628878833" Dec 07 16:07:35 crc kubenswrapper[4716]: I1207 16:07:35.953333 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6c7c66dbd8-dwgtq" podStartSLOduration=4.953306498 podStartE2EDuration="4.953306498s" podCreationTimestamp="2025-12-07 16:07:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:07:35.949378791 +0000 UTC m=+318.639663743" watchObservedRunningTime="2025-12-07 16:07:35.953306498 +0000 UTC m=+318.643591430" Dec 07 16:07:37 crc kubenswrapper[4716]: I1207 16:07:37.928765 4716 generic.go:334] "Generic (PLEG): container finished" podID="6acc423b-537c-4fc2-a098-78376fdc812c" containerID="56af6622941b689ea303da653b8532c81628a968eb1c273a55acfdcfcf8cc881" exitCode=0 Dec 07 16:07:37 crc kubenswrapper[4716]: I1207 16:07:37.928867 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7c9b544d8-h8g9h" event={"ID":"6acc423b-537c-4fc2-a098-78376fdc812c","Type":"ContainerDied","Data":"56af6622941b689ea303da653b8532c81628a968eb1c273a55acfdcfcf8cc881"} Dec 07 16:07:37 crc kubenswrapper[4716]: I1207 16:07:37.940347 4716 generic.go:334] "Generic (PLEG): container finished" podID="ba7ae3c6-4861-4602-bba3-d72858464cb1" containerID="5d9ccfb70b20aed39972a147694629c066608698614a5c95dcf9c4ad946245a3" exitCode=0 Dec 07 16:07:37 crc kubenswrapper[4716]: I1207 16:07:37.940410 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6c7c66dbd8-dwgtq" event={"ID":"ba7ae3c6-4861-4602-bba3-d72858464cb1","Type":"ContainerDied","Data":"5d9ccfb70b20aed39972a147694629c066608698614a5c95dcf9c4ad946245a3"} Dec 07 16:07:38 crc kubenswrapper[4716]: I1207 16:07:38.141904 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6c7c66dbd8-dwgtq" Dec 07 16:07:38 crc kubenswrapper[4716]: I1207 16:07:38.148027 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7c9b544d8-h8g9h" Dec 07 16:07:38 crc kubenswrapper[4716]: I1207 16:07:38.168617 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-75ddb6d6df-c2wt7"] Dec 07 16:07:38 crc kubenswrapper[4716]: E1207 16:07:38.168852 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba7ae3c6-4861-4602-bba3-d72858464cb1" containerName="route-controller-manager" Dec 07 16:07:38 crc kubenswrapper[4716]: I1207 16:07:38.168868 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba7ae3c6-4861-4602-bba3-d72858464cb1" containerName="route-controller-manager" Dec 07 16:07:38 crc kubenswrapper[4716]: E1207 16:07:38.168901 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6acc423b-537c-4fc2-a098-78376fdc812c" containerName="controller-manager" Dec 07 16:07:38 crc kubenswrapper[4716]: I1207 16:07:38.168913 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="6acc423b-537c-4fc2-a098-78376fdc812c" containerName="controller-manager" Dec 07 16:07:38 crc kubenswrapper[4716]: I1207 16:07:38.169043 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba7ae3c6-4861-4602-bba3-d72858464cb1" containerName="route-controller-manager" Dec 07 16:07:38 crc kubenswrapper[4716]: I1207 16:07:38.169067 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="6acc423b-537c-4fc2-a098-78376fdc812c" containerName="controller-manager" Dec 07 16:07:38 crc kubenswrapper[4716]: I1207 16:07:38.169531 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-75ddb6d6df-c2wt7" Dec 07 16:07:38 crc kubenswrapper[4716]: I1207 16:07:38.178903 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-75ddb6d6df-c2wt7"] Dec 07 16:07:38 crc kubenswrapper[4716]: I1207 16:07:38.246450 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dxp46\" (UniqueName: \"kubernetes.io/projected/ba7ae3c6-4861-4602-bba3-d72858464cb1-kube-api-access-dxp46\") pod \"ba7ae3c6-4861-4602-bba3-d72858464cb1\" (UID: \"ba7ae3c6-4861-4602-bba3-d72858464cb1\") " Dec 07 16:07:38 crc kubenswrapper[4716]: I1207 16:07:38.246504 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6acc423b-537c-4fc2-a098-78376fdc812c-proxy-ca-bundles\") pod \"6acc423b-537c-4fc2-a098-78376fdc812c\" (UID: \"6acc423b-537c-4fc2-a098-78376fdc812c\") " Dec 07 16:07:38 crc kubenswrapper[4716]: I1207 16:07:38.246555 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6acc423b-537c-4fc2-a098-78376fdc812c-serving-cert\") pod \"6acc423b-537c-4fc2-a098-78376fdc812c\" (UID: \"6acc423b-537c-4fc2-a098-78376fdc812c\") " Dec 07 16:07:38 crc kubenswrapper[4716]: I1207 16:07:38.246578 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6acc423b-537c-4fc2-a098-78376fdc812c-config\") pod \"6acc423b-537c-4fc2-a098-78376fdc812c\" (UID: \"6acc423b-537c-4fc2-a098-78376fdc812c\") " Dec 07 16:07:38 crc kubenswrapper[4716]: I1207 16:07:38.246660 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ba7ae3c6-4861-4602-bba3-d72858464cb1-serving-cert\") pod \"ba7ae3c6-4861-4602-bba3-d72858464cb1\" (UID: \"ba7ae3c6-4861-4602-bba3-d72858464cb1\") " Dec 07 16:07:38 crc kubenswrapper[4716]: I1207 16:07:38.246685 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6acc423b-537c-4fc2-a098-78376fdc812c-client-ca\") pod \"6acc423b-537c-4fc2-a098-78376fdc812c\" (UID: \"6acc423b-537c-4fc2-a098-78376fdc812c\") " Dec 07 16:07:38 crc kubenswrapper[4716]: I1207 16:07:38.246708 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ba7ae3c6-4861-4602-bba3-d72858464cb1-client-ca\") pod \"ba7ae3c6-4861-4602-bba3-d72858464cb1\" (UID: \"ba7ae3c6-4861-4602-bba3-d72858464cb1\") " Dec 07 16:07:38 crc kubenswrapper[4716]: I1207 16:07:38.246742 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba7ae3c6-4861-4602-bba3-d72858464cb1-config\") pod \"ba7ae3c6-4861-4602-bba3-d72858464cb1\" (UID: \"ba7ae3c6-4861-4602-bba3-d72858464cb1\") " Dec 07 16:07:38 crc kubenswrapper[4716]: I1207 16:07:38.246795 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jn7ql\" (UniqueName: \"kubernetes.io/projected/6acc423b-537c-4fc2-a098-78376fdc812c-kube-api-access-jn7ql\") pod \"6acc423b-537c-4fc2-a098-78376fdc812c\" (UID: \"6acc423b-537c-4fc2-a098-78376fdc812c\") " Dec 07 16:07:38 crc kubenswrapper[4716]: I1207 16:07:38.247331 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d93f592-1d2d-49e1-a945-50d193c95177-config\") pod \"route-controller-manager-75ddb6d6df-c2wt7\" (UID: \"5d93f592-1d2d-49e1-a945-50d193c95177\") " pod="openshift-route-controller-manager/route-controller-manager-75ddb6d6df-c2wt7" Dec 07 16:07:38 crc kubenswrapper[4716]: I1207 16:07:38.247403 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5d93f592-1d2d-49e1-a945-50d193c95177-client-ca\") pod \"route-controller-manager-75ddb6d6df-c2wt7\" (UID: \"5d93f592-1d2d-49e1-a945-50d193c95177\") " pod="openshift-route-controller-manager/route-controller-manager-75ddb6d6df-c2wt7" Dec 07 16:07:38 crc kubenswrapper[4716]: I1207 16:07:38.247440 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7ztx\" (UniqueName: \"kubernetes.io/projected/5d93f592-1d2d-49e1-a945-50d193c95177-kube-api-access-b7ztx\") pod \"route-controller-manager-75ddb6d6df-c2wt7\" (UID: \"5d93f592-1d2d-49e1-a945-50d193c95177\") " pod="openshift-route-controller-manager/route-controller-manager-75ddb6d6df-c2wt7" Dec 07 16:07:38 crc kubenswrapper[4716]: I1207 16:07:38.247471 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5d93f592-1d2d-49e1-a945-50d193c95177-serving-cert\") pod \"route-controller-manager-75ddb6d6df-c2wt7\" (UID: \"5d93f592-1d2d-49e1-a945-50d193c95177\") " pod="openshift-route-controller-manager/route-controller-manager-75ddb6d6df-c2wt7" Dec 07 16:07:38 crc kubenswrapper[4716]: I1207 16:07:38.247651 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6acc423b-537c-4fc2-a098-78376fdc812c-client-ca" (OuterVolumeSpecName: "client-ca") pod "6acc423b-537c-4fc2-a098-78376fdc812c" (UID: "6acc423b-537c-4fc2-a098-78376fdc812c"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:07:38 crc kubenswrapper[4716]: I1207 16:07:38.247695 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6acc423b-537c-4fc2-a098-78376fdc812c-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "6acc423b-537c-4fc2-a098-78376fdc812c" (UID: "6acc423b-537c-4fc2-a098-78376fdc812c"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:07:38 crc kubenswrapper[4716]: I1207 16:07:38.248215 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6acc423b-537c-4fc2-a098-78376fdc812c-config" (OuterVolumeSpecName: "config") pod "6acc423b-537c-4fc2-a098-78376fdc812c" (UID: "6acc423b-537c-4fc2-a098-78376fdc812c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:07:38 crc kubenswrapper[4716]: I1207 16:07:38.248286 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ba7ae3c6-4861-4602-bba3-d72858464cb1-client-ca" (OuterVolumeSpecName: "client-ca") pod "ba7ae3c6-4861-4602-bba3-d72858464cb1" (UID: "ba7ae3c6-4861-4602-bba3-d72858464cb1"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:07:38 crc kubenswrapper[4716]: I1207 16:07:38.248306 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ba7ae3c6-4861-4602-bba3-d72858464cb1-config" (OuterVolumeSpecName: "config") pod "ba7ae3c6-4861-4602-bba3-d72858464cb1" (UID: "ba7ae3c6-4861-4602-bba3-d72858464cb1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:07:38 crc kubenswrapper[4716]: I1207 16:07:38.259239 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba7ae3c6-4861-4602-bba3-d72858464cb1-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "ba7ae3c6-4861-4602-bba3-d72858464cb1" (UID: "ba7ae3c6-4861-4602-bba3-d72858464cb1"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:07:38 crc kubenswrapper[4716]: I1207 16:07:38.259275 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba7ae3c6-4861-4602-bba3-d72858464cb1-kube-api-access-dxp46" (OuterVolumeSpecName: "kube-api-access-dxp46") pod "ba7ae3c6-4861-4602-bba3-d72858464cb1" (UID: "ba7ae3c6-4861-4602-bba3-d72858464cb1"). InnerVolumeSpecName "kube-api-access-dxp46". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:07:38 crc kubenswrapper[4716]: I1207 16:07:38.259295 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6acc423b-537c-4fc2-a098-78376fdc812c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6acc423b-537c-4fc2-a098-78376fdc812c" (UID: "6acc423b-537c-4fc2-a098-78376fdc812c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:07:38 crc kubenswrapper[4716]: I1207 16:07:38.259395 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6acc423b-537c-4fc2-a098-78376fdc812c-kube-api-access-jn7ql" (OuterVolumeSpecName: "kube-api-access-jn7ql") pod "6acc423b-537c-4fc2-a098-78376fdc812c" (UID: "6acc423b-537c-4fc2-a098-78376fdc812c"). InnerVolumeSpecName "kube-api-access-jn7ql". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:07:38 crc kubenswrapper[4716]: I1207 16:07:38.349188 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d93f592-1d2d-49e1-a945-50d193c95177-config\") pod \"route-controller-manager-75ddb6d6df-c2wt7\" (UID: \"5d93f592-1d2d-49e1-a945-50d193c95177\") " pod="openshift-route-controller-manager/route-controller-manager-75ddb6d6df-c2wt7" Dec 07 16:07:38 crc kubenswrapper[4716]: I1207 16:07:38.349250 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5d93f592-1d2d-49e1-a945-50d193c95177-client-ca\") pod \"route-controller-manager-75ddb6d6df-c2wt7\" (UID: \"5d93f592-1d2d-49e1-a945-50d193c95177\") " pod="openshift-route-controller-manager/route-controller-manager-75ddb6d6df-c2wt7" Dec 07 16:07:38 crc kubenswrapper[4716]: I1207 16:07:38.349274 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7ztx\" (UniqueName: \"kubernetes.io/projected/5d93f592-1d2d-49e1-a945-50d193c95177-kube-api-access-b7ztx\") pod \"route-controller-manager-75ddb6d6df-c2wt7\" (UID: \"5d93f592-1d2d-49e1-a945-50d193c95177\") " pod="openshift-route-controller-manager/route-controller-manager-75ddb6d6df-c2wt7" Dec 07 16:07:38 crc kubenswrapper[4716]: I1207 16:07:38.349301 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5d93f592-1d2d-49e1-a945-50d193c95177-serving-cert\") pod \"route-controller-manager-75ddb6d6df-c2wt7\" (UID: \"5d93f592-1d2d-49e1-a945-50d193c95177\") " pod="openshift-route-controller-manager/route-controller-manager-75ddb6d6df-c2wt7" Dec 07 16:07:38 crc kubenswrapper[4716]: I1207 16:07:38.349355 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jn7ql\" (UniqueName: \"kubernetes.io/projected/6acc423b-537c-4fc2-a098-78376fdc812c-kube-api-access-jn7ql\") on node \"crc\" DevicePath \"\"" Dec 07 16:07:38 crc kubenswrapper[4716]: I1207 16:07:38.349388 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dxp46\" (UniqueName: \"kubernetes.io/projected/ba7ae3c6-4861-4602-bba3-d72858464cb1-kube-api-access-dxp46\") on node \"crc\" DevicePath \"\"" Dec 07 16:07:38 crc kubenswrapper[4716]: I1207 16:07:38.349401 4716 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6acc423b-537c-4fc2-a098-78376fdc812c-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 07 16:07:38 crc kubenswrapper[4716]: I1207 16:07:38.349491 4716 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6acc423b-537c-4fc2-a098-78376fdc812c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 07 16:07:38 crc kubenswrapper[4716]: I1207 16:07:38.349505 4716 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6acc423b-537c-4fc2-a098-78376fdc812c-config\") on node \"crc\" DevicePath \"\"" Dec 07 16:07:38 crc kubenswrapper[4716]: I1207 16:07:38.349518 4716 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ba7ae3c6-4861-4602-bba3-d72858464cb1-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 07 16:07:38 crc kubenswrapper[4716]: I1207 16:07:38.349528 4716 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6acc423b-537c-4fc2-a098-78376fdc812c-client-ca\") on node \"crc\" DevicePath \"\"" Dec 07 16:07:38 crc kubenswrapper[4716]: I1207 16:07:38.349543 4716 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ba7ae3c6-4861-4602-bba3-d72858464cb1-client-ca\") on node \"crc\" DevicePath \"\"" Dec 07 16:07:38 crc kubenswrapper[4716]: I1207 16:07:38.349554 4716 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba7ae3c6-4861-4602-bba3-d72858464cb1-config\") on node \"crc\" DevicePath \"\"" Dec 07 16:07:38 crc kubenswrapper[4716]: I1207 16:07:38.350227 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5d93f592-1d2d-49e1-a945-50d193c95177-client-ca\") pod \"route-controller-manager-75ddb6d6df-c2wt7\" (UID: \"5d93f592-1d2d-49e1-a945-50d193c95177\") " pod="openshift-route-controller-manager/route-controller-manager-75ddb6d6df-c2wt7" Dec 07 16:07:38 crc kubenswrapper[4716]: I1207 16:07:38.350582 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d93f592-1d2d-49e1-a945-50d193c95177-config\") pod \"route-controller-manager-75ddb6d6df-c2wt7\" (UID: \"5d93f592-1d2d-49e1-a945-50d193c95177\") " pod="openshift-route-controller-manager/route-controller-manager-75ddb6d6df-c2wt7" Dec 07 16:07:38 crc kubenswrapper[4716]: I1207 16:07:38.352921 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5d93f592-1d2d-49e1-a945-50d193c95177-serving-cert\") pod \"route-controller-manager-75ddb6d6df-c2wt7\" (UID: \"5d93f592-1d2d-49e1-a945-50d193c95177\") " pod="openshift-route-controller-manager/route-controller-manager-75ddb6d6df-c2wt7" Dec 07 16:07:38 crc kubenswrapper[4716]: I1207 16:07:38.366865 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7ztx\" (UniqueName: \"kubernetes.io/projected/5d93f592-1d2d-49e1-a945-50d193c95177-kube-api-access-b7ztx\") pod \"route-controller-manager-75ddb6d6df-c2wt7\" (UID: \"5d93f592-1d2d-49e1-a945-50d193c95177\") " pod="openshift-route-controller-manager/route-controller-manager-75ddb6d6df-c2wt7" Dec 07 16:07:38 crc kubenswrapper[4716]: I1207 16:07:38.486005 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-75ddb6d6df-c2wt7" Dec 07 16:07:38 crc kubenswrapper[4716]: I1207 16:07:38.898289 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-75ddb6d6df-c2wt7"] Dec 07 16:07:38 crc kubenswrapper[4716]: W1207 16:07:38.903234 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5d93f592_1d2d_49e1_a945_50d193c95177.slice/crio-2426f446b954d40160268af040db45fb794c814f84c302d2058d1950b51a780f WatchSource:0}: Error finding container 2426f446b954d40160268af040db45fb794c814f84c302d2058d1950b51a780f: Status 404 returned error can't find the container with id 2426f446b954d40160268af040db45fb794c814f84c302d2058d1950b51a780f Dec 07 16:07:38 crc kubenswrapper[4716]: I1207 16:07:38.946558 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6c7c66dbd8-dwgtq" Dec 07 16:07:38 crc kubenswrapper[4716]: I1207 16:07:38.946579 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6c7c66dbd8-dwgtq" event={"ID":"ba7ae3c6-4861-4602-bba3-d72858464cb1","Type":"ContainerDied","Data":"79f65626ca28c3e24c5f9240ad2a08d394691dfbf784f63ed9fb2214896586ce"} Dec 07 16:07:38 crc kubenswrapper[4716]: I1207 16:07:38.946639 4716 scope.go:117] "RemoveContainer" containerID="5d9ccfb70b20aed39972a147694629c066608698614a5c95dcf9c4ad946245a3" Dec 07 16:07:38 crc kubenswrapper[4716]: I1207 16:07:38.948000 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-75ddb6d6df-c2wt7" event={"ID":"5d93f592-1d2d-49e1-a945-50d193c95177","Type":"ContainerStarted","Data":"2426f446b954d40160268af040db45fb794c814f84c302d2058d1950b51a780f"} Dec 07 16:07:38 crc kubenswrapper[4716]: I1207 16:07:38.949600 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7c9b544d8-h8g9h" event={"ID":"6acc423b-537c-4fc2-a098-78376fdc812c","Type":"ContainerDied","Data":"39d596a9d385e830993a5ef097e1edb70c139b7052c87907f46beb52f948a32c"} Dec 07 16:07:38 crc kubenswrapper[4716]: I1207 16:07:38.949641 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7c9b544d8-h8g9h" Dec 07 16:07:38 crc kubenswrapper[4716]: I1207 16:07:38.965840 4716 scope.go:117] "RemoveContainer" containerID="56af6622941b689ea303da653b8532c81628a968eb1c273a55acfdcfcf8cc881" Dec 07 16:07:38 crc kubenswrapper[4716]: I1207 16:07:38.992703 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6c7c66dbd8-dwgtq"] Dec 07 16:07:38 crc kubenswrapper[4716]: I1207 16:07:38.997466 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6c7c66dbd8-dwgtq"] Dec 07 16:07:39 crc kubenswrapper[4716]: I1207 16:07:39.008968 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7c9b544d8-h8g9h"] Dec 07 16:07:39 crc kubenswrapper[4716]: I1207 16:07:39.012606 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-7c9b544d8-h8g9h"] Dec 07 16:07:39 crc kubenswrapper[4716]: I1207 16:07:39.665456 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6acc423b-537c-4fc2-a098-78376fdc812c" path="/var/lib/kubelet/pods/6acc423b-537c-4fc2-a098-78376fdc812c/volumes" Dec 07 16:07:39 crc kubenswrapper[4716]: I1207 16:07:39.666956 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ba7ae3c6-4861-4602-bba3-d72858464cb1" path="/var/lib/kubelet/pods/ba7ae3c6-4861-4602-bba3-d72858464cb1/volumes" Dec 07 16:07:39 crc kubenswrapper[4716]: I1207 16:07:39.960148 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-75ddb6d6df-c2wt7" event={"ID":"5d93f592-1d2d-49e1-a945-50d193c95177","Type":"ContainerStarted","Data":"978db67a1649e8426c2e61d55ba036aeb55582dde36057afb8c46118c397c526"} Dec 07 16:07:39 crc kubenswrapper[4716]: I1207 16:07:39.961540 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-75ddb6d6df-c2wt7" Dec 07 16:07:39 crc kubenswrapper[4716]: I1207 16:07:39.967553 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-75ddb6d6df-c2wt7" Dec 07 16:07:39 crc kubenswrapper[4716]: I1207 16:07:39.983194 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-75ddb6d6df-c2wt7" podStartSLOduration=6.983161042 podStartE2EDuration="6.983161042s" podCreationTimestamp="2025-12-07 16:07:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:07:39.978460526 +0000 UTC m=+322.668745508" watchObservedRunningTime="2025-12-07 16:07:39.983161042 +0000 UTC m=+322.673445994" Dec 07 16:07:41 crc kubenswrapper[4716]: I1207 16:07:41.052122 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7499785678-9s5cn"] Dec 07 16:07:41 crc kubenswrapper[4716]: I1207 16:07:41.053174 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7499785678-9s5cn" Dec 07 16:07:41 crc kubenswrapper[4716]: I1207 16:07:41.059779 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 07 16:07:41 crc kubenswrapper[4716]: I1207 16:07:41.060660 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 07 16:07:41 crc kubenswrapper[4716]: I1207 16:07:41.062164 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 07 16:07:41 crc kubenswrapper[4716]: I1207 16:07:41.062389 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 07 16:07:41 crc kubenswrapper[4716]: I1207 16:07:41.062497 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 07 16:07:41 crc kubenswrapper[4716]: I1207 16:07:41.063350 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 07 16:07:41 crc kubenswrapper[4716]: I1207 16:07:41.066257 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 07 16:07:41 crc kubenswrapper[4716]: I1207 16:07:41.067562 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7499785678-9s5cn"] Dec 07 16:07:41 crc kubenswrapper[4716]: I1207 16:07:41.079465 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqkbt\" (UniqueName: \"kubernetes.io/projected/0772a86b-4102-4edb-903e-e14d28646e5f-kube-api-access-dqkbt\") pod \"controller-manager-7499785678-9s5cn\" (UID: \"0772a86b-4102-4edb-903e-e14d28646e5f\") " pod="openshift-controller-manager/controller-manager-7499785678-9s5cn" Dec 07 16:07:41 crc kubenswrapper[4716]: I1207 16:07:41.079521 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0772a86b-4102-4edb-903e-e14d28646e5f-client-ca\") pod \"controller-manager-7499785678-9s5cn\" (UID: \"0772a86b-4102-4edb-903e-e14d28646e5f\") " pod="openshift-controller-manager/controller-manager-7499785678-9s5cn" Dec 07 16:07:41 crc kubenswrapper[4716]: I1207 16:07:41.079552 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0772a86b-4102-4edb-903e-e14d28646e5f-serving-cert\") pod \"controller-manager-7499785678-9s5cn\" (UID: \"0772a86b-4102-4edb-903e-e14d28646e5f\") " pod="openshift-controller-manager/controller-manager-7499785678-9s5cn" Dec 07 16:07:41 crc kubenswrapper[4716]: I1207 16:07:41.079625 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0772a86b-4102-4edb-903e-e14d28646e5f-proxy-ca-bundles\") pod \"controller-manager-7499785678-9s5cn\" (UID: \"0772a86b-4102-4edb-903e-e14d28646e5f\") " pod="openshift-controller-manager/controller-manager-7499785678-9s5cn" Dec 07 16:07:41 crc kubenswrapper[4716]: I1207 16:07:41.079701 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0772a86b-4102-4edb-903e-e14d28646e5f-config\") pod \"controller-manager-7499785678-9s5cn\" (UID: \"0772a86b-4102-4edb-903e-e14d28646e5f\") " pod="openshift-controller-manager/controller-manager-7499785678-9s5cn" Dec 07 16:07:41 crc kubenswrapper[4716]: I1207 16:07:41.181420 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0772a86b-4102-4edb-903e-e14d28646e5f-proxy-ca-bundles\") pod \"controller-manager-7499785678-9s5cn\" (UID: \"0772a86b-4102-4edb-903e-e14d28646e5f\") " pod="openshift-controller-manager/controller-manager-7499785678-9s5cn" Dec 07 16:07:41 crc kubenswrapper[4716]: I1207 16:07:41.181464 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0772a86b-4102-4edb-903e-e14d28646e5f-config\") pod \"controller-manager-7499785678-9s5cn\" (UID: \"0772a86b-4102-4edb-903e-e14d28646e5f\") " pod="openshift-controller-manager/controller-manager-7499785678-9s5cn" Dec 07 16:07:41 crc kubenswrapper[4716]: I1207 16:07:41.181511 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqkbt\" (UniqueName: \"kubernetes.io/projected/0772a86b-4102-4edb-903e-e14d28646e5f-kube-api-access-dqkbt\") pod \"controller-manager-7499785678-9s5cn\" (UID: \"0772a86b-4102-4edb-903e-e14d28646e5f\") " pod="openshift-controller-manager/controller-manager-7499785678-9s5cn" Dec 07 16:07:41 crc kubenswrapper[4716]: I1207 16:07:41.181537 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0772a86b-4102-4edb-903e-e14d28646e5f-client-ca\") pod \"controller-manager-7499785678-9s5cn\" (UID: \"0772a86b-4102-4edb-903e-e14d28646e5f\") " pod="openshift-controller-manager/controller-manager-7499785678-9s5cn" Dec 07 16:07:41 crc kubenswrapper[4716]: I1207 16:07:41.181552 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0772a86b-4102-4edb-903e-e14d28646e5f-serving-cert\") pod \"controller-manager-7499785678-9s5cn\" (UID: \"0772a86b-4102-4edb-903e-e14d28646e5f\") " pod="openshift-controller-manager/controller-manager-7499785678-9s5cn" Dec 07 16:07:41 crc kubenswrapper[4716]: I1207 16:07:41.183292 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0772a86b-4102-4edb-903e-e14d28646e5f-client-ca\") pod \"controller-manager-7499785678-9s5cn\" (UID: \"0772a86b-4102-4edb-903e-e14d28646e5f\") " pod="openshift-controller-manager/controller-manager-7499785678-9s5cn" Dec 07 16:07:41 crc kubenswrapper[4716]: I1207 16:07:41.183411 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0772a86b-4102-4edb-903e-e14d28646e5f-config\") pod \"controller-manager-7499785678-9s5cn\" (UID: \"0772a86b-4102-4edb-903e-e14d28646e5f\") " pod="openshift-controller-manager/controller-manager-7499785678-9s5cn" Dec 07 16:07:41 crc kubenswrapper[4716]: I1207 16:07:41.184129 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0772a86b-4102-4edb-903e-e14d28646e5f-proxy-ca-bundles\") pod \"controller-manager-7499785678-9s5cn\" (UID: \"0772a86b-4102-4edb-903e-e14d28646e5f\") " pod="openshift-controller-manager/controller-manager-7499785678-9s5cn" Dec 07 16:07:41 crc kubenswrapper[4716]: I1207 16:07:41.188126 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0772a86b-4102-4edb-903e-e14d28646e5f-serving-cert\") pod \"controller-manager-7499785678-9s5cn\" (UID: \"0772a86b-4102-4edb-903e-e14d28646e5f\") " pod="openshift-controller-manager/controller-manager-7499785678-9s5cn" Dec 07 16:07:41 crc kubenswrapper[4716]: I1207 16:07:41.197298 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqkbt\" (UniqueName: \"kubernetes.io/projected/0772a86b-4102-4edb-903e-e14d28646e5f-kube-api-access-dqkbt\") pod \"controller-manager-7499785678-9s5cn\" (UID: \"0772a86b-4102-4edb-903e-e14d28646e5f\") " pod="openshift-controller-manager/controller-manager-7499785678-9s5cn" Dec 07 16:07:41 crc kubenswrapper[4716]: I1207 16:07:41.383707 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7499785678-9s5cn" Dec 07 16:07:41 crc kubenswrapper[4716]: I1207 16:07:41.802420 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7499785678-9s5cn"] Dec 07 16:07:41 crc kubenswrapper[4716]: I1207 16:07:41.970849 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7499785678-9s5cn" event={"ID":"0772a86b-4102-4edb-903e-e14d28646e5f","Type":"ContainerStarted","Data":"fbc0b03dabc2fa8b3feba0f2be5856dbc9f86acd2e0b734702ff1d510cfda665"} Dec 07 16:07:41 crc kubenswrapper[4716]: I1207 16:07:41.971161 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7499785678-9s5cn" event={"ID":"0772a86b-4102-4edb-903e-e14d28646e5f","Type":"ContainerStarted","Data":"fe5b0cca63db49c92affe616de21fa2eb1fcd0d7d89c357a96c5c32383138e07"} Dec 07 16:07:41 crc kubenswrapper[4716]: I1207 16:07:41.971181 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7499785678-9s5cn" Dec 07 16:07:41 crc kubenswrapper[4716]: I1207 16:07:41.972341 4716 patch_prober.go:28] interesting pod/controller-manager-7499785678-9s5cn container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.66:8443/healthz\": dial tcp 10.217.0.66:8443: connect: connection refused" start-of-body= Dec 07 16:07:41 crc kubenswrapper[4716]: I1207 16:07:41.972389 4716 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-7499785678-9s5cn" podUID="0772a86b-4102-4edb-903e-e14d28646e5f" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.66:8443/healthz\": dial tcp 10.217.0.66:8443: connect: connection refused" Dec 07 16:07:41 crc kubenswrapper[4716]: I1207 16:07:41.987158 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7499785678-9s5cn" podStartSLOduration=6.987136584 podStartE2EDuration="6.987136584s" podCreationTimestamp="2025-12-07 16:07:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:07:41.983460675 +0000 UTC m=+324.673745597" watchObservedRunningTime="2025-12-07 16:07:41.987136584 +0000 UTC m=+324.677421496" Dec 07 16:07:42 crc kubenswrapper[4716]: I1207 16:07:42.980347 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7499785678-9s5cn" Dec 07 16:07:55 crc kubenswrapper[4716]: I1207 16:07:55.433169 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-5btgg"] Dec 07 16:07:55 crc kubenswrapper[4716]: I1207 16:07:55.434927 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-5btgg" Dec 07 16:07:55 crc kubenswrapper[4716]: I1207 16:07:55.448282 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-5btgg"] Dec 07 16:07:55 crc kubenswrapper[4716]: I1207 16:07:55.476210 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2b9be5e9-7b17-4309-9143-9bfe0beae546-trusted-ca\") pod \"image-registry-66df7c8f76-5btgg\" (UID: \"2b9be5e9-7b17-4309-9143-9bfe0beae546\") " pod="openshift-image-registry/image-registry-66df7c8f76-5btgg" Dec 07 16:07:55 crc kubenswrapper[4716]: I1207 16:07:55.476730 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2b9be5e9-7b17-4309-9143-9bfe0beae546-ca-trust-extracted\") pod \"image-registry-66df7c8f76-5btgg\" (UID: \"2b9be5e9-7b17-4309-9143-9bfe0beae546\") " pod="openshift-image-registry/image-registry-66df7c8f76-5btgg" Dec 07 16:07:55 crc kubenswrapper[4716]: I1207 16:07:55.476796 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-5btgg\" (UID: \"2b9be5e9-7b17-4309-9143-9bfe0beae546\") " pod="openshift-image-registry/image-registry-66df7c8f76-5btgg" Dec 07 16:07:55 crc kubenswrapper[4716]: I1207 16:07:55.476818 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2b9be5e9-7b17-4309-9143-9bfe0beae546-registry-certificates\") pod \"image-registry-66df7c8f76-5btgg\" (UID: \"2b9be5e9-7b17-4309-9143-9bfe0beae546\") " pod="openshift-image-registry/image-registry-66df7c8f76-5btgg" Dec 07 16:07:55 crc kubenswrapper[4716]: I1207 16:07:55.476838 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2b9be5e9-7b17-4309-9143-9bfe0beae546-installation-pull-secrets\") pod \"image-registry-66df7c8f76-5btgg\" (UID: \"2b9be5e9-7b17-4309-9143-9bfe0beae546\") " pod="openshift-image-registry/image-registry-66df7c8f76-5btgg" Dec 07 16:07:55 crc kubenswrapper[4716]: I1207 16:07:55.476853 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2b9be5e9-7b17-4309-9143-9bfe0beae546-bound-sa-token\") pod \"image-registry-66df7c8f76-5btgg\" (UID: \"2b9be5e9-7b17-4309-9143-9bfe0beae546\") " pod="openshift-image-registry/image-registry-66df7c8f76-5btgg" Dec 07 16:07:55 crc kubenswrapper[4716]: I1207 16:07:55.476873 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2b9be5e9-7b17-4309-9143-9bfe0beae546-registry-tls\") pod \"image-registry-66df7c8f76-5btgg\" (UID: \"2b9be5e9-7b17-4309-9143-9bfe0beae546\") " pod="openshift-image-registry/image-registry-66df7c8f76-5btgg" Dec 07 16:07:55 crc kubenswrapper[4716]: I1207 16:07:55.476924 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jtn9f\" (UniqueName: \"kubernetes.io/projected/2b9be5e9-7b17-4309-9143-9bfe0beae546-kube-api-access-jtn9f\") pod \"image-registry-66df7c8f76-5btgg\" (UID: \"2b9be5e9-7b17-4309-9143-9bfe0beae546\") " pod="openshift-image-registry/image-registry-66df7c8f76-5btgg" Dec 07 16:07:55 crc kubenswrapper[4716]: I1207 16:07:55.499257 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-5btgg\" (UID: \"2b9be5e9-7b17-4309-9143-9bfe0beae546\") " pod="openshift-image-registry/image-registry-66df7c8f76-5btgg" Dec 07 16:07:55 crc kubenswrapper[4716]: I1207 16:07:55.578514 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2b9be5e9-7b17-4309-9143-9bfe0beae546-installation-pull-secrets\") pod \"image-registry-66df7c8f76-5btgg\" (UID: \"2b9be5e9-7b17-4309-9143-9bfe0beae546\") " pod="openshift-image-registry/image-registry-66df7c8f76-5btgg" Dec 07 16:07:55 crc kubenswrapper[4716]: I1207 16:07:55.578575 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2b9be5e9-7b17-4309-9143-9bfe0beae546-bound-sa-token\") pod \"image-registry-66df7c8f76-5btgg\" (UID: \"2b9be5e9-7b17-4309-9143-9bfe0beae546\") " pod="openshift-image-registry/image-registry-66df7c8f76-5btgg" Dec 07 16:07:55 crc kubenswrapper[4716]: I1207 16:07:55.578616 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2b9be5e9-7b17-4309-9143-9bfe0beae546-registry-tls\") pod \"image-registry-66df7c8f76-5btgg\" (UID: \"2b9be5e9-7b17-4309-9143-9bfe0beae546\") " pod="openshift-image-registry/image-registry-66df7c8f76-5btgg" Dec 07 16:07:55 crc kubenswrapper[4716]: I1207 16:07:55.578685 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jtn9f\" (UniqueName: \"kubernetes.io/projected/2b9be5e9-7b17-4309-9143-9bfe0beae546-kube-api-access-jtn9f\") pod \"image-registry-66df7c8f76-5btgg\" (UID: \"2b9be5e9-7b17-4309-9143-9bfe0beae546\") " pod="openshift-image-registry/image-registry-66df7c8f76-5btgg" Dec 07 16:07:55 crc kubenswrapper[4716]: I1207 16:07:55.578729 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2b9be5e9-7b17-4309-9143-9bfe0beae546-trusted-ca\") pod \"image-registry-66df7c8f76-5btgg\" (UID: \"2b9be5e9-7b17-4309-9143-9bfe0beae546\") " pod="openshift-image-registry/image-registry-66df7c8f76-5btgg" Dec 07 16:07:55 crc kubenswrapper[4716]: I1207 16:07:55.578753 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2b9be5e9-7b17-4309-9143-9bfe0beae546-ca-trust-extracted\") pod \"image-registry-66df7c8f76-5btgg\" (UID: \"2b9be5e9-7b17-4309-9143-9bfe0beae546\") " pod="openshift-image-registry/image-registry-66df7c8f76-5btgg" Dec 07 16:07:55 crc kubenswrapper[4716]: I1207 16:07:55.578800 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2b9be5e9-7b17-4309-9143-9bfe0beae546-registry-certificates\") pod \"image-registry-66df7c8f76-5btgg\" (UID: \"2b9be5e9-7b17-4309-9143-9bfe0beae546\") " pod="openshift-image-registry/image-registry-66df7c8f76-5btgg" Dec 07 16:07:55 crc kubenswrapper[4716]: I1207 16:07:55.579539 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2b9be5e9-7b17-4309-9143-9bfe0beae546-ca-trust-extracted\") pod \"image-registry-66df7c8f76-5btgg\" (UID: \"2b9be5e9-7b17-4309-9143-9bfe0beae546\") " pod="openshift-image-registry/image-registry-66df7c8f76-5btgg" Dec 07 16:07:55 crc kubenswrapper[4716]: I1207 16:07:55.580125 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2b9be5e9-7b17-4309-9143-9bfe0beae546-registry-certificates\") pod \"image-registry-66df7c8f76-5btgg\" (UID: \"2b9be5e9-7b17-4309-9143-9bfe0beae546\") " pod="openshift-image-registry/image-registry-66df7c8f76-5btgg" Dec 07 16:07:55 crc kubenswrapper[4716]: I1207 16:07:55.580185 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2b9be5e9-7b17-4309-9143-9bfe0beae546-trusted-ca\") pod \"image-registry-66df7c8f76-5btgg\" (UID: \"2b9be5e9-7b17-4309-9143-9bfe0beae546\") " pod="openshift-image-registry/image-registry-66df7c8f76-5btgg" Dec 07 16:07:55 crc kubenswrapper[4716]: I1207 16:07:55.585324 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2b9be5e9-7b17-4309-9143-9bfe0beae546-registry-tls\") pod \"image-registry-66df7c8f76-5btgg\" (UID: \"2b9be5e9-7b17-4309-9143-9bfe0beae546\") " pod="openshift-image-registry/image-registry-66df7c8f76-5btgg" Dec 07 16:07:55 crc kubenswrapper[4716]: I1207 16:07:55.589611 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2b9be5e9-7b17-4309-9143-9bfe0beae546-installation-pull-secrets\") pod \"image-registry-66df7c8f76-5btgg\" (UID: \"2b9be5e9-7b17-4309-9143-9bfe0beae546\") " pod="openshift-image-registry/image-registry-66df7c8f76-5btgg" Dec 07 16:07:55 crc kubenswrapper[4716]: I1207 16:07:55.603227 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2b9be5e9-7b17-4309-9143-9bfe0beae546-bound-sa-token\") pod \"image-registry-66df7c8f76-5btgg\" (UID: \"2b9be5e9-7b17-4309-9143-9bfe0beae546\") " pod="openshift-image-registry/image-registry-66df7c8f76-5btgg" Dec 07 16:07:55 crc kubenswrapper[4716]: I1207 16:07:55.603755 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jtn9f\" (UniqueName: \"kubernetes.io/projected/2b9be5e9-7b17-4309-9143-9bfe0beae546-kube-api-access-jtn9f\") pod \"image-registry-66df7c8f76-5btgg\" (UID: \"2b9be5e9-7b17-4309-9143-9bfe0beae546\") " pod="openshift-image-registry/image-registry-66df7c8f76-5btgg" Dec 07 16:07:55 crc kubenswrapper[4716]: I1207 16:07:55.765658 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-5btgg" Dec 07 16:07:56 crc kubenswrapper[4716]: I1207 16:07:56.150627 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-5btgg"] Dec 07 16:07:56 crc kubenswrapper[4716]: W1207 16:07:56.169217 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2b9be5e9_7b17_4309_9143_9bfe0beae546.slice/crio-391a2e8174e571179314880131feff3523622edcae1f4d07caa610a9d105b1b4 WatchSource:0}: Error finding container 391a2e8174e571179314880131feff3523622edcae1f4d07caa610a9d105b1b4: Status 404 returned error can't find the container with id 391a2e8174e571179314880131feff3523622edcae1f4d07caa610a9d105b1b4 Dec 07 16:07:57 crc kubenswrapper[4716]: I1207 16:07:57.073862 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-5btgg" event={"ID":"2b9be5e9-7b17-4309-9143-9bfe0beae546","Type":"ContainerStarted","Data":"9299ac027018e7d9b12c563cf23a801f4bd80758fda77da170d7c8c634bf2f8c"} Dec 07 16:07:57 crc kubenswrapper[4716]: I1207 16:07:57.073945 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-5btgg" event={"ID":"2b9be5e9-7b17-4309-9143-9bfe0beae546","Type":"ContainerStarted","Data":"391a2e8174e571179314880131feff3523622edcae1f4d07caa610a9d105b1b4"} Dec 07 16:07:57 crc kubenswrapper[4716]: I1207 16:07:57.074138 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-5btgg" Dec 07 16:08:15 crc kubenswrapper[4716]: I1207 16:08:15.770589 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-5btgg" Dec 07 16:08:15 crc kubenswrapper[4716]: I1207 16:08:15.794280 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-5btgg" podStartSLOduration=20.7942636 podStartE2EDuration="20.7942636s" podCreationTimestamp="2025-12-07 16:07:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:07:57.1019751 +0000 UTC m=+339.792260022" watchObservedRunningTime="2025-12-07 16:08:15.7942636 +0000 UTC m=+358.484548512" Dec 07 16:08:15 crc kubenswrapper[4716]: I1207 16:08:15.823226 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-4d257"] Dec 07 16:08:22 crc kubenswrapper[4716]: I1207 16:08:22.761156 4716 patch_prober.go:28] interesting pod/machine-config-daemon-zcxxp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 16:08:22 crc kubenswrapper[4716]: I1207 16:08:22.762358 4716 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 16:08:30 crc kubenswrapper[4716]: I1207 16:08:30.665142 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-75ddb6d6df-c2wt7"] Dec 07 16:08:30 crc kubenswrapper[4716]: I1207 16:08:30.666405 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-75ddb6d6df-c2wt7" podUID="5d93f592-1d2d-49e1-a945-50d193c95177" containerName="route-controller-manager" containerID="cri-o://978db67a1649e8426c2e61d55ba036aeb55582dde36057afb8c46118c397c526" gracePeriod=30 Dec 07 16:08:31 crc kubenswrapper[4716]: I1207 16:08:31.001964 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-75ddb6d6df-c2wt7" Dec 07 16:08:31 crc kubenswrapper[4716]: I1207 16:08:31.196560 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5d93f592-1d2d-49e1-a945-50d193c95177-client-ca\") pod \"5d93f592-1d2d-49e1-a945-50d193c95177\" (UID: \"5d93f592-1d2d-49e1-a945-50d193c95177\") " Dec 07 16:08:31 crc kubenswrapper[4716]: I1207 16:08:31.196650 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5d93f592-1d2d-49e1-a945-50d193c95177-serving-cert\") pod \"5d93f592-1d2d-49e1-a945-50d193c95177\" (UID: \"5d93f592-1d2d-49e1-a945-50d193c95177\") " Dec 07 16:08:31 crc kubenswrapper[4716]: I1207 16:08:31.196680 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d93f592-1d2d-49e1-a945-50d193c95177-config\") pod \"5d93f592-1d2d-49e1-a945-50d193c95177\" (UID: \"5d93f592-1d2d-49e1-a945-50d193c95177\") " Dec 07 16:08:31 crc kubenswrapper[4716]: I1207 16:08:31.196770 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b7ztx\" (UniqueName: \"kubernetes.io/projected/5d93f592-1d2d-49e1-a945-50d193c95177-kube-api-access-b7ztx\") pod \"5d93f592-1d2d-49e1-a945-50d193c95177\" (UID: \"5d93f592-1d2d-49e1-a945-50d193c95177\") " Dec 07 16:08:31 crc kubenswrapper[4716]: I1207 16:08:31.197676 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5d93f592-1d2d-49e1-a945-50d193c95177-config" (OuterVolumeSpecName: "config") pod "5d93f592-1d2d-49e1-a945-50d193c95177" (UID: "5d93f592-1d2d-49e1-a945-50d193c95177"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:08:31 crc kubenswrapper[4716]: I1207 16:08:31.198148 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5d93f592-1d2d-49e1-a945-50d193c95177-client-ca" (OuterVolumeSpecName: "client-ca") pod "5d93f592-1d2d-49e1-a945-50d193c95177" (UID: "5d93f592-1d2d-49e1-a945-50d193c95177"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:08:31 crc kubenswrapper[4716]: I1207 16:08:31.202224 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d93f592-1d2d-49e1-a945-50d193c95177-kube-api-access-b7ztx" (OuterVolumeSpecName: "kube-api-access-b7ztx") pod "5d93f592-1d2d-49e1-a945-50d193c95177" (UID: "5d93f592-1d2d-49e1-a945-50d193c95177"). InnerVolumeSpecName "kube-api-access-b7ztx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:08:31 crc kubenswrapper[4716]: I1207 16:08:31.203722 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d93f592-1d2d-49e1-a945-50d193c95177-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5d93f592-1d2d-49e1-a945-50d193c95177" (UID: "5d93f592-1d2d-49e1-a945-50d193c95177"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:08:31 crc kubenswrapper[4716]: I1207 16:08:31.269924 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-75ddb6d6df-c2wt7" Dec 07 16:08:31 crc kubenswrapper[4716]: I1207 16:08:31.269985 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-75ddb6d6df-c2wt7" event={"ID":"5d93f592-1d2d-49e1-a945-50d193c95177","Type":"ContainerDied","Data":"978db67a1649e8426c2e61d55ba036aeb55582dde36057afb8c46118c397c526"} Dec 07 16:08:31 crc kubenswrapper[4716]: I1207 16:08:31.270035 4716 scope.go:117] "RemoveContainer" containerID="978db67a1649e8426c2e61d55ba036aeb55582dde36057afb8c46118c397c526" Dec 07 16:08:31 crc kubenswrapper[4716]: I1207 16:08:31.269916 4716 generic.go:334] "Generic (PLEG): container finished" podID="5d93f592-1d2d-49e1-a945-50d193c95177" containerID="978db67a1649e8426c2e61d55ba036aeb55582dde36057afb8c46118c397c526" exitCode=0 Dec 07 16:08:31 crc kubenswrapper[4716]: I1207 16:08:31.270294 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-75ddb6d6df-c2wt7" event={"ID":"5d93f592-1d2d-49e1-a945-50d193c95177","Type":"ContainerDied","Data":"2426f446b954d40160268af040db45fb794c814f84c302d2058d1950b51a780f"} Dec 07 16:08:31 crc kubenswrapper[4716]: I1207 16:08:31.284842 4716 scope.go:117] "RemoveContainer" containerID="978db67a1649e8426c2e61d55ba036aeb55582dde36057afb8c46118c397c526" Dec 07 16:08:31 crc kubenswrapper[4716]: E1207 16:08:31.285354 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"978db67a1649e8426c2e61d55ba036aeb55582dde36057afb8c46118c397c526\": container with ID starting with 978db67a1649e8426c2e61d55ba036aeb55582dde36057afb8c46118c397c526 not found: ID does not exist" containerID="978db67a1649e8426c2e61d55ba036aeb55582dde36057afb8c46118c397c526" Dec 07 16:08:31 crc kubenswrapper[4716]: I1207 16:08:31.285401 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"978db67a1649e8426c2e61d55ba036aeb55582dde36057afb8c46118c397c526"} err="failed to get container status \"978db67a1649e8426c2e61d55ba036aeb55582dde36057afb8c46118c397c526\": rpc error: code = NotFound desc = could not find container \"978db67a1649e8426c2e61d55ba036aeb55582dde36057afb8c46118c397c526\": container with ID starting with 978db67a1649e8426c2e61d55ba036aeb55582dde36057afb8c46118c397c526 not found: ID does not exist" Dec 07 16:08:31 crc kubenswrapper[4716]: I1207 16:08:31.298293 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b7ztx\" (UniqueName: \"kubernetes.io/projected/5d93f592-1d2d-49e1-a945-50d193c95177-kube-api-access-b7ztx\") on node \"crc\" DevicePath \"\"" Dec 07 16:08:31 crc kubenswrapper[4716]: I1207 16:08:31.298336 4716 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5d93f592-1d2d-49e1-a945-50d193c95177-client-ca\") on node \"crc\" DevicePath \"\"" Dec 07 16:08:31 crc kubenswrapper[4716]: I1207 16:08:31.298348 4716 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5d93f592-1d2d-49e1-a945-50d193c95177-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 07 16:08:31 crc kubenswrapper[4716]: I1207 16:08:31.298379 4716 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d93f592-1d2d-49e1-a945-50d193c95177-config\") on node \"crc\" DevicePath \"\"" Dec 07 16:08:31 crc kubenswrapper[4716]: I1207 16:08:31.303416 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-75ddb6d6df-c2wt7"] Dec 07 16:08:31 crc kubenswrapper[4716]: I1207 16:08:31.307157 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-75ddb6d6df-c2wt7"] Dec 07 16:08:31 crc kubenswrapper[4716]: I1207 16:08:31.669380 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d93f592-1d2d-49e1-a945-50d193c95177" path="/var/lib/kubelet/pods/5d93f592-1d2d-49e1-a945-50d193c95177/volumes" Dec 07 16:08:32 crc kubenswrapper[4716]: I1207 16:08:32.084457 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-898469885-kbfmx"] Dec 07 16:08:32 crc kubenswrapper[4716]: E1207 16:08:32.084694 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d93f592-1d2d-49e1-a945-50d193c95177" containerName="route-controller-manager" Dec 07 16:08:32 crc kubenswrapper[4716]: I1207 16:08:32.084708 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d93f592-1d2d-49e1-a945-50d193c95177" containerName="route-controller-manager" Dec 07 16:08:32 crc kubenswrapper[4716]: I1207 16:08:32.084799 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d93f592-1d2d-49e1-a945-50d193c95177" containerName="route-controller-manager" Dec 07 16:08:32 crc kubenswrapper[4716]: I1207 16:08:32.085204 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-898469885-kbfmx" Dec 07 16:08:32 crc kubenswrapper[4716]: I1207 16:08:32.090704 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 07 16:08:32 crc kubenswrapper[4716]: I1207 16:08:32.091860 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 07 16:08:32 crc kubenswrapper[4716]: I1207 16:08:32.092061 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 07 16:08:32 crc kubenswrapper[4716]: I1207 16:08:32.092288 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 07 16:08:32 crc kubenswrapper[4716]: I1207 16:08:32.092474 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 07 16:08:32 crc kubenswrapper[4716]: I1207 16:08:32.092656 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 07 16:08:32 crc kubenswrapper[4716]: I1207 16:08:32.095847 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-898469885-kbfmx"] Dec 07 16:08:32 crc kubenswrapper[4716]: I1207 16:08:32.208929 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/75578e5a-c6cd-4acc-b86a-3cc2a024652b-serving-cert\") pod \"route-controller-manager-898469885-kbfmx\" (UID: \"75578e5a-c6cd-4acc-b86a-3cc2a024652b\") " pod="openshift-route-controller-manager/route-controller-manager-898469885-kbfmx" Dec 07 16:08:32 crc kubenswrapper[4716]: I1207 16:08:32.208987 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/75578e5a-c6cd-4acc-b86a-3cc2a024652b-config\") pod \"route-controller-manager-898469885-kbfmx\" (UID: \"75578e5a-c6cd-4acc-b86a-3cc2a024652b\") " pod="openshift-route-controller-manager/route-controller-manager-898469885-kbfmx" Dec 07 16:08:32 crc kubenswrapper[4716]: I1207 16:08:32.209016 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/75578e5a-c6cd-4acc-b86a-3cc2a024652b-client-ca\") pod \"route-controller-manager-898469885-kbfmx\" (UID: \"75578e5a-c6cd-4acc-b86a-3cc2a024652b\") " pod="openshift-route-controller-manager/route-controller-manager-898469885-kbfmx" Dec 07 16:08:32 crc kubenswrapper[4716]: I1207 16:08:32.209190 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gkckv\" (UniqueName: \"kubernetes.io/projected/75578e5a-c6cd-4acc-b86a-3cc2a024652b-kube-api-access-gkckv\") pod \"route-controller-manager-898469885-kbfmx\" (UID: \"75578e5a-c6cd-4acc-b86a-3cc2a024652b\") " pod="openshift-route-controller-manager/route-controller-manager-898469885-kbfmx" Dec 07 16:08:32 crc kubenswrapper[4716]: I1207 16:08:32.310605 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/75578e5a-c6cd-4acc-b86a-3cc2a024652b-serving-cert\") pod \"route-controller-manager-898469885-kbfmx\" (UID: \"75578e5a-c6cd-4acc-b86a-3cc2a024652b\") " pod="openshift-route-controller-manager/route-controller-manager-898469885-kbfmx" Dec 07 16:08:32 crc kubenswrapper[4716]: I1207 16:08:32.310676 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/75578e5a-c6cd-4acc-b86a-3cc2a024652b-config\") pod \"route-controller-manager-898469885-kbfmx\" (UID: \"75578e5a-c6cd-4acc-b86a-3cc2a024652b\") " pod="openshift-route-controller-manager/route-controller-manager-898469885-kbfmx" Dec 07 16:08:32 crc kubenswrapper[4716]: I1207 16:08:32.310715 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/75578e5a-c6cd-4acc-b86a-3cc2a024652b-client-ca\") pod \"route-controller-manager-898469885-kbfmx\" (UID: \"75578e5a-c6cd-4acc-b86a-3cc2a024652b\") " pod="openshift-route-controller-manager/route-controller-manager-898469885-kbfmx" Dec 07 16:08:32 crc kubenswrapper[4716]: I1207 16:08:32.310741 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gkckv\" (UniqueName: \"kubernetes.io/projected/75578e5a-c6cd-4acc-b86a-3cc2a024652b-kube-api-access-gkckv\") pod \"route-controller-manager-898469885-kbfmx\" (UID: \"75578e5a-c6cd-4acc-b86a-3cc2a024652b\") " pod="openshift-route-controller-manager/route-controller-manager-898469885-kbfmx" Dec 07 16:08:32 crc kubenswrapper[4716]: I1207 16:08:32.311876 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/75578e5a-c6cd-4acc-b86a-3cc2a024652b-client-ca\") pod \"route-controller-manager-898469885-kbfmx\" (UID: \"75578e5a-c6cd-4acc-b86a-3cc2a024652b\") " pod="openshift-route-controller-manager/route-controller-manager-898469885-kbfmx" Dec 07 16:08:32 crc kubenswrapper[4716]: I1207 16:08:32.312018 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/75578e5a-c6cd-4acc-b86a-3cc2a024652b-config\") pod \"route-controller-manager-898469885-kbfmx\" (UID: \"75578e5a-c6cd-4acc-b86a-3cc2a024652b\") " pod="openshift-route-controller-manager/route-controller-manager-898469885-kbfmx" Dec 07 16:08:32 crc kubenswrapper[4716]: I1207 16:08:32.319830 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/75578e5a-c6cd-4acc-b86a-3cc2a024652b-serving-cert\") pod \"route-controller-manager-898469885-kbfmx\" (UID: \"75578e5a-c6cd-4acc-b86a-3cc2a024652b\") " pod="openshift-route-controller-manager/route-controller-manager-898469885-kbfmx" Dec 07 16:08:32 crc kubenswrapper[4716]: I1207 16:08:32.339211 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gkckv\" (UniqueName: \"kubernetes.io/projected/75578e5a-c6cd-4acc-b86a-3cc2a024652b-kube-api-access-gkckv\") pod \"route-controller-manager-898469885-kbfmx\" (UID: \"75578e5a-c6cd-4acc-b86a-3cc2a024652b\") " pod="openshift-route-controller-manager/route-controller-manager-898469885-kbfmx" Dec 07 16:08:32 crc kubenswrapper[4716]: I1207 16:08:32.411744 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-898469885-kbfmx" Dec 07 16:08:32 crc kubenswrapper[4716]: I1207 16:08:32.838714 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-898469885-kbfmx"] Dec 07 16:08:33 crc kubenswrapper[4716]: I1207 16:08:33.281841 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-898469885-kbfmx" event={"ID":"75578e5a-c6cd-4acc-b86a-3cc2a024652b","Type":"ContainerStarted","Data":"38288ad26ff82a41bc3d2f3a465d2e9f7d9c8673fd8a2a2419754087632844f0"} Dec 07 16:08:33 crc kubenswrapper[4716]: I1207 16:08:33.282182 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-898469885-kbfmx" event={"ID":"75578e5a-c6cd-4acc-b86a-3cc2a024652b","Type":"ContainerStarted","Data":"05e3bb4d44394888820817c04e5337f6d86e8ec3e7e0fe365c3804dcd89c0a2a"} Dec 07 16:08:33 crc kubenswrapper[4716]: I1207 16:08:33.282204 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-898469885-kbfmx" Dec 07 16:08:33 crc kubenswrapper[4716]: I1207 16:08:33.286726 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-898469885-kbfmx" Dec 07 16:08:33 crc kubenswrapper[4716]: I1207 16:08:33.305480 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-898469885-kbfmx" podStartSLOduration=3.305456515 podStartE2EDuration="3.305456515s" podCreationTimestamp="2025-12-07 16:08:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:08:33.302804565 +0000 UTC m=+375.993089477" watchObservedRunningTime="2025-12-07 16:08:33.305456515 +0000 UTC m=+375.995741427" Dec 07 16:08:40 crc kubenswrapper[4716]: I1207 16:08:40.867026 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-4d257" podUID="75569b15-b1c1-46bf-83f2-cd6eb0a38381" containerName="registry" containerID="cri-o://aba2079f6d3cf016e5420b1261ba461a13577e9a5f66685aeb124feec845dad6" gracePeriod=30 Dec 07 16:08:41 crc kubenswrapper[4716]: I1207 16:08:41.222738 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-4d257" Dec 07 16:08:41 crc kubenswrapper[4716]: I1207 16:08:41.322948 4716 generic.go:334] "Generic (PLEG): container finished" podID="75569b15-b1c1-46bf-83f2-cd6eb0a38381" containerID="aba2079f6d3cf016e5420b1261ba461a13577e9a5f66685aeb124feec845dad6" exitCode=0 Dec 07 16:08:41 crc kubenswrapper[4716]: I1207 16:08:41.322992 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-4d257" event={"ID":"75569b15-b1c1-46bf-83f2-cd6eb0a38381","Type":"ContainerDied","Data":"aba2079f6d3cf016e5420b1261ba461a13577e9a5f66685aeb124feec845dad6"} Dec 07 16:08:41 crc kubenswrapper[4716]: I1207 16:08:41.323017 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-4d257" event={"ID":"75569b15-b1c1-46bf-83f2-cd6eb0a38381","Type":"ContainerDied","Data":"84f55b0d6385b8e76bcf9155671e73688b1d730161e0cace0ed36c4b4b026162"} Dec 07 16:08:41 crc kubenswrapper[4716]: I1207 16:08:41.323033 4716 scope.go:117] "RemoveContainer" containerID="aba2079f6d3cf016e5420b1261ba461a13577e9a5f66685aeb124feec845dad6" Dec 07 16:08:41 crc kubenswrapper[4716]: I1207 16:08:41.323160 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-4d257" Dec 07 16:08:41 crc kubenswrapper[4716]: I1207 16:08:41.335110 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/75569b15-b1c1-46bf-83f2-cd6eb0a38381-registry-tls\") pod \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\" (UID: \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\") " Dec 07 16:08:41 crc kubenswrapper[4716]: I1207 16:08:41.335181 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/75569b15-b1c1-46bf-83f2-cd6eb0a38381-bound-sa-token\") pod \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\" (UID: \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\") " Dec 07 16:08:41 crc kubenswrapper[4716]: I1207 16:08:41.335216 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/75569b15-b1c1-46bf-83f2-cd6eb0a38381-installation-pull-secrets\") pod \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\" (UID: \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\") " Dec 07 16:08:41 crc kubenswrapper[4716]: I1207 16:08:41.335253 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/75569b15-b1c1-46bf-83f2-cd6eb0a38381-ca-trust-extracted\") pod \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\" (UID: \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\") " Dec 07 16:08:41 crc kubenswrapper[4716]: I1207 16:08:41.335272 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/75569b15-b1c1-46bf-83f2-cd6eb0a38381-registry-certificates\") pod \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\" (UID: \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\") " Dec 07 16:08:41 crc kubenswrapper[4716]: I1207 16:08:41.335291 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/75569b15-b1c1-46bf-83f2-cd6eb0a38381-trusted-ca\") pod \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\" (UID: \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\") " Dec 07 16:08:41 crc kubenswrapper[4716]: I1207 16:08:41.335400 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\" (UID: \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\") " Dec 07 16:08:41 crc kubenswrapper[4716]: I1207 16:08:41.335452 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hgpp9\" (UniqueName: \"kubernetes.io/projected/75569b15-b1c1-46bf-83f2-cd6eb0a38381-kube-api-access-hgpp9\") pod \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\" (UID: \"75569b15-b1c1-46bf-83f2-cd6eb0a38381\") " Dec 07 16:08:41 crc kubenswrapper[4716]: I1207 16:08:41.336035 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/75569b15-b1c1-46bf-83f2-cd6eb0a38381-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "75569b15-b1c1-46bf-83f2-cd6eb0a38381" (UID: "75569b15-b1c1-46bf-83f2-cd6eb0a38381"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:08:41 crc kubenswrapper[4716]: I1207 16:08:41.336222 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/75569b15-b1c1-46bf-83f2-cd6eb0a38381-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "75569b15-b1c1-46bf-83f2-cd6eb0a38381" (UID: "75569b15-b1c1-46bf-83f2-cd6eb0a38381"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:08:41 crc kubenswrapper[4716]: I1207 16:08:41.341358 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75569b15-b1c1-46bf-83f2-cd6eb0a38381-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "75569b15-b1c1-46bf-83f2-cd6eb0a38381" (UID: "75569b15-b1c1-46bf-83f2-cd6eb0a38381"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:08:41 crc kubenswrapper[4716]: I1207 16:08:41.341458 4716 scope.go:117] "RemoveContainer" containerID="aba2079f6d3cf016e5420b1261ba461a13577e9a5f66685aeb124feec845dad6" Dec 07 16:08:41 crc kubenswrapper[4716]: I1207 16:08:41.341504 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75569b15-b1c1-46bf-83f2-cd6eb0a38381-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "75569b15-b1c1-46bf-83f2-cd6eb0a38381" (UID: "75569b15-b1c1-46bf-83f2-cd6eb0a38381"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:08:41 crc kubenswrapper[4716]: I1207 16:08:41.342389 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75569b15-b1c1-46bf-83f2-cd6eb0a38381-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "75569b15-b1c1-46bf-83f2-cd6eb0a38381" (UID: "75569b15-b1c1-46bf-83f2-cd6eb0a38381"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:08:41 crc kubenswrapper[4716]: I1207 16:08:41.343021 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75569b15-b1c1-46bf-83f2-cd6eb0a38381-kube-api-access-hgpp9" (OuterVolumeSpecName: "kube-api-access-hgpp9") pod "75569b15-b1c1-46bf-83f2-cd6eb0a38381" (UID: "75569b15-b1c1-46bf-83f2-cd6eb0a38381"). InnerVolumeSpecName "kube-api-access-hgpp9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:08:41 crc kubenswrapper[4716]: E1207 16:08:41.344351 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aba2079f6d3cf016e5420b1261ba461a13577e9a5f66685aeb124feec845dad6\": container with ID starting with aba2079f6d3cf016e5420b1261ba461a13577e9a5f66685aeb124feec845dad6 not found: ID does not exist" containerID="aba2079f6d3cf016e5420b1261ba461a13577e9a5f66685aeb124feec845dad6" Dec 07 16:08:41 crc kubenswrapper[4716]: I1207 16:08:41.344420 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aba2079f6d3cf016e5420b1261ba461a13577e9a5f66685aeb124feec845dad6"} err="failed to get container status \"aba2079f6d3cf016e5420b1261ba461a13577e9a5f66685aeb124feec845dad6\": rpc error: code = NotFound desc = could not find container \"aba2079f6d3cf016e5420b1261ba461a13577e9a5f66685aeb124feec845dad6\": container with ID starting with aba2079f6d3cf016e5420b1261ba461a13577e9a5f66685aeb124feec845dad6 not found: ID does not exist" Dec 07 16:08:41 crc kubenswrapper[4716]: I1207 16:08:41.347201 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "75569b15-b1c1-46bf-83f2-cd6eb0a38381" (UID: "75569b15-b1c1-46bf-83f2-cd6eb0a38381"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 07 16:08:41 crc kubenswrapper[4716]: I1207 16:08:41.356545 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/75569b15-b1c1-46bf-83f2-cd6eb0a38381-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "75569b15-b1c1-46bf-83f2-cd6eb0a38381" (UID: "75569b15-b1c1-46bf-83f2-cd6eb0a38381"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:08:41 crc kubenswrapper[4716]: I1207 16:08:41.437022 4716 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/75569b15-b1c1-46bf-83f2-cd6eb0a38381-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 07 16:08:41 crc kubenswrapper[4716]: I1207 16:08:41.437056 4716 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/75569b15-b1c1-46bf-83f2-cd6eb0a38381-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 07 16:08:41 crc kubenswrapper[4716]: I1207 16:08:41.437116 4716 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/75569b15-b1c1-46bf-83f2-cd6eb0a38381-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 07 16:08:41 crc kubenswrapper[4716]: I1207 16:08:41.437131 4716 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/75569b15-b1c1-46bf-83f2-cd6eb0a38381-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 07 16:08:41 crc kubenswrapper[4716]: I1207 16:08:41.437141 4716 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/75569b15-b1c1-46bf-83f2-cd6eb0a38381-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 07 16:08:41 crc kubenswrapper[4716]: I1207 16:08:41.437151 4716 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/75569b15-b1c1-46bf-83f2-cd6eb0a38381-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 07 16:08:41 crc kubenswrapper[4716]: I1207 16:08:41.437192 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hgpp9\" (UniqueName: \"kubernetes.io/projected/75569b15-b1c1-46bf-83f2-cd6eb0a38381-kube-api-access-hgpp9\") on node \"crc\" DevicePath \"\"" Dec 07 16:08:41 crc kubenswrapper[4716]: I1207 16:08:41.670830 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-4d257"] Dec 07 16:08:41 crc kubenswrapper[4716]: I1207 16:08:41.671478 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-4d257"] Dec 07 16:08:43 crc kubenswrapper[4716]: I1207 16:08:43.670826 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75569b15-b1c1-46bf-83f2-cd6eb0a38381" path="/var/lib/kubelet/pods/75569b15-b1c1-46bf-83f2-cd6eb0a38381/volumes" Dec 07 16:08:52 crc kubenswrapper[4716]: I1207 16:08:52.761460 4716 patch_prober.go:28] interesting pod/machine-config-daemon-zcxxp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 16:08:52 crc kubenswrapper[4716]: I1207 16:08:52.762006 4716 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 16:09:22 crc kubenswrapper[4716]: I1207 16:09:22.762049 4716 patch_prober.go:28] interesting pod/machine-config-daemon-zcxxp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 16:09:22 crc kubenswrapper[4716]: I1207 16:09:22.762702 4716 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 16:09:22 crc kubenswrapper[4716]: I1207 16:09:22.762767 4716 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" Dec 07 16:09:22 crc kubenswrapper[4716]: I1207 16:09:22.763913 4716 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ca669356fa57e43a7c10e3a5b76ace008fedad9bdfdd4c2c358a8731bdaaf42b"} pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 07 16:09:22 crc kubenswrapper[4716]: I1207 16:09:22.764020 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" containerName="machine-config-daemon" containerID="cri-o://ca669356fa57e43a7c10e3a5b76ace008fedad9bdfdd4c2c358a8731bdaaf42b" gracePeriod=600 Dec 07 16:09:23 crc kubenswrapper[4716]: I1207 16:09:23.559161 4716 generic.go:334] "Generic (PLEG): container finished" podID="c15b59eb-565d-4556-a4ce-75afdf159dc8" containerID="ca669356fa57e43a7c10e3a5b76ace008fedad9bdfdd4c2c358a8731bdaaf42b" exitCode=0 Dec 07 16:09:23 crc kubenswrapper[4716]: I1207 16:09:23.559268 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" event={"ID":"c15b59eb-565d-4556-a4ce-75afdf159dc8","Type":"ContainerDied","Data":"ca669356fa57e43a7c10e3a5b76ace008fedad9bdfdd4c2c358a8731bdaaf42b"} Dec 07 16:09:23 crc kubenswrapper[4716]: I1207 16:09:23.559605 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" event={"ID":"c15b59eb-565d-4556-a4ce-75afdf159dc8","Type":"ContainerStarted","Data":"def7373f410cd12d2a355b52c07eee66525502687b1341f3cbd5919e77d0a06e"} Dec 07 16:09:23 crc kubenswrapper[4716]: I1207 16:09:23.559636 4716 scope.go:117] "RemoveContainer" containerID="8ccaf6c1e6a978eb978a1c207b6c4c00b23737f78b743f966579aca781e22e0b" Dec 07 16:11:52 crc kubenswrapper[4716]: I1207 16:11:52.760816 4716 patch_prober.go:28] interesting pod/machine-config-daemon-zcxxp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 16:11:52 crc kubenswrapper[4716]: I1207 16:11:52.761560 4716 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 16:12:07 crc kubenswrapper[4716]: I1207 16:12:07.721198 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-cgnm7"] Dec 07 16:12:07 crc kubenswrapper[4716]: E1207 16:12:07.722139 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75569b15-b1c1-46bf-83f2-cd6eb0a38381" containerName="registry" Dec 07 16:12:07 crc kubenswrapper[4716]: I1207 16:12:07.722156 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="75569b15-b1c1-46bf-83f2-cd6eb0a38381" containerName="registry" Dec 07 16:12:07 crc kubenswrapper[4716]: I1207 16:12:07.722266 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="75569b15-b1c1-46bf-83f2-cd6eb0a38381" containerName="registry" Dec 07 16:12:07 crc kubenswrapper[4716]: I1207 16:12:07.722701 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-cgnm7" Dec 07 16:12:07 crc kubenswrapper[4716]: I1207 16:12:07.725918 4716 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-8fd7t" Dec 07 16:12:07 crc kubenswrapper[4716]: I1207 16:12:07.726205 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Dec 07 16:12:07 crc kubenswrapper[4716]: I1207 16:12:07.726418 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Dec 07 16:12:07 crc kubenswrapper[4716]: I1207 16:12:07.733196 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-xb4z7"] Dec 07 16:12:07 crc kubenswrapper[4716]: I1207 16:12:07.733954 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-xb4z7" Dec 07 16:12:07 crc kubenswrapper[4716]: I1207 16:12:07.736186 4716 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-8jp7l" Dec 07 16:12:07 crc kubenswrapper[4716]: I1207 16:12:07.738196 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-cgnm7"] Dec 07 16:12:07 crc kubenswrapper[4716]: I1207 16:12:07.744912 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-ft9cq"] Dec 07 16:12:07 crc kubenswrapper[4716]: I1207 16:12:07.745809 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-ft9cq" Dec 07 16:12:07 crc kubenswrapper[4716]: I1207 16:12:07.747520 4716 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-vhhzq" Dec 07 16:12:07 crc kubenswrapper[4716]: I1207 16:12:07.750153 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-xb4z7"] Dec 07 16:12:07 crc kubenswrapper[4716]: I1207 16:12:07.756833 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-ft9cq"] Dec 07 16:12:07 crc kubenswrapper[4716]: I1207 16:12:07.910215 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vk4kl\" (UniqueName: \"kubernetes.io/projected/badea5ba-19c9-4716-87df-a4ef7248b50a-kube-api-access-vk4kl\") pod \"cert-manager-webhook-5655c58dd6-ft9cq\" (UID: \"badea5ba-19c9-4716-87df-a4ef7248b50a\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-ft9cq" Dec 07 16:12:07 crc kubenswrapper[4716]: I1207 16:12:07.910343 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8cc9h\" (UniqueName: \"kubernetes.io/projected/64d9b2cc-19be-4cd5-81d8-fda3cbf2228e-kube-api-access-8cc9h\") pod \"cert-manager-cainjector-7f985d654d-cgnm7\" (UID: \"64d9b2cc-19be-4cd5-81d8-fda3cbf2228e\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-cgnm7" Dec 07 16:12:07 crc kubenswrapper[4716]: I1207 16:12:07.910476 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ln86j\" (UniqueName: \"kubernetes.io/projected/55db1fd5-6c4f-4f27-888c-4d4b2445e603-kube-api-access-ln86j\") pod \"cert-manager-5b446d88c5-xb4z7\" (UID: \"55db1fd5-6c4f-4f27-888c-4d4b2445e603\") " pod="cert-manager/cert-manager-5b446d88c5-xb4z7" Dec 07 16:12:08 crc kubenswrapper[4716]: I1207 16:12:08.012001 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vk4kl\" (UniqueName: \"kubernetes.io/projected/badea5ba-19c9-4716-87df-a4ef7248b50a-kube-api-access-vk4kl\") pod \"cert-manager-webhook-5655c58dd6-ft9cq\" (UID: \"badea5ba-19c9-4716-87df-a4ef7248b50a\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-ft9cq" Dec 07 16:12:08 crc kubenswrapper[4716]: I1207 16:12:08.012423 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8cc9h\" (UniqueName: \"kubernetes.io/projected/64d9b2cc-19be-4cd5-81d8-fda3cbf2228e-kube-api-access-8cc9h\") pod \"cert-manager-cainjector-7f985d654d-cgnm7\" (UID: \"64d9b2cc-19be-4cd5-81d8-fda3cbf2228e\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-cgnm7" Dec 07 16:12:08 crc kubenswrapper[4716]: I1207 16:12:08.012665 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ln86j\" (UniqueName: \"kubernetes.io/projected/55db1fd5-6c4f-4f27-888c-4d4b2445e603-kube-api-access-ln86j\") pod \"cert-manager-5b446d88c5-xb4z7\" (UID: \"55db1fd5-6c4f-4f27-888c-4d4b2445e603\") " pod="cert-manager/cert-manager-5b446d88c5-xb4z7" Dec 07 16:12:08 crc kubenswrapper[4716]: I1207 16:12:08.035060 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ln86j\" (UniqueName: \"kubernetes.io/projected/55db1fd5-6c4f-4f27-888c-4d4b2445e603-kube-api-access-ln86j\") pod \"cert-manager-5b446d88c5-xb4z7\" (UID: \"55db1fd5-6c4f-4f27-888c-4d4b2445e603\") " pod="cert-manager/cert-manager-5b446d88c5-xb4z7" Dec 07 16:12:08 crc kubenswrapper[4716]: I1207 16:12:08.035479 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vk4kl\" (UniqueName: \"kubernetes.io/projected/badea5ba-19c9-4716-87df-a4ef7248b50a-kube-api-access-vk4kl\") pod \"cert-manager-webhook-5655c58dd6-ft9cq\" (UID: \"badea5ba-19c9-4716-87df-a4ef7248b50a\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-ft9cq" Dec 07 16:12:08 crc kubenswrapper[4716]: I1207 16:12:08.035710 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8cc9h\" (UniqueName: \"kubernetes.io/projected/64d9b2cc-19be-4cd5-81d8-fda3cbf2228e-kube-api-access-8cc9h\") pod \"cert-manager-cainjector-7f985d654d-cgnm7\" (UID: \"64d9b2cc-19be-4cd5-81d8-fda3cbf2228e\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-cgnm7" Dec 07 16:12:08 crc kubenswrapper[4716]: I1207 16:12:08.047203 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-cgnm7" Dec 07 16:12:08 crc kubenswrapper[4716]: I1207 16:12:08.058090 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-xb4z7" Dec 07 16:12:08 crc kubenswrapper[4716]: I1207 16:12:08.066350 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-ft9cq" Dec 07 16:12:08 crc kubenswrapper[4716]: I1207 16:12:08.308283 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-cgnm7"] Dec 07 16:12:08 crc kubenswrapper[4716]: I1207 16:12:08.322330 4716 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 07 16:12:08 crc kubenswrapper[4716]: I1207 16:12:08.468910 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-cgnm7" event={"ID":"64d9b2cc-19be-4cd5-81d8-fda3cbf2228e","Type":"ContainerStarted","Data":"1e54b3a7668ba2044db8f2a42614a19ac77b46dee2e3d1f0931b62360868b5c3"} Dec 07 16:12:08 crc kubenswrapper[4716]: I1207 16:12:08.557648 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-xb4z7"] Dec 07 16:12:08 crc kubenswrapper[4716]: I1207 16:12:08.564796 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-ft9cq"] Dec 07 16:12:08 crc kubenswrapper[4716]: W1207 16:12:08.565038 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod55db1fd5_6c4f_4f27_888c_4d4b2445e603.slice/crio-40d6ed5531d85e84fbb944630d9c63688acc620f8865d8f210a6574c5f0c39fc WatchSource:0}: Error finding container 40d6ed5531d85e84fbb944630d9c63688acc620f8865d8f210a6574c5f0c39fc: Status 404 returned error can't find the container with id 40d6ed5531d85e84fbb944630d9c63688acc620f8865d8f210a6574c5f0c39fc Dec 07 16:12:08 crc kubenswrapper[4716]: W1207 16:12:08.568944 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbadea5ba_19c9_4716_87df_a4ef7248b50a.slice/crio-43da9ef8bb07f128a8778eec1cd02a6667d7c69982552bb0d60256a4c9159fcb WatchSource:0}: Error finding container 43da9ef8bb07f128a8778eec1cd02a6667d7c69982552bb0d60256a4c9159fcb: Status 404 returned error can't find the container with id 43da9ef8bb07f128a8778eec1cd02a6667d7c69982552bb0d60256a4c9159fcb Dec 07 16:12:09 crc kubenswrapper[4716]: I1207 16:12:09.476769 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-ft9cq" event={"ID":"badea5ba-19c9-4716-87df-a4ef7248b50a","Type":"ContainerStarted","Data":"43da9ef8bb07f128a8778eec1cd02a6667d7c69982552bb0d60256a4c9159fcb"} Dec 07 16:12:09 crc kubenswrapper[4716]: I1207 16:12:09.477960 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-xb4z7" event={"ID":"55db1fd5-6c4f-4f27-888c-4d4b2445e603","Type":"ContainerStarted","Data":"40d6ed5531d85e84fbb944630d9c63688acc620f8865d8f210a6574c5f0c39fc"} Dec 07 16:12:10 crc kubenswrapper[4716]: I1207 16:12:10.484341 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-cgnm7" event={"ID":"64d9b2cc-19be-4cd5-81d8-fda3cbf2228e","Type":"ContainerStarted","Data":"bd221f4f6e46065cc16728a1f5b3c1d0651384b062a53c409262a86debe71b79"} Dec 07 16:12:10 crc kubenswrapper[4716]: I1207 16:12:10.497743 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-cgnm7" podStartSLOduration=1.716248855 podStartE2EDuration="3.497728278s" podCreationTimestamp="2025-12-07 16:12:07 +0000 UTC" firstStartedPulling="2025-12-07 16:12:08.322049167 +0000 UTC m=+591.012334079" lastFinishedPulling="2025-12-07 16:12:10.10352859 +0000 UTC m=+592.793813502" observedRunningTime="2025-12-07 16:12:10.494454831 +0000 UTC m=+593.184739743" watchObservedRunningTime="2025-12-07 16:12:10.497728278 +0000 UTC m=+593.188013190" Dec 07 16:12:12 crc kubenswrapper[4716]: I1207 16:12:12.497145 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-ft9cq" event={"ID":"badea5ba-19c9-4716-87df-a4ef7248b50a","Type":"ContainerStarted","Data":"1e954a0b1449beef53c5f09313a81238d96ef0757ad436fa68cc22ece84537f9"} Dec 07 16:12:12 crc kubenswrapper[4716]: I1207 16:12:12.497565 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-ft9cq" Dec 07 16:12:12 crc kubenswrapper[4716]: I1207 16:12:12.499410 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-xb4z7" event={"ID":"55db1fd5-6c4f-4f27-888c-4d4b2445e603","Type":"ContainerStarted","Data":"abe1fa9ee0d0920ff23729247a04d23e133aed4bdc598618373a6eb3fbc13668"} Dec 07 16:12:12 crc kubenswrapper[4716]: I1207 16:12:12.518344 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-ft9cq" podStartSLOduration=2.419089293 podStartE2EDuration="5.518323733s" podCreationTimestamp="2025-12-07 16:12:07 +0000 UTC" firstStartedPulling="2025-12-07 16:12:08.5713044 +0000 UTC m=+591.261589352" lastFinishedPulling="2025-12-07 16:12:11.67053888 +0000 UTC m=+594.360823792" observedRunningTime="2025-12-07 16:12:12.516477224 +0000 UTC m=+595.206762156" watchObservedRunningTime="2025-12-07 16:12:12.518323733 +0000 UTC m=+595.208608645" Dec 07 16:12:12 crc kubenswrapper[4716]: I1207 16:12:12.536281 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-xb4z7" podStartSLOduration=2.369494789 podStartE2EDuration="5.536262693s" podCreationTimestamp="2025-12-07 16:12:07 +0000 UTC" firstStartedPulling="2025-12-07 16:12:08.567353915 +0000 UTC m=+591.257638837" lastFinishedPulling="2025-12-07 16:12:11.734121809 +0000 UTC m=+594.424406741" observedRunningTime="2025-12-07 16:12:12.533223832 +0000 UTC m=+595.223508744" watchObservedRunningTime="2025-12-07 16:12:12.536262693 +0000 UTC m=+595.226547605" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.507375 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-qvc7p"] Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.508481 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" podUID="c54d618a-dfec-4f2a-a288-f5188c8a128a" containerName="ovn-controller" containerID="cri-o://dcae78a6133d97f6a32dd7421ccec4333b970a237404e056a5cc601f8eefdc8d" gracePeriod=30 Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.508586 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" podUID="c54d618a-dfec-4f2a-a288-f5188c8a128a" containerName="northd" containerID="cri-o://b73be664f5412c98b89d971a2cac37c494151c293fdb44a98b390978658a6643" gracePeriod=30 Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.508629 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" podUID="c54d618a-dfec-4f2a-a288-f5188c8a128a" containerName="kube-rbac-proxy-node" containerID="cri-o://60324dc7d701bc064e14ae4f474bce10f6ee4ff3c3c942eab749e7166082cf9d" gracePeriod=30 Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.508649 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" podUID="c54d618a-dfec-4f2a-a288-f5188c8a128a" containerName="ovn-acl-logging" containerID="cri-o://a847d04e22e98091c0037a89054b9236228b701b8c7ba31940a56df0f37b95c6" gracePeriod=30 Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.508589 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" podUID="c54d618a-dfec-4f2a-a288-f5188c8a128a" containerName="sbdb" containerID="cri-o://bcfdffe822635b0ae13b253e6133449fe00461b2792be43d913c4a77219b8255" gracePeriod=30 Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.508811 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" podUID="c54d618a-dfec-4f2a-a288-f5188c8a128a" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://2ed4f4d3564674bf5cb4e7a76dc0fd95e203ecc8950798e59a1fb6c9ab4a20f7" gracePeriod=30 Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.508813 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" podUID="c54d618a-dfec-4f2a-a288-f5188c8a128a" containerName="nbdb" containerID="cri-o://def9ac840fb8ecf2778bed88ea980eddca7c5b2aba4bd1afb967d62fec3f7bae" gracePeriod=30 Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.572623 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" podUID="c54d618a-dfec-4f2a-a288-f5188c8a128a" containerName="ovnkube-controller" containerID="cri-o://e462e925550f957641f83d75b2990297a4cfd973b9227fb08dc610a75b1d6e79" gracePeriod=30 Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.796671 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qvc7p_c54d618a-dfec-4f2a-a288-f5188c8a128a/ovnkube-controller/3.log" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.798983 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qvc7p_c54d618a-dfec-4f2a-a288-f5188c8a128a/ovn-acl-logging/0.log" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.799742 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qvc7p_c54d618a-dfec-4f2a-a288-f5188c8a128a/ovn-controller/0.log" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.801030 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.831774 4716 scope.go:117] "RemoveContainer" containerID="60324dc7d701bc064e14ae4f474bce10f6ee4ff3c3c942eab749e7166082cf9d" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.847520 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-node-log\") pod \"c54d618a-dfec-4f2a-a288-f5188c8a128a\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.847611 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-node-log" (OuterVolumeSpecName: "node-log") pod "c54d618a-dfec-4f2a-a288-f5188c8a128a" (UID: "c54d618a-dfec-4f2a-a288-f5188c8a128a"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.847676 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-host-cni-bin\") pod \"c54d618a-dfec-4f2a-a288-f5188c8a128a\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.847702 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-run-systemd\") pod \"c54d618a-dfec-4f2a-a288-f5188c8a128a\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.847746 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "c54d618a-dfec-4f2a-a288-f5188c8a128a" (UID: "c54d618a-dfec-4f2a-a288-f5188c8a128a"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.847777 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-run-ovn\") pod \"c54d618a-dfec-4f2a-a288-f5188c8a128a\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.847792 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-log-socket\") pod \"c54d618a-dfec-4f2a-a288-f5188c8a128a\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.847837 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "c54d618a-dfec-4f2a-a288-f5188c8a128a" (UID: "c54d618a-dfec-4f2a-a288-f5188c8a128a"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.847965 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-host-slash" (OuterVolumeSpecName: "host-slash") pod "c54d618a-dfec-4f2a-a288-f5188c8a128a" (UID: "c54d618a-dfec-4f2a-a288-f5188c8a128a"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.847958 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-log-socket" (OuterVolumeSpecName: "log-socket") pod "c54d618a-dfec-4f2a-a288-f5188c8a128a" (UID: "c54d618a-dfec-4f2a-a288-f5188c8a128a"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.847808 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-host-slash\") pod \"c54d618a-dfec-4f2a-a288-f5188c8a128a\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.848484 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-host-kubelet\") pod \"c54d618a-dfec-4f2a-a288-f5188c8a128a\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.848510 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-host-run-netns\") pod \"c54d618a-dfec-4f2a-a288-f5188c8a128a\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.848526 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-var-lib-openvswitch\") pod \"c54d618a-dfec-4f2a-a288-f5188c8a128a\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.848572 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c54d618a-dfec-4f2a-a288-f5188c8a128a-ovnkube-script-lib\") pod \"c54d618a-dfec-4f2a-a288-f5188c8a128a\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.848583 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "c54d618a-dfec-4f2a-a288-f5188c8a128a" (UID: "c54d618a-dfec-4f2a-a288-f5188c8a128a"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.848598 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-run-openvswitch\") pod \"c54d618a-dfec-4f2a-a288-f5188c8a128a\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.848574 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "c54d618a-dfec-4f2a-a288-f5188c8a128a" (UID: "c54d618a-dfec-4f2a-a288-f5188c8a128a"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.848621 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-systemd-units\") pod \"c54d618a-dfec-4f2a-a288-f5188c8a128a\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.848659 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c54d618a-dfec-4f2a-a288-f5188c8a128a-ovn-node-metrics-cert\") pod \"c54d618a-dfec-4f2a-a288-f5188c8a128a\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.848659 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "c54d618a-dfec-4f2a-a288-f5188c8a128a" (UID: "c54d618a-dfec-4f2a-a288-f5188c8a128a"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.848666 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "c54d618a-dfec-4f2a-a288-f5188c8a128a" (UID: "c54d618a-dfec-4f2a-a288-f5188c8a128a"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.848700 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "c54d618a-dfec-4f2a-a288-f5188c8a128a" (UID: "c54d618a-dfec-4f2a-a288-f5188c8a128a"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.848683 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-host-var-lib-cni-networks-ovn-kubernetes\") pod \"c54d618a-dfec-4f2a-a288-f5188c8a128a\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.848710 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "c54d618a-dfec-4f2a-a288-f5188c8a128a" (UID: "c54d618a-dfec-4f2a-a288-f5188c8a128a"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.848803 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-host-run-ovn-kubernetes\") pod \"c54d618a-dfec-4f2a-a288-f5188c8a128a\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.848847 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c54d618a-dfec-4f2a-a288-f5188c8a128a-ovnkube-config\") pod \"c54d618a-dfec-4f2a-a288-f5188c8a128a\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.848890 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2qphz\" (UniqueName: \"kubernetes.io/projected/c54d618a-dfec-4f2a-a288-f5188c8a128a-kube-api-access-2qphz\") pod \"c54d618a-dfec-4f2a-a288-f5188c8a128a\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.848924 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c54d618a-dfec-4f2a-a288-f5188c8a128a-env-overrides\") pod \"c54d618a-dfec-4f2a-a288-f5188c8a128a\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.848953 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-host-cni-netd\") pod \"c54d618a-dfec-4f2a-a288-f5188c8a128a\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.848981 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-etc-openvswitch\") pod \"c54d618a-dfec-4f2a-a288-f5188c8a128a\" (UID: \"c54d618a-dfec-4f2a-a288-f5188c8a128a\") " Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.849268 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-mwqsx"] Dec 07 16:12:17 crc kubenswrapper[4716]: E1207 16:12:17.849445 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c54d618a-dfec-4f2a-a288-f5188c8a128a" containerName="kubecfg-setup" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.849459 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="c54d618a-dfec-4f2a-a288-f5188c8a128a" containerName="kubecfg-setup" Dec 07 16:12:17 crc kubenswrapper[4716]: E1207 16:12:17.849467 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c54d618a-dfec-4f2a-a288-f5188c8a128a" containerName="nbdb" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.849472 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="c54d618a-dfec-4f2a-a288-f5188c8a128a" containerName="nbdb" Dec 07 16:12:17 crc kubenswrapper[4716]: E1207 16:12:17.849485 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c54d618a-dfec-4f2a-a288-f5188c8a128a" containerName="ovn-controller" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.848886 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "c54d618a-dfec-4f2a-a288-f5188c8a128a" (UID: "c54d618a-dfec-4f2a-a288-f5188c8a128a"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.849491 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="c54d618a-dfec-4f2a-a288-f5188c8a128a" containerName="ovn-controller" Dec 07 16:12:17 crc kubenswrapper[4716]: E1207 16:12:17.849510 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c54d618a-dfec-4f2a-a288-f5188c8a128a" containerName="ovnkube-controller" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.849516 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="c54d618a-dfec-4f2a-a288-f5188c8a128a" containerName="ovnkube-controller" Dec 07 16:12:17 crc kubenswrapper[4716]: E1207 16:12:17.849523 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c54d618a-dfec-4f2a-a288-f5188c8a128a" containerName="kube-rbac-proxy-ovn-metrics" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.849529 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="c54d618a-dfec-4f2a-a288-f5188c8a128a" containerName="kube-rbac-proxy-ovn-metrics" Dec 07 16:12:17 crc kubenswrapper[4716]: E1207 16:12:17.849536 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c54d618a-dfec-4f2a-a288-f5188c8a128a" containerName="ovnkube-controller" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.849542 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="c54d618a-dfec-4f2a-a288-f5188c8a128a" containerName="ovnkube-controller" Dec 07 16:12:17 crc kubenswrapper[4716]: E1207 16:12:17.849549 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c54d618a-dfec-4f2a-a288-f5188c8a128a" containerName="northd" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.849554 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="c54d618a-dfec-4f2a-a288-f5188c8a128a" containerName="northd" Dec 07 16:12:17 crc kubenswrapper[4716]: E1207 16:12:17.849561 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c54d618a-dfec-4f2a-a288-f5188c8a128a" containerName="kube-rbac-proxy-node" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.849566 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="c54d618a-dfec-4f2a-a288-f5188c8a128a" containerName="kube-rbac-proxy-node" Dec 07 16:12:17 crc kubenswrapper[4716]: E1207 16:12:17.849573 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c54d618a-dfec-4f2a-a288-f5188c8a128a" containerName="ovn-acl-logging" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.849580 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="c54d618a-dfec-4f2a-a288-f5188c8a128a" containerName="ovn-acl-logging" Dec 07 16:12:17 crc kubenswrapper[4716]: E1207 16:12:17.849588 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c54d618a-dfec-4f2a-a288-f5188c8a128a" containerName="ovnkube-controller" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.849593 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="c54d618a-dfec-4f2a-a288-f5188c8a128a" containerName="ovnkube-controller" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.849442 4716 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-run-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.849618 4716 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-systemd-units\") on node \"crc\" DevicePath \"\"" Dec 07 16:12:17 crc kubenswrapper[4716]: E1207 16:12:17.849601 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c54d618a-dfec-4f2a-a288-f5188c8a128a" containerName="sbdb" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.849703 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="c54d618a-dfec-4f2a-a288-f5188c8a128a" containerName="sbdb" Dec 07 16:12:17 crc kubenswrapper[4716]: E1207 16:12:17.850007 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c54d618a-dfec-4f2a-a288-f5188c8a128a" containerName="ovnkube-controller" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.850030 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="c54d618a-dfec-4f2a-a288-f5188c8a128a" containerName="ovnkube-controller" Dec 07 16:12:17 crc kubenswrapper[4716]: E1207 16:12:17.850059 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c54d618a-dfec-4f2a-a288-f5188c8a128a" containerName="ovnkube-controller" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.850075 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="c54d618a-dfec-4f2a-a288-f5188c8a128a" containerName="ovnkube-controller" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.848943 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c54d618a-dfec-4f2a-a288-f5188c8a128a-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "c54d618a-dfec-4f2a-a288-f5188c8a128a" (UID: "c54d618a-dfec-4f2a-a288-f5188c8a128a"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.849170 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "c54d618a-dfec-4f2a-a288-f5188c8a128a" (UID: "c54d618a-dfec-4f2a-a288-f5188c8a128a"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.849480 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "c54d618a-dfec-4f2a-a288-f5188c8a128a" (UID: "c54d618a-dfec-4f2a-a288-f5188c8a128a"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.849575 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c54d618a-dfec-4f2a-a288-f5188c8a128a-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "c54d618a-dfec-4f2a-a288-f5188c8a128a" (UID: "c54d618a-dfec-4f2a-a288-f5188c8a128a"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.850139 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c54d618a-dfec-4f2a-a288-f5188c8a128a-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "c54d618a-dfec-4f2a-a288-f5188c8a128a" (UID: "c54d618a-dfec-4f2a-a288-f5188c8a128a"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.850464 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="c54d618a-dfec-4f2a-a288-f5188c8a128a" containerName="ovnkube-controller" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.850491 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="c54d618a-dfec-4f2a-a288-f5188c8a128a" containerName="kube-rbac-proxy-node" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.850514 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="c54d618a-dfec-4f2a-a288-f5188c8a128a" containerName="nbdb" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.850540 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="c54d618a-dfec-4f2a-a288-f5188c8a128a" containerName="northd" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.850558 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="c54d618a-dfec-4f2a-a288-f5188c8a128a" containerName="ovn-acl-logging" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.850581 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="c54d618a-dfec-4f2a-a288-f5188c8a128a" containerName="ovnkube-controller" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.850603 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="c54d618a-dfec-4f2a-a288-f5188c8a128a" containerName="sbdb" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.849638 4716 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.850626 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="c54d618a-dfec-4f2a-a288-f5188c8a128a" containerName="ovnkube-controller" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.850751 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="c54d618a-dfec-4f2a-a288-f5188c8a128a" containerName="ovnkube-controller" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.850777 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="c54d618a-dfec-4f2a-a288-f5188c8a128a" containerName="ovn-controller" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.850798 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="c54d618a-dfec-4f2a-a288-f5188c8a128a" containerName="ovnkube-controller" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.850671 4716 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-node-log\") on node \"crc\" DevicePath \"\"" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.850825 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="c54d618a-dfec-4f2a-a288-f5188c8a128a" containerName="kube-rbac-proxy-ovn-metrics" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.850851 4716 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-host-cni-bin\") on node \"crc\" DevicePath \"\"" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.853496 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c54d618a-dfec-4f2a-a288-f5188c8a128a-kube-api-access-2qphz" (OuterVolumeSpecName: "kube-api-access-2qphz") pod "c54d618a-dfec-4f2a-a288-f5188c8a128a" (UID: "c54d618a-dfec-4f2a-a288-f5188c8a128a"). InnerVolumeSpecName "kube-api-access-2qphz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.860713 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c54d618a-dfec-4f2a-a288-f5188c8a128a-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "c54d618a-dfec-4f2a-a288-f5188c8a128a" (UID: "c54d618a-dfec-4f2a-a288-f5188c8a128a"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.853039 4716 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.869060 4716 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-log-socket\") on node \"crc\" DevicePath \"\"" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.869126 4716 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-host-slash\") on node \"crc\" DevicePath \"\"" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.869149 4716 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-host-kubelet\") on node \"crc\" DevicePath \"\"" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.869179 4716 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-host-run-netns\") on node \"crc\" DevicePath \"\"" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.869200 4716 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.870066 4716 scope.go:117] "RemoveContainer" containerID="bcfdffe822635b0ae13b253e6133449fe00461b2792be43d913c4a77219b8255" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.874544 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.877971 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "c54d618a-dfec-4f2a-a288-f5188c8a128a" (UID: "c54d618a-dfec-4f2a-a288-f5188c8a128a"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.911385 4716 scope.go:117] "RemoveContainer" containerID="a847d04e22e98091c0037a89054b9236228b701b8c7ba31940a56df0f37b95c6" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.930493 4716 scope.go:117] "RemoveContainer" containerID="def9ac840fb8ecf2778bed88ea980eddca7c5b2aba4bd1afb967d62fec3f7bae" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.957965 4716 scope.go:117] "RemoveContainer" containerID="b73be664f5412c98b89d971a2cac37c494151c293fdb44a98b390978658a6643" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.970164 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4e16e2f1-2dd7-441b-9115-1ff638a4de50-etc-openvswitch\") pod \"ovnkube-node-mwqsx\" (UID: \"4e16e2f1-2dd7-441b-9115-1ff638a4de50\") " pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.970199 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bnrvr\" (UniqueName: \"kubernetes.io/projected/4e16e2f1-2dd7-441b-9115-1ff638a4de50-kube-api-access-bnrvr\") pod \"ovnkube-node-mwqsx\" (UID: \"4e16e2f1-2dd7-441b-9115-1ff638a4de50\") " pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.970244 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4e16e2f1-2dd7-441b-9115-1ff638a4de50-host-run-ovn-kubernetes\") pod \"ovnkube-node-mwqsx\" (UID: \"4e16e2f1-2dd7-441b-9115-1ff638a4de50\") " pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.970275 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/4e16e2f1-2dd7-441b-9115-1ff638a4de50-host-kubelet\") pod \"ovnkube-node-mwqsx\" (UID: \"4e16e2f1-2dd7-441b-9115-1ff638a4de50\") " pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.970297 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/4e16e2f1-2dd7-441b-9115-1ff638a4de50-run-systemd\") pod \"ovnkube-node-mwqsx\" (UID: \"4e16e2f1-2dd7-441b-9115-1ff638a4de50\") " pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.970363 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/4e16e2f1-2dd7-441b-9115-1ff638a4de50-node-log\") pod \"ovnkube-node-mwqsx\" (UID: \"4e16e2f1-2dd7-441b-9115-1ff638a4de50\") " pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.970382 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/4e16e2f1-2dd7-441b-9115-1ff638a4de50-host-slash\") pod \"ovnkube-node-mwqsx\" (UID: \"4e16e2f1-2dd7-441b-9115-1ff638a4de50\") " pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.970400 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/4e16e2f1-2dd7-441b-9115-1ff638a4de50-systemd-units\") pod \"ovnkube-node-mwqsx\" (UID: \"4e16e2f1-2dd7-441b-9115-1ff638a4de50\") " pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.970414 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/4e16e2f1-2dd7-441b-9115-1ff638a4de50-ovnkube-script-lib\") pod \"ovnkube-node-mwqsx\" (UID: \"4e16e2f1-2dd7-441b-9115-1ff638a4de50\") " pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.970434 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4e16e2f1-2dd7-441b-9115-1ff638a4de50-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-mwqsx\" (UID: \"4e16e2f1-2dd7-441b-9115-1ff638a4de50\") " pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.970538 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/4e16e2f1-2dd7-441b-9115-1ff638a4de50-host-cni-netd\") pod \"ovnkube-node-mwqsx\" (UID: \"4e16e2f1-2dd7-441b-9115-1ff638a4de50\") " pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.970626 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4e16e2f1-2dd7-441b-9115-1ff638a4de50-host-cni-bin\") pod \"ovnkube-node-mwqsx\" (UID: \"4e16e2f1-2dd7-441b-9115-1ff638a4de50\") " pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.970677 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/4e16e2f1-2dd7-441b-9115-1ff638a4de50-log-socket\") pod \"ovnkube-node-mwqsx\" (UID: \"4e16e2f1-2dd7-441b-9115-1ff638a4de50\") " pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.970744 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4e16e2f1-2dd7-441b-9115-1ff638a4de50-env-overrides\") pod \"ovnkube-node-mwqsx\" (UID: \"4e16e2f1-2dd7-441b-9115-1ff638a4de50\") " pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.970774 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4e16e2f1-2dd7-441b-9115-1ff638a4de50-ovnkube-config\") pod \"ovnkube-node-mwqsx\" (UID: \"4e16e2f1-2dd7-441b-9115-1ff638a4de50\") " pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.970801 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4e16e2f1-2dd7-441b-9115-1ff638a4de50-var-lib-openvswitch\") pod \"ovnkube-node-mwqsx\" (UID: \"4e16e2f1-2dd7-441b-9115-1ff638a4de50\") " pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.970821 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4e16e2f1-2dd7-441b-9115-1ff638a4de50-host-run-netns\") pod \"ovnkube-node-mwqsx\" (UID: \"4e16e2f1-2dd7-441b-9115-1ff638a4de50\") " pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.970837 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4e16e2f1-2dd7-441b-9115-1ff638a4de50-run-openvswitch\") pod \"ovnkube-node-mwqsx\" (UID: \"4e16e2f1-2dd7-441b-9115-1ff638a4de50\") " pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.970866 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/4e16e2f1-2dd7-441b-9115-1ff638a4de50-run-ovn\") pod \"ovnkube-node-mwqsx\" (UID: \"4e16e2f1-2dd7-441b-9115-1ff638a4de50\") " pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.970883 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4e16e2f1-2dd7-441b-9115-1ff638a4de50-ovn-node-metrics-cert\") pod \"ovnkube-node-mwqsx\" (UID: \"4e16e2f1-2dd7-441b-9115-1ff638a4de50\") " pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.970953 4716 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c54d618a-dfec-4f2a-a288-f5188c8a128a-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.970964 4716 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c54d618a-dfec-4f2a-a288-f5188c8a128a-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.970978 4716 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.970988 4716 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c54d618a-dfec-4f2a-a288-f5188c8a128a-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.970997 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2qphz\" (UniqueName: \"kubernetes.io/projected/c54d618a-dfec-4f2a-a288-f5188c8a128a-kube-api-access-2qphz\") on node \"crc\" DevicePath \"\"" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.971034 4716 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c54d618a-dfec-4f2a-a288-f5188c8a128a-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.971043 4716 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-host-cni-netd\") on node \"crc\" DevicePath \"\"" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.971051 4716 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.971062 4716 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c54d618a-dfec-4f2a-a288-f5188c8a128a-run-systemd\") on node \"crc\" DevicePath \"\"" Dec 07 16:12:17 crc kubenswrapper[4716]: I1207 16:12:17.977161 4716 scope.go:117] "RemoveContainer" containerID="dcae78a6133d97f6a32dd7421ccec4333b970a237404e056a5cc601f8eefdc8d" Dec 07 16:12:18 crc kubenswrapper[4716]: I1207 16:12:18.000697 4716 scope.go:117] "RemoveContainer" containerID="9bfe2ffce7a7b7c320474117734fc05246cc2b4fd744833e896e8e7c83f5e4d8" Dec 07 16:12:18 crc kubenswrapper[4716]: I1207 16:12:18.021539 4716 scope.go:117] "RemoveContainer" containerID="c60794c12cb6e7647d4c1d0bee76e214a71d9cb818d9c98dbf9880fb030ed54b" Dec 07 16:12:18 crc kubenswrapper[4716]: I1207 16:12:18.043326 4716 scope.go:117] "RemoveContainer" containerID="e462e925550f957641f83d75b2990297a4cfd973b9227fb08dc610a75b1d6e79" Dec 07 16:12:18 crc kubenswrapper[4716]: I1207 16:12:18.061669 4716 scope.go:117] "RemoveContainer" containerID="2ed4f4d3564674bf5cb4e7a76dc0fd95e203ecc8950798e59a1fb6c9ab4a20f7" Dec 07 16:12:18 crc kubenswrapper[4716]: I1207 16:12:18.070385 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-ft9cq" Dec 07 16:12:18 crc kubenswrapper[4716]: I1207 16:12:18.071625 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4e16e2f1-2dd7-441b-9115-1ff638a4de50-env-overrides\") pod \"ovnkube-node-mwqsx\" (UID: \"4e16e2f1-2dd7-441b-9115-1ff638a4de50\") " pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" Dec 07 16:12:18 crc kubenswrapper[4716]: I1207 16:12:18.071659 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4e16e2f1-2dd7-441b-9115-1ff638a4de50-ovnkube-config\") pod \"ovnkube-node-mwqsx\" (UID: \"4e16e2f1-2dd7-441b-9115-1ff638a4de50\") " pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" Dec 07 16:12:18 crc kubenswrapper[4716]: I1207 16:12:18.071684 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4e16e2f1-2dd7-441b-9115-1ff638a4de50-var-lib-openvswitch\") pod \"ovnkube-node-mwqsx\" (UID: \"4e16e2f1-2dd7-441b-9115-1ff638a4de50\") " pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" Dec 07 16:12:18 crc kubenswrapper[4716]: I1207 16:12:18.071700 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4e16e2f1-2dd7-441b-9115-1ff638a4de50-host-run-netns\") pod \"ovnkube-node-mwqsx\" (UID: \"4e16e2f1-2dd7-441b-9115-1ff638a4de50\") " pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" Dec 07 16:12:18 crc kubenswrapper[4716]: I1207 16:12:18.071716 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4e16e2f1-2dd7-441b-9115-1ff638a4de50-run-openvswitch\") pod \"ovnkube-node-mwqsx\" (UID: \"4e16e2f1-2dd7-441b-9115-1ff638a4de50\") " pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" Dec 07 16:12:18 crc kubenswrapper[4716]: I1207 16:12:18.071735 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/4e16e2f1-2dd7-441b-9115-1ff638a4de50-run-ovn\") pod \"ovnkube-node-mwqsx\" (UID: \"4e16e2f1-2dd7-441b-9115-1ff638a4de50\") " pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" Dec 07 16:12:18 crc kubenswrapper[4716]: I1207 16:12:18.071752 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4e16e2f1-2dd7-441b-9115-1ff638a4de50-ovn-node-metrics-cert\") pod \"ovnkube-node-mwqsx\" (UID: \"4e16e2f1-2dd7-441b-9115-1ff638a4de50\") " pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" Dec 07 16:12:18 crc kubenswrapper[4716]: I1207 16:12:18.071771 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4e16e2f1-2dd7-441b-9115-1ff638a4de50-etc-openvswitch\") pod \"ovnkube-node-mwqsx\" (UID: \"4e16e2f1-2dd7-441b-9115-1ff638a4de50\") " pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" Dec 07 16:12:18 crc kubenswrapper[4716]: I1207 16:12:18.071787 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bnrvr\" (UniqueName: \"kubernetes.io/projected/4e16e2f1-2dd7-441b-9115-1ff638a4de50-kube-api-access-bnrvr\") pod \"ovnkube-node-mwqsx\" (UID: \"4e16e2f1-2dd7-441b-9115-1ff638a4de50\") " pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" Dec 07 16:12:18 crc kubenswrapper[4716]: I1207 16:12:18.071804 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4e16e2f1-2dd7-441b-9115-1ff638a4de50-host-run-ovn-kubernetes\") pod \"ovnkube-node-mwqsx\" (UID: \"4e16e2f1-2dd7-441b-9115-1ff638a4de50\") " pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" Dec 07 16:12:18 crc kubenswrapper[4716]: I1207 16:12:18.071823 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/4e16e2f1-2dd7-441b-9115-1ff638a4de50-host-kubelet\") pod \"ovnkube-node-mwqsx\" (UID: \"4e16e2f1-2dd7-441b-9115-1ff638a4de50\") " pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" Dec 07 16:12:18 crc kubenswrapper[4716]: I1207 16:12:18.071830 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4e16e2f1-2dd7-441b-9115-1ff638a4de50-var-lib-openvswitch\") pod \"ovnkube-node-mwqsx\" (UID: \"4e16e2f1-2dd7-441b-9115-1ff638a4de50\") " pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" Dec 07 16:12:18 crc kubenswrapper[4716]: I1207 16:12:18.071906 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4e16e2f1-2dd7-441b-9115-1ff638a4de50-host-run-netns\") pod \"ovnkube-node-mwqsx\" (UID: \"4e16e2f1-2dd7-441b-9115-1ff638a4de50\") " pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" Dec 07 16:12:18 crc kubenswrapper[4716]: I1207 16:12:18.071863 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/4e16e2f1-2dd7-441b-9115-1ff638a4de50-run-systemd\") pod \"ovnkube-node-mwqsx\" (UID: \"4e16e2f1-2dd7-441b-9115-1ff638a4de50\") " pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" Dec 07 16:12:18 crc kubenswrapper[4716]: I1207 16:12:18.071889 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/4e16e2f1-2dd7-441b-9115-1ff638a4de50-run-systemd\") pod \"ovnkube-node-mwqsx\" (UID: \"4e16e2f1-2dd7-441b-9115-1ff638a4de50\") " pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" Dec 07 16:12:18 crc kubenswrapper[4716]: I1207 16:12:18.071970 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4e16e2f1-2dd7-441b-9115-1ff638a4de50-run-openvswitch\") pod \"ovnkube-node-mwqsx\" (UID: \"4e16e2f1-2dd7-441b-9115-1ff638a4de50\") " pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" Dec 07 16:12:18 crc kubenswrapper[4716]: I1207 16:12:18.071993 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/4e16e2f1-2dd7-441b-9115-1ff638a4de50-run-ovn\") pod \"ovnkube-node-mwqsx\" (UID: \"4e16e2f1-2dd7-441b-9115-1ff638a4de50\") " pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" Dec 07 16:12:18 crc kubenswrapper[4716]: I1207 16:12:18.071994 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/4e16e2f1-2dd7-441b-9115-1ff638a4de50-node-log\") pod \"ovnkube-node-mwqsx\" (UID: \"4e16e2f1-2dd7-441b-9115-1ff638a4de50\") " pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" Dec 07 16:12:18 crc kubenswrapper[4716]: I1207 16:12:18.072032 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/4e16e2f1-2dd7-441b-9115-1ff638a4de50-host-slash\") pod \"ovnkube-node-mwqsx\" (UID: \"4e16e2f1-2dd7-441b-9115-1ff638a4de50\") " pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" Dec 07 16:12:18 crc kubenswrapper[4716]: I1207 16:12:18.072102 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/4e16e2f1-2dd7-441b-9115-1ff638a4de50-node-log\") pod \"ovnkube-node-mwqsx\" (UID: \"4e16e2f1-2dd7-441b-9115-1ff638a4de50\") " pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" Dec 07 16:12:18 crc kubenswrapper[4716]: I1207 16:12:18.072107 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/4e16e2f1-2dd7-441b-9115-1ff638a4de50-host-slash\") pod \"ovnkube-node-mwqsx\" (UID: \"4e16e2f1-2dd7-441b-9115-1ff638a4de50\") " pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" Dec 07 16:12:18 crc kubenswrapper[4716]: I1207 16:12:18.072218 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4e16e2f1-2dd7-441b-9115-1ff638a4de50-env-overrides\") pod \"ovnkube-node-mwqsx\" (UID: \"4e16e2f1-2dd7-441b-9115-1ff638a4de50\") " pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" Dec 07 16:12:18 crc kubenswrapper[4716]: I1207 16:12:18.072260 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4e16e2f1-2dd7-441b-9115-1ff638a4de50-etc-openvswitch\") pod \"ovnkube-node-mwqsx\" (UID: \"4e16e2f1-2dd7-441b-9115-1ff638a4de50\") " pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" Dec 07 16:12:18 crc kubenswrapper[4716]: I1207 16:12:18.072285 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4e16e2f1-2dd7-441b-9115-1ff638a4de50-host-run-ovn-kubernetes\") pod \"ovnkube-node-mwqsx\" (UID: \"4e16e2f1-2dd7-441b-9115-1ff638a4de50\") " pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" Dec 07 16:12:18 crc kubenswrapper[4716]: I1207 16:12:18.072308 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/4e16e2f1-2dd7-441b-9115-1ff638a4de50-host-kubelet\") pod \"ovnkube-node-mwqsx\" (UID: \"4e16e2f1-2dd7-441b-9115-1ff638a4de50\") " pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" Dec 07 16:12:18 crc kubenswrapper[4716]: I1207 16:12:18.072405 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4e16e2f1-2dd7-441b-9115-1ff638a4de50-ovnkube-config\") pod \"ovnkube-node-mwqsx\" (UID: \"4e16e2f1-2dd7-441b-9115-1ff638a4de50\") " pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" Dec 07 16:12:18 crc kubenswrapper[4716]: I1207 16:12:18.072412 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/4e16e2f1-2dd7-441b-9115-1ff638a4de50-systemd-units\") pod \"ovnkube-node-mwqsx\" (UID: \"4e16e2f1-2dd7-441b-9115-1ff638a4de50\") " pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" Dec 07 16:12:18 crc kubenswrapper[4716]: I1207 16:12:18.072440 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/4e16e2f1-2dd7-441b-9115-1ff638a4de50-systemd-units\") pod \"ovnkube-node-mwqsx\" (UID: \"4e16e2f1-2dd7-441b-9115-1ff638a4de50\") " pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" Dec 07 16:12:18 crc kubenswrapper[4716]: I1207 16:12:18.072519 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/4e16e2f1-2dd7-441b-9115-1ff638a4de50-ovnkube-script-lib\") pod \"ovnkube-node-mwqsx\" (UID: \"4e16e2f1-2dd7-441b-9115-1ff638a4de50\") " pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" Dec 07 16:12:18 crc kubenswrapper[4716]: I1207 16:12:18.072589 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4e16e2f1-2dd7-441b-9115-1ff638a4de50-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-mwqsx\" (UID: \"4e16e2f1-2dd7-441b-9115-1ff638a4de50\") " pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" Dec 07 16:12:18 crc kubenswrapper[4716]: I1207 16:12:18.072614 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/4e16e2f1-2dd7-441b-9115-1ff638a4de50-host-cni-netd\") pod \"ovnkube-node-mwqsx\" (UID: \"4e16e2f1-2dd7-441b-9115-1ff638a4de50\") " pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" Dec 07 16:12:18 crc kubenswrapper[4716]: I1207 16:12:18.072641 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4e16e2f1-2dd7-441b-9115-1ff638a4de50-host-cni-bin\") pod \"ovnkube-node-mwqsx\" (UID: \"4e16e2f1-2dd7-441b-9115-1ff638a4de50\") " pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" Dec 07 16:12:18 crc kubenswrapper[4716]: I1207 16:12:18.072662 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/4e16e2f1-2dd7-441b-9115-1ff638a4de50-log-socket\") pod \"ovnkube-node-mwqsx\" (UID: \"4e16e2f1-2dd7-441b-9115-1ff638a4de50\") " pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" Dec 07 16:12:18 crc kubenswrapper[4716]: I1207 16:12:18.072737 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/4e16e2f1-2dd7-441b-9115-1ff638a4de50-host-cni-netd\") pod \"ovnkube-node-mwqsx\" (UID: \"4e16e2f1-2dd7-441b-9115-1ff638a4de50\") " pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" Dec 07 16:12:18 crc kubenswrapper[4716]: I1207 16:12:18.072744 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/4e16e2f1-2dd7-441b-9115-1ff638a4de50-log-socket\") pod \"ovnkube-node-mwqsx\" (UID: \"4e16e2f1-2dd7-441b-9115-1ff638a4de50\") " pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" Dec 07 16:12:18 crc kubenswrapper[4716]: I1207 16:12:18.072763 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4e16e2f1-2dd7-441b-9115-1ff638a4de50-host-cni-bin\") pod \"ovnkube-node-mwqsx\" (UID: \"4e16e2f1-2dd7-441b-9115-1ff638a4de50\") " pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" Dec 07 16:12:18 crc kubenswrapper[4716]: I1207 16:12:18.072806 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4e16e2f1-2dd7-441b-9115-1ff638a4de50-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-mwqsx\" (UID: \"4e16e2f1-2dd7-441b-9115-1ff638a4de50\") " pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" Dec 07 16:12:18 crc kubenswrapper[4716]: I1207 16:12:18.072948 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/4e16e2f1-2dd7-441b-9115-1ff638a4de50-ovnkube-script-lib\") pod \"ovnkube-node-mwqsx\" (UID: \"4e16e2f1-2dd7-441b-9115-1ff638a4de50\") " pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" Dec 07 16:12:18 crc kubenswrapper[4716]: I1207 16:12:18.075583 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4e16e2f1-2dd7-441b-9115-1ff638a4de50-ovn-node-metrics-cert\") pod \"ovnkube-node-mwqsx\" (UID: \"4e16e2f1-2dd7-441b-9115-1ff638a4de50\") " pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" Dec 07 16:12:18 crc kubenswrapper[4716]: I1207 16:12:18.081429 4716 scope.go:117] "RemoveContainer" containerID="6fca3b2ea503451cfaf7afb00445fe0adb0acc87d436d1eebba5068452b8e7cb" Dec 07 16:12:18 crc kubenswrapper[4716]: I1207 16:12:18.094629 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bnrvr\" (UniqueName: \"kubernetes.io/projected/4e16e2f1-2dd7-441b-9115-1ff638a4de50-kube-api-access-bnrvr\") pod \"ovnkube-node-mwqsx\" (UID: \"4e16e2f1-2dd7-441b-9115-1ff638a4de50\") " pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" Dec 07 16:12:18 crc kubenswrapper[4716]: I1207 16:12:18.214266 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" Dec 07 16:12:18 crc kubenswrapper[4716]: W1207 16:12:18.241171 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4e16e2f1_2dd7_441b_9115_1ff638a4de50.slice/crio-189b487169658d46da1ac58bb3d92a1ac5de862fd3ea662f12647b9fa07c11a5 WatchSource:0}: Error finding container 189b487169658d46da1ac58bb3d92a1ac5de862fd3ea662f12647b9fa07c11a5: Status 404 returned error can't find the container with id 189b487169658d46da1ac58bb3d92a1ac5de862fd3ea662f12647b9fa07c11a5 Dec 07 16:12:18 crc kubenswrapper[4716]: I1207 16:12:18.541601 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" event={"ID":"c54d618a-dfec-4f2a-a288-f5188c8a128a","Type":"ContainerDied","Data":"e462e925550f957641f83d75b2990297a4cfd973b9227fb08dc610a75b1d6e79"} Dec 07 16:12:18 crc kubenswrapper[4716]: I1207 16:12:18.542232 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" event={"ID":"c54d618a-dfec-4f2a-a288-f5188c8a128a","Type":"ContainerDied","Data":"bcfdffe822635b0ae13b253e6133449fe00461b2792be43d913c4a77219b8255"} Dec 07 16:12:18 crc kubenswrapper[4716]: I1207 16:12:18.542278 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" event={"ID":"c54d618a-dfec-4f2a-a288-f5188c8a128a","Type":"ContainerDied","Data":"def9ac840fb8ecf2778bed88ea980eddca7c5b2aba4bd1afb967d62fec3f7bae"} Dec 07 16:12:18 crc kubenswrapper[4716]: I1207 16:12:18.542319 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" event={"ID":"c54d618a-dfec-4f2a-a288-f5188c8a128a","Type":"ContainerDied","Data":"b73be664f5412c98b89d971a2cac37c494151c293fdb44a98b390978658a6643"} Dec 07 16:12:18 crc kubenswrapper[4716]: I1207 16:12:18.542345 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" event={"ID":"c54d618a-dfec-4f2a-a288-f5188c8a128a","Type":"ContainerDied","Data":"2ed4f4d3564674bf5cb4e7a76dc0fd95e203ecc8950798e59a1fb6c9ab4a20f7"} Dec 07 16:12:18 crc kubenswrapper[4716]: I1207 16:12:18.542372 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" event={"ID":"c54d618a-dfec-4f2a-a288-f5188c8a128a","Type":"ContainerDied","Data":"60324dc7d701bc064e14ae4f474bce10f6ee4ff3c3c942eab749e7166082cf9d"} Dec 07 16:12:18 crc kubenswrapper[4716]: I1207 16:12:18.542399 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" event={"ID":"c54d618a-dfec-4f2a-a288-f5188c8a128a","Type":"ContainerDied","Data":"a847d04e22e98091c0037a89054b9236228b701b8c7ba31940a56df0f37b95c6"} Dec 07 16:12:18 crc kubenswrapper[4716]: I1207 16:12:18.542423 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" event={"ID":"c54d618a-dfec-4f2a-a288-f5188c8a128a","Type":"ContainerDied","Data":"dcae78a6133d97f6a32dd7421ccec4333b970a237404e056a5cc601f8eefdc8d"} Dec 07 16:12:18 crc kubenswrapper[4716]: I1207 16:12:18.542452 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" event={"ID":"c54d618a-dfec-4f2a-a288-f5188c8a128a","Type":"ContainerDied","Data":"ca8d06419e72db2db45ffd4d586e3eac926332b14273bb82e37118f39d5ebd3f"} Dec 07 16:12:18 crc kubenswrapper[4716]: I1207 16:12:18.543867 4716 generic.go:334] "Generic (PLEG): container finished" podID="4e16e2f1-2dd7-441b-9115-1ff638a4de50" containerID="cb9966ccf1a028f751cc0b5f4fd3157dd47a79f96db7c7f973204eb7cc17afec" exitCode=0 Dec 07 16:12:18 crc kubenswrapper[4716]: I1207 16:12:18.543940 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" event={"ID":"4e16e2f1-2dd7-441b-9115-1ff638a4de50","Type":"ContainerDied","Data":"cb9966ccf1a028f751cc0b5f4fd3157dd47a79f96db7c7f973204eb7cc17afec"} Dec 07 16:12:18 crc kubenswrapper[4716]: I1207 16:12:18.543993 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" event={"ID":"4e16e2f1-2dd7-441b-9115-1ff638a4de50","Type":"ContainerStarted","Data":"189b487169658d46da1ac58bb3d92a1ac5de862fd3ea662f12647b9fa07c11a5"} Dec 07 16:12:18 crc kubenswrapper[4716]: I1207 16:12:18.550567 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-k4wr2_0baf8575-26cf-4a72-8a6d-ae94c078eaad/kube-multus/2.log" Dec 07 16:12:18 crc kubenswrapper[4716]: I1207 16:12:18.550650 4716 generic.go:334] "Generic (PLEG): container finished" podID="0baf8575-26cf-4a72-8a6d-ae94c078eaad" containerID="a7122711012345eed920500d2894cb3e2cb321f8b1c45845be71f106475783c8" exitCode=2 Dec 07 16:12:18 crc kubenswrapper[4716]: I1207 16:12:18.550721 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-k4wr2" event={"ID":"0baf8575-26cf-4a72-8a6d-ae94c078eaad","Type":"ContainerDied","Data":"a7122711012345eed920500d2894cb3e2cb321f8b1c45845be71f106475783c8"} Dec 07 16:12:18 crc kubenswrapper[4716]: I1207 16:12:18.550796 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-qvc7p" Dec 07 16:12:18 crc kubenswrapper[4716]: I1207 16:12:18.551540 4716 scope.go:117] "RemoveContainer" containerID="a7122711012345eed920500d2894cb3e2cb321f8b1c45845be71f106475783c8" Dec 07 16:12:18 crc kubenswrapper[4716]: E1207 16:12:18.551740 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-k4wr2_openshift-multus(0baf8575-26cf-4a72-8a6d-ae94c078eaad)\"" pod="openshift-multus/multus-k4wr2" podUID="0baf8575-26cf-4a72-8a6d-ae94c078eaad" Dec 07 16:12:18 crc kubenswrapper[4716]: I1207 16:12:18.610312 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-qvc7p"] Dec 07 16:12:18 crc kubenswrapper[4716]: I1207 16:12:18.616125 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-qvc7p"] Dec 07 16:12:19 crc kubenswrapper[4716]: I1207 16:12:19.562286 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" event={"ID":"4e16e2f1-2dd7-441b-9115-1ff638a4de50","Type":"ContainerStarted","Data":"4b457397e93015cb26747c951c45a0c38a749ed5b3ad3f2b2d0575e1df421fc4"} Dec 07 16:12:19 crc kubenswrapper[4716]: I1207 16:12:19.562699 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" event={"ID":"4e16e2f1-2dd7-441b-9115-1ff638a4de50","Type":"ContainerStarted","Data":"934f9aae7d36e396fa1e187d50055dc2a63ae613cef020a8af70d17e5fbe0454"} Dec 07 16:12:19 crc kubenswrapper[4716]: I1207 16:12:19.562710 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" event={"ID":"4e16e2f1-2dd7-441b-9115-1ff638a4de50","Type":"ContainerStarted","Data":"da61cb2b7b812ee710d49822e99b30e7f60504e319cd484cbb6bb19ee2e65bff"} Dec 07 16:12:19 crc kubenswrapper[4716]: I1207 16:12:19.562723 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" event={"ID":"4e16e2f1-2dd7-441b-9115-1ff638a4de50","Type":"ContainerStarted","Data":"4567f4449d791e4fb767bcaa7336d9be7a09d823d758022437aee992ee0dfef2"} Dec 07 16:12:19 crc kubenswrapper[4716]: I1207 16:12:19.562733 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" event={"ID":"4e16e2f1-2dd7-441b-9115-1ff638a4de50","Type":"ContainerStarted","Data":"5d094942cd1e1f2275ae58f36b082053ea50bb152c6b7002f29ccd13a9978286"} Dec 07 16:12:19 crc kubenswrapper[4716]: I1207 16:12:19.562741 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" event={"ID":"4e16e2f1-2dd7-441b-9115-1ff638a4de50","Type":"ContainerStarted","Data":"122ba5bf0cac38bfa672e76df83e59afc52eadc16dcdb6ebb69a9027fd2030c3"} Dec 07 16:12:19 crc kubenswrapper[4716]: I1207 16:12:19.669861 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c54d618a-dfec-4f2a-a288-f5188c8a128a" path="/var/lib/kubelet/pods/c54d618a-dfec-4f2a-a288-f5188c8a128a/volumes" Dec 07 16:12:22 crc kubenswrapper[4716]: I1207 16:12:22.589030 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" event={"ID":"4e16e2f1-2dd7-441b-9115-1ff638a4de50","Type":"ContainerStarted","Data":"2e92aa4085b13153b120aced7b50f576e688cabb833bc3f3fd7738bec432f93a"} Dec 07 16:12:22 crc kubenswrapper[4716]: I1207 16:12:22.762226 4716 patch_prober.go:28] interesting pod/machine-config-daemon-zcxxp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 16:12:22 crc kubenswrapper[4716]: I1207 16:12:22.762329 4716 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 16:12:24 crc kubenswrapper[4716]: I1207 16:12:24.603053 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" event={"ID":"4e16e2f1-2dd7-441b-9115-1ff638a4de50","Type":"ContainerStarted","Data":"fb68465e175098bb814e7ad95a1b4e77f3c6d98465772a4783661420286cb6e4"} Dec 07 16:12:24 crc kubenswrapper[4716]: I1207 16:12:24.603622 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" Dec 07 16:12:24 crc kubenswrapper[4716]: I1207 16:12:24.603635 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" Dec 07 16:12:24 crc kubenswrapper[4716]: I1207 16:12:24.630477 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" podStartSLOduration=7.630426328 podStartE2EDuration="7.630426328s" podCreationTimestamp="2025-12-07 16:12:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:12:24.627789937 +0000 UTC m=+607.318074849" watchObservedRunningTime="2025-12-07 16:12:24.630426328 +0000 UTC m=+607.320711240" Dec 07 16:12:24 crc kubenswrapper[4716]: I1207 16:12:24.638806 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" Dec 07 16:12:25 crc kubenswrapper[4716]: I1207 16:12:25.608936 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" Dec 07 16:12:25 crc kubenswrapper[4716]: I1207 16:12:25.639744 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" Dec 07 16:12:30 crc kubenswrapper[4716]: I1207 16:12:30.658638 4716 scope.go:117] "RemoveContainer" containerID="a7122711012345eed920500d2894cb3e2cb321f8b1c45845be71f106475783c8" Dec 07 16:12:30 crc kubenswrapper[4716]: E1207 16:12:30.661303 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-k4wr2_openshift-multus(0baf8575-26cf-4a72-8a6d-ae94c078eaad)\"" pod="openshift-multus/multus-k4wr2" podUID="0baf8575-26cf-4a72-8a6d-ae94c078eaad" Dec 07 16:12:45 crc kubenswrapper[4716]: I1207 16:12:45.657835 4716 scope.go:117] "RemoveContainer" containerID="a7122711012345eed920500d2894cb3e2cb321f8b1c45845be71f106475783c8" Dec 07 16:12:46 crc kubenswrapper[4716]: I1207 16:12:46.763861 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-k4wr2_0baf8575-26cf-4a72-8a6d-ae94c078eaad/kube-multus/2.log" Dec 07 16:12:46 crc kubenswrapper[4716]: I1207 16:12:46.764888 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-k4wr2" event={"ID":"0baf8575-26cf-4a72-8a6d-ae94c078eaad","Type":"ContainerStarted","Data":"03bfcba0fd73a26589b50eb329c191b61534f7afc1e694d0e166d78ad998127a"} Dec 07 16:12:48 crc kubenswrapper[4716]: I1207 16:12:48.235764 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-mwqsx" Dec 07 16:12:52 crc kubenswrapper[4716]: I1207 16:12:52.761677 4716 patch_prober.go:28] interesting pod/machine-config-daemon-zcxxp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 16:12:52 crc kubenswrapper[4716]: I1207 16:12:52.762328 4716 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 16:12:52 crc kubenswrapper[4716]: I1207 16:12:52.762380 4716 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" Dec 07 16:12:52 crc kubenswrapper[4716]: I1207 16:12:52.762988 4716 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"def7373f410cd12d2a355b52c07eee66525502687b1341f3cbd5919e77d0a06e"} pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 07 16:12:52 crc kubenswrapper[4716]: I1207 16:12:52.763100 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" containerName="machine-config-daemon" containerID="cri-o://def7373f410cd12d2a355b52c07eee66525502687b1341f3cbd5919e77d0a06e" gracePeriod=600 Dec 07 16:12:53 crc kubenswrapper[4716]: I1207 16:12:53.814856 4716 generic.go:334] "Generic (PLEG): container finished" podID="c15b59eb-565d-4556-a4ce-75afdf159dc8" containerID="def7373f410cd12d2a355b52c07eee66525502687b1341f3cbd5919e77d0a06e" exitCode=0 Dec 07 16:12:53 crc kubenswrapper[4716]: I1207 16:12:53.814989 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" event={"ID":"c15b59eb-565d-4556-a4ce-75afdf159dc8","Type":"ContainerDied","Data":"def7373f410cd12d2a355b52c07eee66525502687b1341f3cbd5919e77d0a06e"} Dec 07 16:12:53 crc kubenswrapper[4716]: I1207 16:12:53.815192 4716 scope.go:117] "RemoveContainer" containerID="ca669356fa57e43a7c10e3a5b76ace008fedad9bdfdd4c2c358a8731bdaaf42b" Dec 07 16:12:54 crc kubenswrapper[4716]: I1207 16:12:54.822547 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" event={"ID":"c15b59eb-565d-4556-a4ce-75afdf159dc8","Type":"ContainerStarted","Data":"8ec0c9fb49e1eaea181d9e297f07686725c05f58dac3a79c7a2ee3f00f979908"} Dec 07 16:12:57 crc kubenswrapper[4716]: I1207 16:12:57.486684 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fh9qqw"] Dec 07 16:12:57 crc kubenswrapper[4716]: I1207 16:12:57.488695 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fh9qqw" Dec 07 16:12:57 crc kubenswrapper[4716]: I1207 16:12:57.491145 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 07 16:12:57 crc kubenswrapper[4716]: I1207 16:12:57.499592 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fh9qqw"] Dec 07 16:12:57 crc kubenswrapper[4716]: I1207 16:12:57.511488 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7ecea701-6ded-4f27-844f-fd9dfcb92f29-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fh9qqw\" (UID: \"7ecea701-6ded-4f27-844f-fd9dfcb92f29\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fh9qqw" Dec 07 16:12:57 crc kubenswrapper[4716]: I1207 16:12:57.511540 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7ecea701-6ded-4f27-844f-fd9dfcb92f29-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fh9qqw\" (UID: \"7ecea701-6ded-4f27-844f-fd9dfcb92f29\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fh9qqw" Dec 07 16:12:57 crc kubenswrapper[4716]: I1207 16:12:57.511571 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9c5r\" (UniqueName: \"kubernetes.io/projected/7ecea701-6ded-4f27-844f-fd9dfcb92f29-kube-api-access-h9c5r\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fh9qqw\" (UID: \"7ecea701-6ded-4f27-844f-fd9dfcb92f29\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fh9qqw" Dec 07 16:12:57 crc kubenswrapper[4716]: I1207 16:12:57.613149 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7ecea701-6ded-4f27-844f-fd9dfcb92f29-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fh9qqw\" (UID: \"7ecea701-6ded-4f27-844f-fd9dfcb92f29\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fh9qqw" Dec 07 16:12:57 crc kubenswrapper[4716]: I1207 16:12:57.613196 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7ecea701-6ded-4f27-844f-fd9dfcb92f29-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fh9qqw\" (UID: \"7ecea701-6ded-4f27-844f-fd9dfcb92f29\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fh9qqw" Dec 07 16:12:57 crc kubenswrapper[4716]: I1207 16:12:57.613228 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9c5r\" (UniqueName: \"kubernetes.io/projected/7ecea701-6ded-4f27-844f-fd9dfcb92f29-kube-api-access-h9c5r\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fh9qqw\" (UID: \"7ecea701-6ded-4f27-844f-fd9dfcb92f29\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fh9qqw" Dec 07 16:12:57 crc kubenswrapper[4716]: I1207 16:12:57.613825 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7ecea701-6ded-4f27-844f-fd9dfcb92f29-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fh9qqw\" (UID: \"7ecea701-6ded-4f27-844f-fd9dfcb92f29\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fh9qqw" Dec 07 16:12:57 crc kubenswrapper[4716]: I1207 16:12:57.614049 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7ecea701-6ded-4f27-844f-fd9dfcb92f29-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fh9qqw\" (UID: \"7ecea701-6ded-4f27-844f-fd9dfcb92f29\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fh9qqw" Dec 07 16:12:57 crc kubenswrapper[4716]: I1207 16:12:57.635064 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9c5r\" (UniqueName: \"kubernetes.io/projected/7ecea701-6ded-4f27-844f-fd9dfcb92f29-kube-api-access-h9c5r\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fh9qqw\" (UID: \"7ecea701-6ded-4f27-844f-fd9dfcb92f29\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fh9qqw" Dec 07 16:12:57 crc kubenswrapper[4716]: I1207 16:12:57.806144 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fh9qqw" Dec 07 16:12:58 crc kubenswrapper[4716]: I1207 16:12:58.057336 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fh9qqw"] Dec 07 16:12:58 crc kubenswrapper[4716]: I1207 16:12:58.849976 4716 generic.go:334] "Generic (PLEG): container finished" podID="7ecea701-6ded-4f27-844f-fd9dfcb92f29" containerID="557fbd71179ecc1e44303d07962b9ab8ac9b905d2cdf059dcd7bd440da588104" exitCode=0 Dec 07 16:12:58 crc kubenswrapper[4716]: I1207 16:12:58.850045 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fh9qqw" event={"ID":"7ecea701-6ded-4f27-844f-fd9dfcb92f29","Type":"ContainerDied","Data":"557fbd71179ecc1e44303d07962b9ab8ac9b905d2cdf059dcd7bd440da588104"} Dec 07 16:12:58 crc kubenswrapper[4716]: I1207 16:12:58.850357 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fh9qqw" event={"ID":"7ecea701-6ded-4f27-844f-fd9dfcb92f29","Type":"ContainerStarted","Data":"f43a1ecd52058bbb4bd5e7b29e5bc378456f3a22b7b9098a70f92da0e7c9ad83"} Dec 07 16:13:00 crc kubenswrapper[4716]: I1207 16:13:00.866546 4716 generic.go:334] "Generic (PLEG): container finished" podID="7ecea701-6ded-4f27-844f-fd9dfcb92f29" containerID="2ad8eb95a1bb4b97051f0a32f7f621a9954bd07166035e29f1d739cb59ad0a64" exitCode=0 Dec 07 16:13:00 crc kubenswrapper[4716]: I1207 16:13:00.866577 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fh9qqw" event={"ID":"7ecea701-6ded-4f27-844f-fd9dfcb92f29","Type":"ContainerDied","Data":"2ad8eb95a1bb4b97051f0a32f7f621a9954bd07166035e29f1d739cb59ad0a64"} Dec 07 16:13:01 crc kubenswrapper[4716]: I1207 16:13:01.878203 4716 generic.go:334] "Generic (PLEG): container finished" podID="7ecea701-6ded-4f27-844f-fd9dfcb92f29" containerID="92d5a5e58f7929b3a537f3240cce72852ac5137faac7c27c47541e47a2919084" exitCode=0 Dec 07 16:13:01 crc kubenswrapper[4716]: I1207 16:13:01.878414 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fh9qqw" event={"ID":"7ecea701-6ded-4f27-844f-fd9dfcb92f29","Type":"ContainerDied","Data":"92d5a5e58f7929b3a537f3240cce72852ac5137faac7c27c47541e47a2919084"} Dec 07 16:13:03 crc kubenswrapper[4716]: I1207 16:13:03.146658 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fh9qqw" Dec 07 16:13:03 crc kubenswrapper[4716]: I1207 16:13:03.186109 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7ecea701-6ded-4f27-844f-fd9dfcb92f29-util\") pod \"7ecea701-6ded-4f27-844f-fd9dfcb92f29\" (UID: \"7ecea701-6ded-4f27-844f-fd9dfcb92f29\") " Dec 07 16:13:03 crc kubenswrapper[4716]: I1207 16:13:03.186231 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h9c5r\" (UniqueName: \"kubernetes.io/projected/7ecea701-6ded-4f27-844f-fd9dfcb92f29-kube-api-access-h9c5r\") pod \"7ecea701-6ded-4f27-844f-fd9dfcb92f29\" (UID: \"7ecea701-6ded-4f27-844f-fd9dfcb92f29\") " Dec 07 16:13:03 crc kubenswrapper[4716]: I1207 16:13:03.186325 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7ecea701-6ded-4f27-844f-fd9dfcb92f29-bundle\") pod \"7ecea701-6ded-4f27-844f-fd9dfcb92f29\" (UID: \"7ecea701-6ded-4f27-844f-fd9dfcb92f29\") " Dec 07 16:13:03 crc kubenswrapper[4716]: I1207 16:13:03.187192 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ecea701-6ded-4f27-844f-fd9dfcb92f29-bundle" (OuterVolumeSpecName: "bundle") pod "7ecea701-6ded-4f27-844f-fd9dfcb92f29" (UID: "7ecea701-6ded-4f27-844f-fd9dfcb92f29"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:13:03 crc kubenswrapper[4716]: I1207 16:13:03.196378 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ecea701-6ded-4f27-844f-fd9dfcb92f29-kube-api-access-h9c5r" (OuterVolumeSpecName: "kube-api-access-h9c5r") pod "7ecea701-6ded-4f27-844f-fd9dfcb92f29" (UID: "7ecea701-6ded-4f27-844f-fd9dfcb92f29"). InnerVolumeSpecName "kube-api-access-h9c5r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:13:03 crc kubenswrapper[4716]: I1207 16:13:03.200823 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ecea701-6ded-4f27-844f-fd9dfcb92f29-util" (OuterVolumeSpecName: "util") pod "7ecea701-6ded-4f27-844f-fd9dfcb92f29" (UID: "7ecea701-6ded-4f27-844f-fd9dfcb92f29"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:13:03 crc kubenswrapper[4716]: I1207 16:13:03.287384 4716 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7ecea701-6ded-4f27-844f-fd9dfcb92f29-util\") on node \"crc\" DevicePath \"\"" Dec 07 16:13:03 crc kubenswrapper[4716]: I1207 16:13:03.287421 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h9c5r\" (UniqueName: \"kubernetes.io/projected/7ecea701-6ded-4f27-844f-fd9dfcb92f29-kube-api-access-h9c5r\") on node \"crc\" DevicePath \"\"" Dec 07 16:13:03 crc kubenswrapper[4716]: I1207 16:13:03.287436 4716 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7ecea701-6ded-4f27-844f-fd9dfcb92f29-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 16:13:03 crc kubenswrapper[4716]: I1207 16:13:03.894290 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fh9qqw" Dec 07 16:13:03 crc kubenswrapper[4716]: I1207 16:13:03.894175 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fh9qqw" event={"ID":"7ecea701-6ded-4f27-844f-fd9dfcb92f29","Type":"ContainerDied","Data":"f43a1ecd52058bbb4bd5e7b29e5bc378456f3a22b7b9098a70f92da0e7c9ad83"} Dec 07 16:13:03 crc kubenswrapper[4716]: I1207 16:13:03.895478 4716 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f43a1ecd52058bbb4bd5e7b29e5bc378456f3a22b7b9098a70f92da0e7c9ad83" Dec 07 16:13:05 crc kubenswrapper[4716]: I1207 16:13:05.167266 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-s55qn"] Dec 07 16:13:05 crc kubenswrapper[4716]: E1207 16:13:05.167689 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ecea701-6ded-4f27-844f-fd9dfcb92f29" containerName="extract" Dec 07 16:13:05 crc kubenswrapper[4716]: I1207 16:13:05.167701 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ecea701-6ded-4f27-844f-fd9dfcb92f29" containerName="extract" Dec 07 16:13:05 crc kubenswrapper[4716]: E1207 16:13:05.167718 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ecea701-6ded-4f27-844f-fd9dfcb92f29" containerName="util" Dec 07 16:13:05 crc kubenswrapper[4716]: I1207 16:13:05.167724 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ecea701-6ded-4f27-844f-fd9dfcb92f29" containerName="util" Dec 07 16:13:05 crc kubenswrapper[4716]: E1207 16:13:05.167733 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ecea701-6ded-4f27-844f-fd9dfcb92f29" containerName="pull" Dec 07 16:13:05 crc kubenswrapper[4716]: I1207 16:13:05.167739 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ecea701-6ded-4f27-844f-fd9dfcb92f29" containerName="pull" Dec 07 16:13:05 crc kubenswrapper[4716]: I1207 16:13:05.167823 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ecea701-6ded-4f27-844f-fd9dfcb92f29" containerName="extract" Dec 07 16:13:05 crc kubenswrapper[4716]: I1207 16:13:05.168207 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-s55qn" Dec 07 16:13:05 crc kubenswrapper[4716]: I1207 16:13:05.169649 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Dec 07 16:13:05 crc kubenswrapper[4716]: I1207 16:13:05.169785 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-rcs2l" Dec 07 16:13:05 crc kubenswrapper[4716]: I1207 16:13:05.170636 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Dec 07 16:13:05 crc kubenswrapper[4716]: I1207 16:13:05.206109 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-s55qn"] Dec 07 16:13:05 crc kubenswrapper[4716]: I1207 16:13:05.206847 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rkbzb\" (UniqueName: \"kubernetes.io/projected/08d6dea5-f7a4-4895-b768-f6feebceab45-kube-api-access-rkbzb\") pod \"nmstate-operator-5b5b58f5c8-s55qn\" (UID: \"08d6dea5-f7a4-4895-b768-f6feebceab45\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-s55qn" Dec 07 16:13:05 crc kubenswrapper[4716]: I1207 16:13:05.307665 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rkbzb\" (UniqueName: \"kubernetes.io/projected/08d6dea5-f7a4-4895-b768-f6feebceab45-kube-api-access-rkbzb\") pod \"nmstate-operator-5b5b58f5c8-s55qn\" (UID: \"08d6dea5-f7a4-4895-b768-f6feebceab45\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-s55qn" Dec 07 16:13:05 crc kubenswrapper[4716]: I1207 16:13:05.323153 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rkbzb\" (UniqueName: \"kubernetes.io/projected/08d6dea5-f7a4-4895-b768-f6feebceab45-kube-api-access-rkbzb\") pod \"nmstate-operator-5b5b58f5c8-s55qn\" (UID: \"08d6dea5-f7a4-4895-b768-f6feebceab45\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-s55qn" Dec 07 16:13:05 crc kubenswrapper[4716]: I1207 16:13:05.482239 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-s55qn" Dec 07 16:13:05 crc kubenswrapper[4716]: W1207 16:13:05.962332 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod08d6dea5_f7a4_4895_b768_f6feebceab45.slice/crio-48a2a8c8824d64e070ffe96b95bed4d64ac108bfe28b8e4a412120bdd81bc7d4 WatchSource:0}: Error finding container 48a2a8c8824d64e070ffe96b95bed4d64ac108bfe28b8e4a412120bdd81bc7d4: Status 404 returned error can't find the container with id 48a2a8c8824d64e070ffe96b95bed4d64ac108bfe28b8e4a412120bdd81bc7d4 Dec 07 16:13:05 crc kubenswrapper[4716]: I1207 16:13:05.967520 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-s55qn"] Dec 07 16:13:06 crc kubenswrapper[4716]: I1207 16:13:06.914917 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-s55qn" event={"ID":"08d6dea5-f7a4-4895-b768-f6feebceab45","Type":"ContainerStarted","Data":"48a2a8c8824d64e070ffe96b95bed4d64ac108bfe28b8e4a412120bdd81bc7d4"} Dec 07 16:13:08 crc kubenswrapper[4716]: I1207 16:13:08.930447 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-s55qn" event={"ID":"08d6dea5-f7a4-4895-b768-f6feebceab45","Type":"ContainerStarted","Data":"20ec38a2a5e6aadc70d344afc7177f91b344a45c42f5fcc7122f64d535a8b007"} Dec 07 16:13:08 crc kubenswrapper[4716]: I1207 16:13:08.963696 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-s55qn" podStartSLOduration=1.965767584 podStartE2EDuration="3.963674382s" podCreationTimestamp="2025-12-07 16:13:05 +0000 UTC" firstStartedPulling="2025-12-07 16:13:05.965051663 +0000 UTC m=+648.655336615" lastFinishedPulling="2025-12-07 16:13:07.962958501 +0000 UTC m=+650.653243413" observedRunningTime="2025-12-07 16:13:08.959044829 +0000 UTC m=+651.649329781" watchObservedRunningTime="2025-12-07 16:13:08.963674382 +0000 UTC m=+651.653959314" Dec 07 16:13:09 crc kubenswrapper[4716]: I1207 16:13:09.966186 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-slf82"] Dec 07 16:13:09 crc kubenswrapper[4716]: I1207 16:13:09.967192 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-slf82" Dec 07 16:13:09 crc kubenswrapper[4716]: I1207 16:13:09.969160 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-rkbl6" Dec 07 16:13:09 crc kubenswrapper[4716]: I1207 16:13:09.979794 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-8tpxz"] Dec 07 16:13:09 crc kubenswrapper[4716]: I1207 16:13:09.980655 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-8tpxz" Dec 07 16:13:09 crc kubenswrapper[4716]: I1207 16:13:09.982564 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Dec 07 16:13:09 crc kubenswrapper[4716]: I1207 16:13:09.984727 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-slf82"] Dec 07 16:13:10 crc kubenswrapper[4716]: I1207 16:13:10.017120 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-8tpxz"] Dec 07 16:13:10 crc kubenswrapper[4716]: I1207 16:13:10.019636 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-89zc4"] Dec 07 16:13:10 crc kubenswrapper[4716]: I1207 16:13:10.020456 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-89zc4" Dec 07 16:13:10 crc kubenswrapper[4716]: I1207 16:13:10.142259 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-rqbh5"] Dec 07 16:13:10 crc kubenswrapper[4716]: I1207 16:13:10.143140 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-rqbh5" Dec 07 16:13:10 crc kubenswrapper[4716]: I1207 16:13:10.145123 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Dec 07 16:13:10 crc kubenswrapper[4716]: I1207 16:13:10.145613 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Dec 07 16:13:10 crc kubenswrapper[4716]: I1207 16:13:10.146175 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-gjdfr" Dec 07 16:13:10 crc kubenswrapper[4716]: I1207 16:13:10.155234 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-rqbh5"] Dec 07 16:13:10 crc kubenswrapper[4716]: I1207 16:13:10.164806 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dg9jj\" (UniqueName: \"kubernetes.io/projected/f54315f3-de9b-48b5-9baf-94e78414c0e2-kube-api-access-dg9jj\") pod \"nmstate-metrics-7f946cbc9-slf82\" (UID: \"f54315f3-de9b-48b5-9baf-94e78414c0e2\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-slf82" Dec 07 16:13:10 crc kubenswrapper[4716]: I1207 16:13:10.164865 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/e4dfc168-4ec1-40d0-9d1a-d7f04063f189-dbus-socket\") pod \"nmstate-handler-89zc4\" (UID: \"e4dfc168-4ec1-40d0-9d1a-d7f04063f189\") " pod="openshift-nmstate/nmstate-handler-89zc4" Dec 07 16:13:10 crc kubenswrapper[4716]: I1207 16:13:10.164903 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/e4dfc168-4ec1-40d0-9d1a-d7f04063f189-ovs-socket\") pod \"nmstate-handler-89zc4\" (UID: \"e4dfc168-4ec1-40d0-9d1a-d7f04063f189\") " pod="openshift-nmstate/nmstate-handler-89zc4" Dec 07 16:13:10 crc kubenswrapper[4716]: I1207 16:13:10.164947 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2vnkr\" (UniqueName: \"kubernetes.io/projected/e4dfc168-4ec1-40d0-9d1a-d7f04063f189-kube-api-access-2vnkr\") pod \"nmstate-handler-89zc4\" (UID: \"e4dfc168-4ec1-40d0-9d1a-d7f04063f189\") " pod="openshift-nmstate/nmstate-handler-89zc4" Dec 07 16:13:10 crc kubenswrapper[4716]: I1207 16:13:10.164971 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/3546fcc6-820f-4601-9c0e-b652481582d3-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-8tpxz\" (UID: \"3546fcc6-820f-4601-9c0e-b652481582d3\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-8tpxz" Dec 07 16:13:10 crc kubenswrapper[4716]: I1207 16:13:10.165018 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bj2v2\" (UniqueName: \"kubernetes.io/projected/3546fcc6-820f-4601-9c0e-b652481582d3-kube-api-access-bj2v2\") pod \"nmstate-webhook-5f6d4c5ccb-8tpxz\" (UID: \"3546fcc6-820f-4601-9c0e-b652481582d3\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-8tpxz" Dec 07 16:13:10 crc kubenswrapper[4716]: I1207 16:13:10.165042 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/e4dfc168-4ec1-40d0-9d1a-d7f04063f189-nmstate-lock\") pod \"nmstate-handler-89zc4\" (UID: \"e4dfc168-4ec1-40d0-9d1a-d7f04063f189\") " pod="openshift-nmstate/nmstate-handler-89zc4" Dec 07 16:13:10 crc kubenswrapper[4716]: I1207 16:13:10.266744 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/e4dfc168-4ec1-40d0-9d1a-d7f04063f189-ovs-socket\") pod \"nmstate-handler-89zc4\" (UID: \"e4dfc168-4ec1-40d0-9d1a-d7f04063f189\") " pod="openshift-nmstate/nmstate-handler-89zc4" Dec 07 16:13:10 crc kubenswrapper[4716]: I1207 16:13:10.266807 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/682c523b-c5ad-4e1c-9fd9-969911ef2242-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-rqbh5\" (UID: \"682c523b-c5ad-4e1c-9fd9-969911ef2242\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-rqbh5" Dec 07 16:13:10 crc kubenswrapper[4716]: I1207 16:13:10.266844 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/3546fcc6-820f-4601-9c0e-b652481582d3-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-8tpxz\" (UID: \"3546fcc6-820f-4601-9c0e-b652481582d3\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-8tpxz" Dec 07 16:13:10 crc kubenswrapper[4716]: I1207 16:13:10.266860 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2vnkr\" (UniqueName: \"kubernetes.io/projected/e4dfc168-4ec1-40d0-9d1a-d7f04063f189-kube-api-access-2vnkr\") pod \"nmstate-handler-89zc4\" (UID: \"e4dfc168-4ec1-40d0-9d1a-d7f04063f189\") " pod="openshift-nmstate/nmstate-handler-89zc4" Dec 07 16:13:10 crc kubenswrapper[4716]: I1207 16:13:10.266885 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bc8f5\" (UniqueName: \"kubernetes.io/projected/682c523b-c5ad-4e1c-9fd9-969911ef2242-kube-api-access-bc8f5\") pod \"nmstate-console-plugin-7fbb5f6569-rqbh5\" (UID: \"682c523b-c5ad-4e1c-9fd9-969911ef2242\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-rqbh5" Dec 07 16:13:10 crc kubenswrapper[4716]: I1207 16:13:10.266882 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/e4dfc168-4ec1-40d0-9d1a-d7f04063f189-ovs-socket\") pod \"nmstate-handler-89zc4\" (UID: \"e4dfc168-4ec1-40d0-9d1a-d7f04063f189\") " pod="openshift-nmstate/nmstate-handler-89zc4" Dec 07 16:13:10 crc kubenswrapper[4716]: I1207 16:13:10.266921 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bj2v2\" (UniqueName: \"kubernetes.io/projected/3546fcc6-820f-4601-9c0e-b652481582d3-kube-api-access-bj2v2\") pod \"nmstate-webhook-5f6d4c5ccb-8tpxz\" (UID: \"3546fcc6-820f-4601-9c0e-b652481582d3\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-8tpxz" Dec 07 16:13:10 crc kubenswrapper[4716]: I1207 16:13:10.266941 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/e4dfc168-4ec1-40d0-9d1a-d7f04063f189-nmstate-lock\") pod \"nmstate-handler-89zc4\" (UID: \"e4dfc168-4ec1-40d0-9d1a-d7f04063f189\") " pod="openshift-nmstate/nmstate-handler-89zc4" Dec 07 16:13:10 crc kubenswrapper[4716]: E1207 16:13:10.267044 4716 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Dec 07 16:13:10 crc kubenswrapper[4716]: I1207 16:13:10.267056 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/e4dfc168-4ec1-40d0-9d1a-d7f04063f189-nmstate-lock\") pod \"nmstate-handler-89zc4\" (UID: \"e4dfc168-4ec1-40d0-9d1a-d7f04063f189\") " pod="openshift-nmstate/nmstate-handler-89zc4" Dec 07 16:13:10 crc kubenswrapper[4716]: E1207 16:13:10.267134 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3546fcc6-820f-4601-9c0e-b652481582d3-tls-key-pair podName:3546fcc6-820f-4601-9c0e-b652481582d3 nodeName:}" failed. No retries permitted until 2025-12-07 16:13:10.767113876 +0000 UTC m=+653.457398788 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/3546fcc6-820f-4601-9c0e-b652481582d3-tls-key-pair") pod "nmstate-webhook-5f6d4c5ccb-8tpxz" (UID: "3546fcc6-820f-4601-9c0e-b652481582d3") : secret "openshift-nmstate-webhook" not found Dec 07 16:13:10 crc kubenswrapper[4716]: I1207 16:13:10.267293 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dg9jj\" (UniqueName: \"kubernetes.io/projected/f54315f3-de9b-48b5-9baf-94e78414c0e2-kube-api-access-dg9jj\") pod \"nmstate-metrics-7f946cbc9-slf82\" (UID: \"f54315f3-de9b-48b5-9baf-94e78414c0e2\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-slf82" Dec 07 16:13:10 crc kubenswrapper[4716]: I1207 16:13:10.267367 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/682c523b-c5ad-4e1c-9fd9-969911ef2242-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-rqbh5\" (UID: \"682c523b-c5ad-4e1c-9fd9-969911ef2242\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-rqbh5" Dec 07 16:13:10 crc kubenswrapper[4716]: I1207 16:13:10.267397 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/e4dfc168-4ec1-40d0-9d1a-d7f04063f189-dbus-socket\") pod \"nmstate-handler-89zc4\" (UID: \"e4dfc168-4ec1-40d0-9d1a-d7f04063f189\") " pod="openshift-nmstate/nmstate-handler-89zc4" Dec 07 16:13:10 crc kubenswrapper[4716]: I1207 16:13:10.267683 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/e4dfc168-4ec1-40d0-9d1a-d7f04063f189-dbus-socket\") pod \"nmstate-handler-89zc4\" (UID: \"e4dfc168-4ec1-40d0-9d1a-d7f04063f189\") " pod="openshift-nmstate/nmstate-handler-89zc4" Dec 07 16:13:10 crc kubenswrapper[4716]: I1207 16:13:10.291648 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2vnkr\" (UniqueName: \"kubernetes.io/projected/e4dfc168-4ec1-40d0-9d1a-d7f04063f189-kube-api-access-2vnkr\") pod \"nmstate-handler-89zc4\" (UID: \"e4dfc168-4ec1-40d0-9d1a-d7f04063f189\") " pod="openshift-nmstate/nmstate-handler-89zc4" Dec 07 16:13:10 crc kubenswrapper[4716]: I1207 16:13:10.291863 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bj2v2\" (UniqueName: \"kubernetes.io/projected/3546fcc6-820f-4601-9c0e-b652481582d3-kube-api-access-bj2v2\") pod \"nmstate-webhook-5f6d4c5ccb-8tpxz\" (UID: \"3546fcc6-820f-4601-9c0e-b652481582d3\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-8tpxz" Dec 07 16:13:10 crc kubenswrapper[4716]: I1207 16:13:10.291899 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dg9jj\" (UniqueName: \"kubernetes.io/projected/f54315f3-de9b-48b5-9baf-94e78414c0e2-kube-api-access-dg9jj\") pod \"nmstate-metrics-7f946cbc9-slf82\" (UID: \"f54315f3-de9b-48b5-9baf-94e78414c0e2\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-slf82" Dec 07 16:13:10 crc kubenswrapper[4716]: I1207 16:13:10.294518 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-slf82" Dec 07 16:13:10 crc kubenswrapper[4716]: I1207 16:13:10.323844 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-7546cdc554-llcqg"] Dec 07 16:13:10 crc kubenswrapper[4716]: I1207 16:13:10.325048 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7546cdc554-llcqg" Dec 07 16:13:10 crc kubenswrapper[4716]: I1207 16:13:10.339574 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-89zc4" Dec 07 16:13:10 crc kubenswrapper[4716]: I1207 16:13:10.343902 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-7546cdc554-llcqg"] Dec 07 16:13:10 crc kubenswrapper[4716]: I1207 16:13:10.371141 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b57c69b4-2bd5-499b-809d-6b60c7f802b3-service-ca\") pod \"console-7546cdc554-llcqg\" (UID: \"b57c69b4-2bd5-499b-809d-6b60c7f802b3\") " pod="openshift-console/console-7546cdc554-llcqg" Dec 07 16:13:10 crc kubenswrapper[4716]: I1207 16:13:10.371191 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bc8f5\" (UniqueName: \"kubernetes.io/projected/682c523b-c5ad-4e1c-9fd9-969911ef2242-kube-api-access-bc8f5\") pod \"nmstate-console-plugin-7fbb5f6569-rqbh5\" (UID: \"682c523b-c5ad-4e1c-9fd9-969911ef2242\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-rqbh5" Dec 07 16:13:10 crc kubenswrapper[4716]: I1207 16:13:10.371212 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b57c69b4-2bd5-499b-809d-6b60c7f802b3-console-oauth-config\") pod \"console-7546cdc554-llcqg\" (UID: \"b57c69b4-2bd5-499b-809d-6b60c7f802b3\") " pod="openshift-console/console-7546cdc554-llcqg" Dec 07 16:13:10 crc kubenswrapper[4716]: I1207 16:13:10.371248 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b57c69b4-2bd5-499b-809d-6b60c7f802b3-console-serving-cert\") pod \"console-7546cdc554-llcqg\" (UID: \"b57c69b4-2bd5-499b-809d-6b60c7f802b3\") " pod="openshift-console/console-7546cdc554-llcqg" Dec 07 16:13:10 crc kubenswrapper[4716]: I1207 16:13:10.371267 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6nqn\" (UniqueName: \"kubernetes.io/projected/b57c69b4-2bd5-499b-809d-6b60c7f802b3-kube-api-access-v6nqn\") pod \"console-7546cdc554-llcqg\" (UID: \"b57c69b4-2bd5-499b-809d-6b60c7f802b3\") " pod="openshift-console/console-7546cdc554-llcqg" Dec 07 16:13:10 crc kubenswrapper[4716]: I1207 16:13:10.371302 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/682c523b-c5ad-4e1c-9fd9-969911ef2242-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-rqbh5\" (UID: \"682c523b-c5ad-4e1c-9fd9-969911ef2242\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-rqbh5" Dec 07 16:13:10 crc kubenswrapper[4716]: I1207 16:13:10.371337 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b57c69b4-2bd5-499b-809d-6b60c7f802b3-trusted-ca-bundle\") pod \"console-7546cdc554-llcqg\" (UID: \"b57c69b4-2bd5-499b-809d-6b60c7f802b3\") " pod="openshift-console/console-7546cdc554-llcqg" Dec 07 16:13:10 crc kubenswrapper[4716]: I1207 16:13:10.371355 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b57c69b4-2bd5-499b-809d-6b60c7f802b3-oauth-serving-cert\") pod \"console-7546cdc554-llcqg\" (UID: \"b57c69b4-2bd5-499b-809d-6b60c7f802b3\") " pod="openshift-console/console-7546cdc554-llcqg" Dec 07 16:13:10 crc kubenswrapper[4716]: I1207 16:13:10.371380 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/682c523b-c5ad-4e1c-9fd9-969911ef2242-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-rqbh5\" (UID: \"682c523b-c5ad-4e1c-9fd9-969911ef2242\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-rqbh5" Dec 07 16:13:10 crc kubenswrapper[4716]: I1207 16:13:10.371400 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b57c69b4-2bd5-499b-809d-6b60c7f802b3-console-config\") pod \"console-7546cdc554-llcqg\" (UID: \"b57c69b4-2bd5-499b-809d-6b60c7f802b3\") " pod="openshift-console/console-7546cdc554-llcqg" Dec 07 16:13:10 crc kubenswrapper[4716]: E1207 16:13:10.372240 4716 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Dec 07 16:13:10 crc kubenswrapper[4716]: E1207 16:13:10.372304 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/682c523b-c5ad-4e1c-9fd9-969911ef2242-plugin-serving-cert podName:682c523b-c5ad-4e1c-9fd9-969911ef2242 nodeName:}" failed. No retries permitted until 2025-12-07 16:13:10.872287587 +0000 UTC m=+653.562572499 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/682c523b-c5ad-4e1c-9fd9-969911ef2242-plugin-serving-cert") pod "nmstate-console-plugin-7fbb5f6569-rqbh5" (UID: "682c523b-c5ad-4e1c-9fd9-969911ef2242") : secret "plugin-serving-cert" not found Dec 07 16:13:10 crc kubenswrapper[4716]: I1207 16:13:10.372392 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/682c523b-c5ad-4e1c-9fd9-969911ef2242-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-rqbh5\" (UID: \"682c523b-c5ad-4e1c-9fd9-969911ef2242\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-rqbh5" Dec 07 16:13:10 crc kubenswrapper[4716]: I1207 16:13:10.393818 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bc8f5\" (UniqueName: \"kubernetes.io/projected/682c523b-c5ad-4e1c-9fd9-969911ef2242-kube-api-access-bc8f5\") pod \"nmstate-console-plugin-7fbb5f6569-rqbh5\" (UID: \"682c523b-c5ad-4e1c-9fd9-969911ef2242\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-rqbh5" Dec 07 16:13:10 crc kubenswrapper[4716]: I1207 16:13:10.471855 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b57c69b4-2bd5-499b-809d-6b60c7f802b3-oauth-serving-cert\") pod \"console-7546cdc554-llcqg\" (UID: \"b57c69b4-2bd5-499b-809d-6b60c7f802b3\") " pod="openshift-console/console-7546cdc554-llcqg" Dec 07 16:13:10 crc kubenswrapper[4716]: I1207 16:13:10.471936 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b57c69b4-2bd5-499b-809d-6b60c7f802b3-console-config\") pod \"console-7546cdc554-llcqg\" (UID: \"b57c69b4-2bd5-499b-809d-6b60c7f802b3\") " pod="openshift-console/console-7546cdc554-llcqg" Dec 07 16:13:10 crc kubenswrapper[4716]: I1207 16:13:10.472529 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b57c69b4-2bd5-499b-809d-6b60c7f802b3-service-ca\") pod \"console-7546cdc554-llcqg\" (UID: \"b57c69b4-2bd5-499b-809d-6b60c7f802b3\") " pod="openshift-console/console-7546cdc554-llcqg" Dec 07 16:13:10 crc kubenswrapper[4716]: I1207 16:13:10.472569 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b57c69b4-2bd5-499b-809d-6b60c7f802b3-console-oauth-config\") pod \"console-7546cdc554-llcqg\" (UID: \"b57c69b4-2bd5-499b-809d-6b60c7f802b3\") " pod="openshift-console/console-7546cdc554-llcqg" Dec 07 16:13:10 crc kubenswrapper[4716]: I1207 16:13:10.472641 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b57c69b4-2bd5-499b-809d-6b60c7f802b3-console-serving-cert\") pod \"console-7546cdc554-llcqg\" (UID: \"b57c69b4-2bd5-499b-809d-6b60c7f802b3\") " pod="openshift-console/console-7546cdc554-llcqg" Dec 07 16:13:10 crc kubenswrapper[4716]: I1207 16:13:10.472664 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6nqn\" (UniqueName: \"kubernetes.io/projected/b57c69b4-2bd5-499b-809d-6b60c7f802b3-kube-api-access-v6nqn\") pod \"console-7546cdc554-llcqg\" (UID: \"b57c69b4-2bd5-499b-809d-6b60c7f802b3\") " pod="openshift-console/console-7546cdc554-llcqg" Dec 07 16:13:10 crc kubenswrapper[4716]: I1207 16:13:10.472772 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b57c69b4-2bd5-499b-809d-6b60c7f802b3-trusted-ca-bundle\") pod \"console-7546cdc554-llcqg\" (UID: \"b57c69b4-2bd5-499b-809d-6b60c7f802b3\") " pod="openshift-console/console-7546cdc554-llcqg" Dec 07 16:13:10 crc kubenswrapper[4716]: I1207 16:13:10.473523 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b57c69b4-2bd5-499b-809d-6b60c7f802b3-oauth-serving-cert\") pod \"console-7546cdc554-llcqg\" (UID: \"b57c69b4-2bd5-499b-809d-6b60c7f802b3\") " pod="openshift-console/console-7546cdc554-llcqg" Dec 07 16:13:10 crc kubenswrapper[4716]: I1207 16:13:10.474306 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b57c69b4-2bd5-499b-809d-6b60c7f802b3-trusted-ca-bundle\") pod \"console-7546cdc554-llcqg\" (UID: \"b57c69b4-2bd5-499b-809d-6b60c7f802b3\") " pod="openshift-console/console-7546cdc554-llcqg" Dec 07 16:13:10 crc kubenswrapper[4716]: I1207 16:13:10.475043 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b57c69b4-2bd5-499b-809d-6b60c7f802b3-service-ca\") pod \"console-7546cdc554-llcqg\" (UID: \"b57c69b4-2bd5-499b-809d-6b60c7f802b3\") " pod="openshift-console/console-7546cdc554-llcqg" Dec 07 16:13:10 crc kubenswrapper[4716]: I1207 16:13:10.475527 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b57c69b4-2bd5-499b-809d-6b60c7f802b3-console-config\") pod \"console-7546cdc554-llcqg\" (UID: \"b57c69b4-2bd5-499b-809d-6b60c7f802b3\") " pod="openshift-console/console-7546cdc554-llcqg" Dec 07 16:13:10 crc kubenswrapper[4716]: I1207 16:13:10.479018 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b57c69b4-2bd5-499b-809d-6b60c7f802b3-console-serving-cert\") pod \"console-7546cdc554-llcqg\" (UID: \"b57c69b4-2bd5-499b-809d-6b60c7f802b3\") " pod="openshift-console/console-7546cdc554-llcqg" Dec 07 16:13:10 crc kubenswrapper[4716]: I1207 16:13:10.486051 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b57c69b4-2bd5-499b-809d-6b60c7f802b3-console-oauth-config\") pod \"console-7546cdc554-llcqg\" (UID: \"b57c69b4-2bd5-499b-809d-6b60c7f802b3\") " pod="openshift-console/console-7546cdc554-llcqg" Dec 07 16:13:10 crc kubenswrapper[4716]: I1207 16:13:10.489487 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6nqn\" (UniqueName: \"kubernetes.io/projected/b57c69b4-2bd5-499b-809d-6b60c7f802b3-kube-api-access-v6nqn\") pod \"console-7546cdc554-llcqg\" (UID: \"b57c69b4-2bd5-499b-809d-6b60c7f802b3\") " pod="openshift-console/console-7546cdc554-llcqg" Dec 07 16:13:10 crc kubenswrapper[4716]: I1207 16:13:10.534351 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-slf82"] Dec 07 16:13:10 crc kubenswrapper[4716]: I1207 16:13:10.671864 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7546cdc554-llcqg" Dec 07 16:13:10 crc kubenswrapper[4716]: I1207 16:13:10.780168 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/3546fcc6-820f-4601-9c0e-b652481582d3-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-8tpxz\" (UID: \"3546fcc6-820f-4601-9c0e-b652481582d3\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-8tpxz" Dec 07 16:13:10 crc kubenswrapper[4716]: I1207 16:13:10.786017 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/3546fcc6-820f-4601-9c0e-b652481582d3-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-8tpxz\" (UID: \"3546fcc6-820f-4601-9c0e-b652481582d3\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-8tpxz" Dec 07 16:13:10 crc kubenswrapper[4716]: I1207 16:13:10.881464 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/682c523b-c5ad-4e1c-9fd9-969911ef2242-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-rqbh5\" (UID: \"682c523b-c5ad-4e1c-9fd9-969911ef2242\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-rqbh5" Dec 07 16:13:10 crc kubenswrapper[4716]: I1207 16:13:10.884703 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/682c523b-c5ad-4e1c-9fd9-969911ef2242-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-rqbh5\" (UID: \"682c523b-c5ad-4e1c-9fd9-969911ef2242\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-rqbh5" Dec 07 16:13:10 crc kubenswrapper[4716]: I1207 16:13:10.885951 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-7546cdc554-llcqg"] Dec 07 16:13:10 crc kubenswrapper[4716]: I1207 16:13:10.906015 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-8tpxz" Dec 07 16:13:10 crc kubenswrapper[4716]: I1207 16:13:10.943134 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7546cdc554-llcqg" event={"ID":"b57c69b4-2bd5-499b-809d-6b60c7f802b3","Type":"ContainerStarted","Data":"19e96daf04a9ea6356e50e638db943d684635876cb20c4bfbd8b00a75a543e65"} Dec 07 16:13:10 crc kubenswrapper[4716]: I1207 16:13:10.944329 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-slf82" event={"ID":"f54315f3-de9b-48b5-9baf-94e78414c0e2","Type":"ContainerStarted","Data":"3ff1d497aa73fb45bb1647f9a80ed5f99f0c64d4b3ed5043f38f597b1e1e65e4"} Dec 07 16:13:10 crc kubenswrapper[4716]: I1207 16:13:10.945403 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-89zc4" event={"ID":"e4dfc168-4ec1-40d0-9d1a-d7f04063f189","Type":"ContainerStarted","Data":"d8ee519033dd8d75dcaa917a1ea1a4de1fefa494d77b5d871243f70e31175a0a"} Dec 07 16:13:11 crc kubenswrapper[4716]: I1207 16:13:11.095999 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-rqbh5" Dec 07 16:13:11 crc kubenswrapper[4716]: I1207 16:13:11.119322 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-8tpxz"] Dec 07 16:13:11 crc kubenswrapper[4716]: W1207 16:13:11.123026 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3546fcc6_820f_4601_9c0e_b652481582d3.slice/crio-243688e393e1e83cb8757db57dd2c7000808d0ca4f8f75cf9fbaa44817e01f66 WatchSource:0}: Error finding container 243688e393e1e83cb8757db57dd2c7000808d0ca4f8f75cf9fbaa44817e01f66: Status 404 returned error can't find the container with id 243688e393e1e83cb8757db57dd2c7000808d0ca4f8f75cf9fbaa44817e01f66 Dec 07 16:13:11 crc kubenswrapper[4716]: I1207 16:13:11.270312 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-rqbh5"] Dec 07 16:13:11 crc kubenswrapper[4716]: W1207 16:13:11.274859 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod682c523b_c5ad_4e1c_9fd9_969911ef2242.slice/crio-599d84c6026c0bba5e25ff76c303156d35efe2b42fd038a904d61bddfc72a27d WatchSource:0}: Error finding container 599d84c6026c0bba5e25ff76c303156d35efe2b42fd038a904d61bddfc72a27d: Status 404 returned error can't find the container with id 599d84c6026c0bba5e25ff76c303156d35efe2b42fd038a904d61bddfc72a27d Dec 07 16:13:11 crc kubenswrapper[4716]: I1207 16:13:11.951256 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-8tpxz" event={"ID":"3546fcc6-820f-4601-9c0e-b652481582d3","Type":"ContainerStarted","Data":"243688e393e1e83cb8757db57dd2c7000808d0ca4f8f75cf9fbaa44817e01f66"} Dec 07 16:13:11 crc kubenswrapper[4716]: I1207 16:13:11.953555 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7546cdc554-llcqg" event={"ID":"b57c69b4-2bd5-499b-809d-6b60c7f802b3","Type":"ContainerStarted","Data":"60e4c22544e2b3083f218046976b4614a7ac30575b8228ac85f1809f56e1375a"} Dec 07 16:13:11 crc kubenswrapper[4716]: I1207 16:13:11.955626 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-rqbh5" event={"ID":"682c523b-c5ad-4e1c-9fd9-969911ef2242","Type":"ContainerStarted","Data":"599d84c6026c0bba5e25ff76c303156d35efe2b42fd038a904d61bddfc72a27d"} Dec 07 16:13:11 crc kubenswrapper[4716]: I1207 16:13:11.971382 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-7546cdc554-llcqg" podStartSLOduration=1.971362444 podStartE2EDuration="1.971362444s" podCreationTimestamp="2025-12-07 16:13:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:13:11.969097323 +0000 UTC m=+654.659382235" watchObservedRunningTime="2025-12-07 16:13:11.971362444 +0000 UTC m=+654.661647356" Dec 07 16:13:13 crc kubenswrapper[4716]: I1207 16:13:13.980908 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-89zc4" event={"ID":"e4dfc168-4ec1-40d0-9d1a-d7f04063f189","Type":"ContainerStarted","Data":"0167aa4ccaa8157a2b1cb97c6306c4e0206a206881b58a0bd67f6d7aab5583a7"} Dec 07 16:13:13 crc kubenswrapper[4716]: I1207 16:13:13.981505 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-89zc4" Dec 07 16:13:13 crc kubenswrapper[4716]: I1207 16:13:13.984431 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-8tpxz" event={"ID":"3546fcc6-820f-4601-9c0e-b652481582d3","Type":"ContainerStarted","Data":"86626197d4708cd052afd70b7cc8391d469a59c0d65e551c95d7579b015cb623"} Dec 07 16:13:13 crc kubenswrapper[4716]: I1207 16:13:13.984543 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-8tpxz" Dec 07 16:13:13 crc kubenswrapper[4716]: I1207 16:13:13.986504 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-rqbh5" event={"ID":"682c523b-c5ad-4e1c-9fd9-969911ef2242","Type":"ContainerStarted","Data":"f6e011bcaca9b2f20ae74f9fc259758bde265bb2a1ecec9f7ba85c1eebe76867"} Dec 07 16:13:13 crc kubenswrapper[4716]: I1207 16:13:13.987833 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-slf82" event={"ID":"f54315f3-de9b-48b5-9baf-94e78414c0e2","Type":"ContainerStarted","Data":"11275d60dfece00f43c3d2bfc41f77a1c876feed00d0d0cec85e7e634fd6b867"} Dec 07 16:13:14 crc kubenswrapper[4716]: I1207 16:13:14.000734 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-89zc4" podStartSLOduration=2.591584961 podStartE2EDuration="5.000720202s" podCreationTimestamp="2025-12-07 16:13:09 +0000 UTC" firstStartedPulling="2025-12-07 16:13:10.389388214 +0000 UTC m=+653.079673126" lastFinishedPulling="2025-12-07 16:13:12.798523445 +0000 UTC m=+655.488808367" observedRunningTime="2025-12-07 16:13:13.998500924 +0000 UTC m=+656.688785886" watchObservedRunningTime="2025-12-07 16:13:14.000720202 +0000 UTC m=+656.691005114" Dec 07 16:13:14 crc kubenswrapper[4716]: I1207 16:13:14.021309 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-8tpxz" podStartSLOduration=3.334545403 podStartE2EDuration="5.021287873s" podCreationTimestamp="2025-12-07 16:13:09 +0000 UTC" firstStartedPulling="2025-12-07 16:13:11.127965288 +0000 UTC m=+653.818250200" lastFinishedPulling="2025-12-07 16:13:12.814707758 +0000 UTC m=+655.504992670" observedRunningTime="2025-12-07 16:13:14.020241714 +0000 UTC m=+656.710526626" watchObservedRunningTime="2025-12-07 16:13:14.021287873 +0000 UTC m=+656.711572785" Dec 07 16:13:14 crc kubenswrapper[4716]: I1207 16:13:14.072213 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-rqbh5" podStartSLOduration=1.5854309469999999 podStartE2EDuration="4.072187753s" podCreationTimestamp="2025-12-07 16:13:10 +0000 UTC" firstStartedPulling="2025-12-07 16:13:11.277099865 +0000 UTC m=+653.967384787" lastFinishedPulling="2025-12-07 16:13:13.763856671 +0000 UTC m=+656.454141593" observedRunningTime="2025-12-07 16:13:14.069423789 +0000 UTC m=+656.759708791" watchObservedRunningTime="2025-12-07 16:13:14.072187753 +0000 UTC m=+656.762472705" Dec 07 16:13:16 crc kubenswrapper[4716]: I1207 16:13:16.007713 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-slf82" event={"ID":"f54315f3-de9b-48b5-9baf-94e78414c0e2","Type":"ContainerStarted","Data":"13f499d7577bc330d427ee5e7ff3bf9dcfe85eb8003e0b3f52f8ffba8d77ad39"} Dec 07 16:13:16 crc kubenswrapper[4716]: I1207 16:13:16.039769 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-slf82" podStartSLOduration=2.432581472 podStartE2EDuration="7.03974619s" podCreationTimestamp="2025-12-07 16:13:09 +0000 UTC" firstStartedPulling="2025-12-07 16:13:10.533300022 +0000 UTC m=+653.223584974" lastFinishedPulling="2025-12-07 16:13:15.14046479 +0000 UTC m=+657.830749692" observedRunningTime="2025-12-07 16:13:16.035172828 +0000 UTC m=+658.725457800" watchObservedRunningTime="2025-12-07 16:13:16.03974619 +0000 UTC m=+658.730031112" Dec 07 16:13:20 crc kubenswrapper[4716]: I1207 16:13:20.378197 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-89zc4" Dec 07 16:13:20 crc kubenswrapper[4716]: I1207 16:13:20.672334 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-7546cdc554-llcqg" Dec 07 16:13:20 crc kubenswrapper[4716]: I1207 16:13:20.672403 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-7546cdc554-llcqg" Dec 07 16:13:20 crc kubenswrapper[4716]: I1207 16:13:20.680412 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-7546cdc554-llcqg" Dec 07 16:13:21 crc kubenswrapper[4716]: I1207 16:13:21.049911 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-7546cdc554-llcqg" Dec 07 16:13:21 crc kubenswrapper[4716]: I1207 16:13:21.124920 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-ndp5v"] Dec 07 16:13:30 crc kubenswrapper[4716]: I1207 16:13:30.915192 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-8tpxz" Dec 07 16:13:46 crc kubenswrapper[4716]: I1207 16:13:46.198799 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-ndp5v" podUID="cbf3f592-bbb1-4596-8ae1-0e62aa366f0b" containerName="console" containerID="cri-o://98b5a90edae2b9e0c3df9cc09a9dd99f59934f6e38e7da5499cddff1743f5663" gracePeriod=15 Dec 07 16:13:46 crc kubenswrapper[4716]: I1207 16:13:46.571738 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-ndp5v_cbf3f592-bbb1-4596-8ae1-0e62aa366f0b/console/0.log" Dec 07 16:13:46 crc kubenswrapper[4716]: I1207 16:13:46.572180 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-ndp5v" Dec 07 16:13:46 crc kubenswrapper[4716]: I1207 16:13:46.716628 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cbf3f592-bbb1-4596-8ae1-0e62aa366f0b-trusted-ca-bundle\") pod \"cbf3f592-bbb1-4596-8ae1-0e62aa366f0b\" (UID: \"cbf3f592-bbb1-4596-8ae1-0e62aa366f0b\") " Dec 07 16:13:46 crc kubenswrapper[4716]: I1207 16:13:46.716717 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/cbf3f592-bbb1-4596-8ae1-0e62aa366f0b-console-oauth-config\") pod \"cbf3f592-bbb1-4596-8ae1-0e62aa366f0b\" (UID: \"cbf3f592-bbb1-4596-8ae1-0e62aa366f0b\") " Dec 07 16:13:46 crc kubenswrapper[4716]: I1207 16:13:46.716750 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tmljs\" (UniqueName: \"kubernetes.io/projected/cbf3f592-bbb1-4596-8ae1-0e62aa366f0b-kube-api-access-tmljs\") pod \"cbf3f592-bbb1-4596-8ae1-0e62aa366f0b\" (UID: \"cbf3f592-bbb1-4596-8ae1-0e62aa366f0b\") " Dec 07 16:13:46 crc kubenswrapper[4716]: I1207 16:13:46.716787 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/cbf3f592-bbb1-4596-8ae1-0e62aa366f0b-console-config\") pod \"cbf3f592-bbb1-4596-8ae1-0e62aa366f0b\" (UID: \"cbf3f592-bbb1-4596-8ae1-0e62aa366f0b\") " Dec 07 16:13:46 crc kubenswrapper[4716]: I1207 16:13:46.716828 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/cbf3f592-bbb1-4596-8ae1-0e62aa366f0b-service-ca\") pod \"cbf3f592-bbb1-4596-8ae1-0e62aa366f0b\" (UID: \"cbf3f592-bbb1-4596-8ae1-0e62aa366f0b\") " Dec 07 16:13:46 crc kubenswrapper[4716]: I1207 16:13:46.716849 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/cbf3f592-bbb1-4596-8ae1-0e62aa366f0b-oauth-serving-cert\") pod \"cbf3f592-bbb1-4596-8ae1-0e62aa366f0b\" (UID: \"cbf3f592-bbb1-4596-8ae1-0e62aa366f0b\") " Dec 07 16:13:46 crc kubenswrapper[4716]: I1207 16:13:46.716918 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/cbf3f592-bbb1-4596-8ae1-0e62aa366f0b-console-serving-cert\") pod \"cbf3f592-bbb1-4596-8ae1-0e62aa366f0b\" (UID: \"cbf3f592-bbb1-4596-8ae1-0e62aa366f0b\") " Dec 07 16:13:46 crc kubenswrapper[4716]: I1207 16:13:46.717558 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cbf3f592-bbb1-4596-8ae1-0e62aa366f0b-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "cbf3f592-bbb1-4596-8ae1-0e62aa366f0b" (UID: "cbf3f592-bbb1-4596-8ae1-0e62aa366f0b"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:13:46 crc kubenswrapper[4716]: I1207 16:13:46.718196 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cbf3f592-bbb1-4596-8ae1-0e62aa366f0b-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "cbf3f592-bbb1-4596-8ae1-0e62aa366f0b" (UID: "cbf3f592-bbb1-4596-8ae1-0e62aa366f0b"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:13:46 crc kubenswrapper[4716]: I1207 16:13:46.718212 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cbf3f592-bbb1-4596-8ae1-0e62aa366f0b-console-config" (OuterVolumeSpecName: "console-config") pod "cbf3f592-bbb1-4596-8ae1-0e62aa366f0b" (UID: "cbf3f592-bbb1-4596-8ae1-0e62aa366f0b"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:13:46 crc kubenswrapper[4716]: I1207 16:13:46.718222 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cbf3f592-bbb1-4596-8ae1-0e62aa366f0b-service-ca" (OuterVolumeSpecName: "service-ca") pod "cbf3f592-bbb1-4596-8ae1-0e62aa366f0b" (UID: "cbf3f592-bbb1-4596-8ae1-0e62aa366f0b"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:13:46 crc kubenswrapper[4716]: I1207 16:13:46.722982 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cbf3f592-bbb1-4596-8ae1-0e62aa366f0b-kube-api-access-tmljs" (OuterVolumeSpecName: "kube-api-access-tmljs") pod "cbf3f592-bbb1-4596-8ae1-0e62aa366f0b" (UID: "cbf3f592-bbb1-4596-8ae1-0e62aa366f0b"). InnerVolumeSpecName "kube-api-access-tmljs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:13:46 crc kubenswrapper[4716]: I1207 16:13:46.724319 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cbf3f592-bbb1-4596-8ae1-0e62aa366f0b-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "cbf3f592-bbb1-4596-8ae1-0e62aa366f0b" (UID: "cbf3f592-bbb1-4596-8ae1-0e62aa366f0b"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:13:46 crc kubenswrapper[4716]: I1207 16:13:46.724548 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cbf3f592-bbb1-4596-8ae1-0e62aa366f0b-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "cbf3f592-bbb1-4596-8ae1-0e62aa366f0b" (UID: "cbf3f592-bbb1-4596-8ae1-0e62aa366f0b"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:13:46 crc kubenswrapper[4716]: I1207 16:13:46.778364 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c2vdd"] Dec 07 16:13:46 crc kubenswrapper[4716]: E1207 16:13:46.778740 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbf3f592-bbb1-4596-8ae1-0e62aa366f0b" containerName="console" Dec 07 16:13:46 crc kubenswrapper[4716]: I1207 16:13:46.778805 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbf3f592-bbb1-4596-8ae1-0e62aa366f0b" containerName="console" Dec 07 16:13:46 crc kubenswrapper[4716]: I1207 16:13:46.778955 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="cbf3f592-bbb1-4596-8ae1-0e62aa366f0b" containerName="console" Dec 07 16:13:46 crc kubenswrapper[4716]: I1207 16:13:46.779698 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c2vdd" Dec 07 16:13:46 crc kubenswrapper[4716]: I1207 16:13:46.781646 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 07 16:13:46 crc kubenswrapper[4716]: I1207 16:13:46.789843 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c2vdd"] Dec 07 16:13:46 crc kubenswrapper[4716]: I1207 16:13:46.818440 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/19363c75-987f-4c3b-bbca-28a9cc33238b-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c2vdd\" (UID: \"19363c75-987f-4c3b-bbca-28a9cc33238b\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c2vdd" Dec 07 16:13:46 crc kubenswrapper[4716]: I1207 16:13:46.818491 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4blr\" (UniqueName: \"kubernetes.io/projected/19363c75-987f-4c3b-bbca-28a9cc33238b-kube-api-access-x4blr\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c2vdd\" (UID: \"19363c75-987f-4c3b-bbca-28a9cc33238b\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c2vdd" Dec 07 16:13:46 crc kubenswrapper[4716]: I1207 16:13:46.818515 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/19363c75-987f-4c3b-bbca-28a9cc33238b-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c2vdd\" (UID: \"19363c75-987f-4c3b-bbca-28a9cc33238b\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c2vdd" Dec 07 16:13:46 crc kubenswrapper[4716]: I1207 16:13:46.818583 4716 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/cbf3f592-bbb1-4596-8ae1-0e62aa366f0b-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 07 16:13:46 crc kubenswrapper[4716]: I1207 16:13:46.818595 4716 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cbf3f592-bbb1-4596-8ae1-0e62aa366f0b-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 16:13:46 crc kubenswrapper[4716]: I1207 16:13:46.818604 4716 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/cbf3f592-bbb1-4596-8ae1-0e62aa366f0b-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 07 16:13:46 crc kubenswrapper[4716]: I1207 16:13:46.818613 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tmljs\" (UniqueName: \"kubernetes.io/projected/cbf3f592-bbb1-4596-8ae1-0e62aa366f0b-kube-api-access-tmljs\") on node \"crc\" DevicePath \"\"" Dec 07 16:13:46 crc kubenswrapper[4716]: I1207 16:13:46.818625 4716 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/cbf3f592-bbb1-4596-8ae1-0e62aa366f0b-console-config\") on node \"crc\" DevicePath \"\"" Dec 07 16:13:46 crc kubenswrapper[4716]: I1207 16:13:46.818636 4716 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/cbf3f592-bbb1-4596-8ae1-0e62aa366f0b-service-ca\") on node \"crc\" DevicePath \"\"" Dec 07 16:13:46 crc kubenswrapper[4716]: I1207 16:13:46.818681 4716 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/cbf3f592-bbb1-4596-8ae1-0e62aa366f0b-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 07 16:13:46 crc kubenswrapper[4716]: I1207 16:13:46.919524 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/19363c75-987f-4c3b-bbca-28a9cc33238b-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c2vdd\" (UID: \"19363c75-987f-4c3b-bbca-28a9cc33238b\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c2vdd" Dec 07 16:13:46 crc kubenswrapper[4716]: I1207 16:13:46.919892 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4blr\" (UniqueName: \"kubernetes.io/projected/19363c75-987f-4c3b-bbca-28a9cc33238b-kube-api-access-x4blr\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c2vdd\" (UID: \"19363c75-987f-4c3b-bbca-28a9cc33238b\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c2vdd" Dec 07 16:13:46 crc kubenswrapper[4716]: I1207 16:13:46.920060 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/19363c75-987f-4c3b-bbca-28a9cc33238b-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c2vdd\" (UID: \"19363c75-987f-4c3b-bbca-28a9cc33238b\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c2vdd" Dec 07 16:13:46 crc kubenswrapper[4716]: I1207 16:13:46.920163 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/19363c75-987f-4c3b-bbca-28a9cc33238b-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c2vdd\" (UID: \"19363c75-987f-4c3b-bbca-28a9cc33238b\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c2vdd" Dec 07 16:13:46 crc kubenswrapper[4716]: I1207 16:13:46.920472 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/19363c75-987f-4c3b-bbca-28a9cc33238b-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c2vdd\" (UID: \"19363c75-987f-4c3b-bbca-28a9cc33238b\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c2vdd" Dec 07 16:13:46 crc kubenswrapper[4716]: I1207 16:13:46.935981 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4blr\" (UniqueName: \"kubernetes.io/projected/19363c75-987f-4c3b-bbca-28a9cc33238b-kube-api-access-x4blr\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c2vdd\" (UID: \"19363c75-987f-4c3b-bbca-28a9cc33238b\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c2vdd" Dec 07 16:13:47 crc kubenswrapper[4716]: I1207 16:13:47.111113 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c2vdd" Dec 07 16:13:47 crc kubenswrapper[4716]: I1207 16:13:47.224741 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-ndp5v_cbf3f592-bbb1-4596-8ae1-0e62aa366f0b/console/0.log" Dec 07 16:13:47 crc kubenswrapper[4716]: I1207 16:13:47.224810 4716 generic.go:334] "Generic (PLEG): container finished" podID="cbf3f592-bbb1-4596-8ae1-0e62aa366f0b" containerID="98b5a90edae2b9e0c3df9cc09a9dd99f59934f6e38e7da5499cddff1743f5663" exitCode=2 Dec 07 16:13:47 crc kubenswrapper[4716]: I1207 16:13:47.224853 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-ndp5v" event={"ID":"cbf3f592-bbb1-4596-8ae1-0e62aa366f0b","Type":"ContainerDied","Data":"98b5a90edae2b9e0c3df9cc09a9dd99f59934f6e38e7da5499cddff1743f5663"} Dec 07 16:13:47 crc kubenswrapper[4716]: I1207 16:13:47.224886 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-ndp5v" Dec 07 16:13:47 crc kubenswrapper[4716]: I1207 16:13:47.224907 4716 scope.go:117] "RemoveContainer" containerID="98b5a90edae2b9e0c3df9cc09a9dd99f59934f6e38e7da5499cddff1743f5663" Dec 07 16:13:47 crc kubenswrapper[4716]: I1207 16:13:47.224890 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-ndp5v" event={"ID":"cbf3f592-bbb1-4596-8ae1-0e62aa366f0b","Type":"ContainerDied","Data":"af061ef8ee4d7573b0865c8a6560467f821a1a39b9e853cda55704fe1185a924"} Dec 07 16:13:47 crc kubenswrapper[4716]: I1207 16:13:47.256806 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-ndp5v"] Dec 07 16:13:47 crc kubenswrapper[4716]: I1207 16:13:47.261977 4716 scope.go:117] "RemoveContainer" containerID="98b5a90edae2b9e0c3df9cc09a9dd99f59934f6e38e7da5499cddff1743f5663" Dec 07 16:13:47 crc kubenswrapper[4716]: E1207 16:13:47.262564 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98b5a90edae2b9e0c3df9cc09a9dd99f59934f6e38e7da5499cddff1743f5663\": container with ID starting with 98b5a90edae2b9e0c3df9cc09a9dd99f59934f6e38e7da5499cddff1743f5663 not found: ID does not exist" containerID="98b5a90edae2b9e0c3df9cc09a9dd99f59934f6e38e7da5499cddff1743f5663" Dec 07 16:13:47 crc kubenswrapper[4716]: I1207 16:13:47.262649 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98b5a90edae2b9e0c3df9cc09a9dd99f59934f6e38e7da5499cddff1743f5663"} err="failed to get container status \"98b5a90edae2b9e0c3df9cc09a9dd99f59934f6e38e7da5499cddff1743f5663\": rpc error: code = NotFound desc = could not find container \"98b5a90edae2b9e0c3df9cc09a9dd99f59934f6e38e7da5499cddff1743f5663\": container with ID starting with 98b5a90edae2b9e0c3df9cc09a9dd99f59934f6e38e7da5499cddff1743f5663 not found: ID does not exist" Dec 07 16:13:47 crc kubenswrapper[4716]: I1207 16:13:47.262574 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-ndp5v"] Dec 07 16:13:47 crc kubenswrapper[4716]: I1207 16:13:47.362303 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c2vdd"] Dec 07 16:13:47 crc kubenswrapper[4716]: W1207 16:13:47.369889 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod19363c75_987f_4c3b_bbca_28a9cc33238b.slice/crio-55024c17f53534613ae015e324f12dda58772a618895bf67fec66dc08369ec82 WatchSource:0}: Error finding container 55024c17f53534613ae015e324f12dda58772a618895bf67fec66dc08369ec82: Status 404 returned error can't find the container with id 55024c17f53534613ae015e324f12dda58772a618895bf67fec66dc08369ec82 Dec 07 16:13:47 crc kubenswrapper[4716]: I1207 16:13:47.673688 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cbf3f592-bbb1-4596-8ae1-0e62aa366f0b" path="/var/lib/kubelet/pods/cbf3f592-bbb1-4596-8ae1-0e62aa366f0b/volumes" Dec 07 16:13:48 crc kubenswrapper[4716]: I1207 16:13:48.239210 4716 generic.go:334] "Generic (PLEG): container finished" podID="19363c75-987f-4c3b-bbca-28a9cc33238b" containerID="f739a1fd11820bdfc6cdac53a55527823fdb507154fa62b50c51920b6a7140ba" exitCode=0 Dec 07 16:13:48 crc kubenswrapper[4716]: I1207 16:13:48.239279 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c2vdd" event={"ID":"19363c75-987f-4c3b-bbca-28a9cc33238b","Type":"ContainerDied","Data":"f739a1fd11820bdfc6cdac53a55527823fdb507154fa62b50c51920b6a7140ba"} Dec 07 16:13:48 crc kubenswrapper[4716]: I1207 16:13:48.239324 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c2vdd" event={"ID":"19363c75-987f-4c3b-bbca-28a9cc33238b","Type":"ContainerStarted","Data":"55024c17f53534613ae015e324f12dda58772a618895bf67fec66dc08369ec82"} Dec 07 16:13:50 crc kubenswrapper[4716]: I1207 16:13:50.250371 4716 generic.go:334] "Generic (PLEG): container finished" podID="19363c75-987f-4c3b-bbca-28a9cc33238b" containerID="5efde99e85d899545793fab17ffb0965d291485ba6a0b4cdaf25df9db4c1f2fe" exitCode=0 Dec 07 16:13:50 crc kubenswrapper[4716]: I1207 16:13:50.250462 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c2vdd" event={"ID":"19363c75-987f-4c3b-bbca-28a9cc33238b","Type":"ContainerDied","Data":"5efde99e85d899545793fab17ffb0965d291485ba6a0b4cdaf25df9db4c1f2fe"} Dec 07 16:13:51 crc kubenswrapper[4716]: I1207 16:13:51.259755 4716 generic.go:334] "Generic (PLEG): container finished" podID="19363c75-987f-4c3b-bbca-28a9cc33238b" containerID="11422b2389463a9124312a69a6542e34e77cb51bfb3abaa4ea47e58bba6d57ea" exitCode=0 Dec 07 16:13:51 crc kubenswrapper[4716]: I1207 16:13:51.259925 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c2vdd" event={"ID":"19363c75-987f-4c3b-bbca-28a9cc33238b","Type":"ContainerDied","Data":"11422b2389463a9124312a69a6542e34e77cb51bfb3abaa4ea47e58bba6d57ea"} Dec 07 16:13:52 crc kubenswrapper[4716]: I1207 16:13:52.507259 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c2vdd" Dec 07 16:13:52 crc kubenswrapper[4716]: I1207 16:13:52.692529 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/19363c75-987f-4c3b-bbca-28a9cc33238b-util\") pod \"19363c75-987f-4c3b-bbca-28a9cc33238b\" (UID: \"19363c75-987f-4c3b-bbca-28a9cc33238b\") " Dec 07 16:13:52 crc kubenswrapper[4716]: I1207 16:13:52.692628 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4blr\" (UniqueName: \"kubernetes.io/projected/19363c75-987f-4c3b-bbca-28a9cc33238b-kube-api-access-x4blr\") pod \"19363c75-987f-4c3b-bbca-28a9cc33238b\" (UID: \"19363c75-987f-4c3b-bbca-28a9cc33238b\") " Dec 07 16:13:52 crc kubenswrapper[4716]: I1207 16:13:52.692681 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/19363c75-987f-4c3b-bbca-28a9cc33238b-bundle\") pod \"19363c75-987f-4c3b-bbca-28a9cc33238b\" (UID: \"19363c75-987f-4c3b-bbca-28a9cc33238b\") " Dec 07 16:13:52 crc kubenswrapper[4716]: I1207 16:13:52.694101 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/19363c75-987f-4c3b-bbca-28a9cc33238b-bundle" (OuterVolumeSpecName: "bundle") pod "19363c75-987f-4c3b-bbca-28a9cc33238b" (UID: "19363c75-987f-4c3b-bbca-28a9cc33238b"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:13:52 crc kubenswrapper[4716]: I1207 16:13:52.698677 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19363c75-987f-4c3b-bbca-28a9cc33238b-kube-api-access-x4blr" (OuterVolumeSpecName: "kube-api-access-x4blr") pod "19363c75-987f-4c3b-bbca-28a9cc33238b" (UID: "19363c75-987f-4c3b-bbca-28a9cc33238b"). InnerVolumeSpecName "kube-api-access-x4blr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:13:52 crc kubenswrapper[4716]: I1207 16:13:52.719315 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/19363c75-987f-4c3b-bbca-28a9cc33238b-util" (OuterVolumeSpecName: "util") pod "19363c75-987f-4c3b-bbca-28a9cc33238b" (UID: "19363c75-987f-4c3b-bbca-28a9cc33238b"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:13:52 crc kubenswrapper[4716]: I1207 16:13:52.794212 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4blr\" (UniqueName: \"kubernetes.io/projected/19363c75-987f-4c3b-bbca-28a9cc33238b-kube-api-access-x4blr\") on node \"crc\" DevicePath \"\"" Dec 07 16:13:52 crc kubenswrapper[4716]: I1207 16:13:52.794541 4716 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/19363c75-987f-4c3b-bbca-28a9cc33238b-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 16:13:52 crc kubenswrapper[4716]: I1207 16:13:52.795315 4716 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/19363c75-987f-4c3b-bbca-28a9cc33238b-util\") on node \"crc\" DevicePath \"\"" Dec 07 16:13:53 crc kubenswrapper[4716]: I1207 16:13:53.272365 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c2vdd" event={"ID":"19363c75-987f-4c3b-bbca-28a9cc33238b","Type":"ContainerDied","Data":"55024c17f53534613ae015e324f12dda58772a618895bf67fec66dc08369ec82"} Dec 07 16:13:53 crc kubenswrapper[4716]: I1207 16:13:53.272412 4716 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="55024c17f53534613ae015e324f12dda58772a618895bf67fec66dc08369ec82" Dec 07 16:13:53 crc kubenswrapper[4716]: I1207 16:13:53.272411 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c2vdd" Dec 07 16:14:05 crc kubenswrapper[4716]: I1207 16:14:05.230112 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-655c466c6d-8pvn6"] Dec 07 16:14:05 crc kubenswrapper[4716]: E1207 16:14:05.231052 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19363c75-987f-4c3b-bbca-28a9cc33238b" containerName="pull" Dec 07 16:14:05 crc kubenswrapper[4716]: I1207 16:14:05.231067 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="19363c75-987f-4c3b-bbca-28a9cc33238b" containerName="pull" Dec 07 16:14:05 crc kubenswrapper[4716]: E1207 16:14:05.231096 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19363c75-987f-4c3b-bbca-28a9cc33238b" containerName="util" Dec 07 16:14:05 crc kubenswrapper[4716]: I1207 16:14:05.231102 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="19363c75-987f-4c3b-bbca-28a9cc33238b" containerName="util" Dec 07 16:14:05 crc kubenswrapper[4716]: E1207 16:14:05.231124 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19363c75-987f-4c3b-bbca-28a9cc33238b" containerName="extract" Dec 07 16:14:05 crc kubenswrapper[4716]: I1207 16:14:05.231130 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="19363c75-987f-4c3b-bbca-28a9cc33238b" containerName="extract" Dec 07 16:14:05 crc kubenswrapper[4716]: I1207 16:14:05.231241 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="19363c75-987f-4c3b-bbca-28a9cc33238b" containerName="extract" Dec 07 16:14:05 crc kubenswrapper[4716]: I1207 16:14:05.231739 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-655c466c6d-8pvn6" Dec 07 16:14:05 crc kubenswrapper[4716]: I1207 16:14:05.234335 4716 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Dec 07 16:14:05 crc kubenswrapper[4716]: I1207 16:14:05.234536 4716 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-57v4z" Dec 07 16:14:05 crc kubenswrapper[4716]: I1207 16:14:05.234661 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Dec 07 16:14:05 crc kubenswrapper[4716]: I1207 16:14:05.234838 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Dec 07 16:14:05 crc kubenswrapper[4716]: I1207 16:14:05.237655 4716 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Dec 07 16:14:05 crc kubenswrapper[4716]: I1207 16:14:05.246316 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-655c466c6d-8pvn6"] Dec 07 16:14:05 crc kubenswrapper[4716]: I1207 16:14:05.354159 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/049c277a-93fc-42db-952b-b1549f6a599c-apiservice-cert\") pod \"metallb-operator-controller-manager-655c466c6d-8pvn6\" (UID: \"049c277a-93fc-42db-952b-b1549f6a599c\") " pod="metallb-system/metallb-operator-controller-manager-655c466c6d-8pvn6" Dec 07 16:14:05 crc kubenswrapper[4716]: I1207 16:14:05.354221 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/049c277a-93fc-42db-952b-b1549f6a599c-webhook-cert\") pod \"metallb-operator-controller-manager-655c466c6d-8pvn6\" (UID: \"049c277a-93fc-42db-952b-b1549f6a599c\") " pod="metallb-system/metallb-operator-controller-manager-655c466c6d-8pvn6" Dec 07 16:14:05 crc kubenswrapper[4716]: I1207 16:14:05.354696 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2mrf8\" (UniqueName: \"kubernetes.io/projected/049c277a-93fc-42db-952b-b1549f6a599c-kube-api-access-2mrf8\") pod \"metallb-operator-controller-manager-655c466c6d-8pvn6\" (UID: \"049c277a-93fc-42db-952b-b1549f6a599c\") " pod="metallb-system/metallb-operator-controller-manager-655c466c6d-8pvn6" Dec 07 16:14:05 crc kubenswrapper[4716]: I1207 16:14:05.456163 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2mrf8\" (UniqueName: \"kubernetes.io/projected/049c277a-93fc-42db-952b-b1549f6a599c-kube-api-access-2mrf8\") pod \"metallb-operator-controller-manager-655c466c6d-8pvn6\" (UID: \"049c277a-93fc-42db-952b-b1549f6a599c\") " pod="metallb-system/metallb-operator-controller-manager-655c466c6d-8pvn6" Dec 07 16:14:05 crc kubenswrapper[4716]: I1207 16:14:05.456254 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/049c277a-93fc-42db-952b-b1549f6a599c-apiservice-cert\") pod \"metallb-operator-controller-manager-655c466c6d-8pvn6\" (UID: \"049c277a-93fc-42db-952b-b1549f6a599c\") " pod="metallb-system/metallb-operator-controller-manager-655c466c6d-8pvn6" Dec 07 16:14:05 crc kubenswrapper[4716]: I1207 16:14:05.456280 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/049c277a-93fc-42db-952b-b1549f6a599c-webhook-cert\") pod \"metallb-operator-controller-manager-655c466c6d-8pvn6\" (UID: \"049c277a-93fc-42db-952b-b1549f6a599c\") " pod="metallb-system/metallb-operator-controller-manager-655c466c6d-8pvn6" Dec 07 16:14:05 crc kubenswrapper[4716]: I1207 16:14:05.462803 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/049c277a-93fc-42db-952b-b1549f6a599c-apiservice-cert\") pod \"metallb-operator-controller-manager-655c466c6d-8pvn6\" (UID: \"049c277a-93fc-42db-952b-b1549f6a599c\") " pod="metallb-system/metallb-operator-controller-manager-655c466c6d-8pvn6" Dec 07 16:14:05 crc kubenswrapper[4716]: I1207 16:14:05.473205 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/049c277a-93fc-42db-952b-b1549f6a599c-webhook-cert\") pod \"metallb-operator-controller-manager-655c466c6d-8pvn6\" (UID: \"049c277a-93fc-42db-952b-b1549f6a599c\") " pod="metallb-system/metallb-operator-controller-manager-655c466c6d-8pvn6" Dec 07 16:14:05 crc kubenswrapper[4716]: I1207 16:14:05.484959 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2mrf8\" (UniqueName: \"kubernetes.io/projected/049c277a-93fc-42db-952b-b1549f6a599c-kube-api-access-2mrf8\") pod \"metallb-operator-controller-manager-655c466c6d-8pvn6\" (UID: \"049c277a-93fc-42db-952b-b1549f6a599c\") " pod="metallb-system/metallb-operator-controller-manager-655c466c6d-8pvn6" Dec 07 16:14:05 crc kubenswrapper[4716]: I1207 16:14:05.489711 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-65fbf78689-nbmlk"] Dec 07 16:14:05 crc kubenswrapper[4716]: I1207 16:14:05.490584 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-65fbf78689-nbmlk" Dec 07 16:14:05 crc kubenswrapper[4716]: I1207 16:14:05.493176 4716 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 07 16:14:05 crc kubenswrapper[4716]: I1207 16:14:05.493428 4716 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Dec 07 16:14:05 crc kubenswrapper[4716]: I1207 16:14:05.493439 4716 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-8xh8r" Dec 07 16:14:05 crc kubenswrapper[4716]: I1207 16:14:05.501531 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-65fbf78689-nbmlk"] Dec 07 16:14:05 crc kubenswrapper[4716]: I1207 16:14:05.549912 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-655c466c6d-8pvn6" Dec 07 16:14:05 crc kubenswrapper[4716]: I1207 16:14:05.556979 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t7dl2\" (UniqueName: \"kubernetes.io/projected/b6bf63fb-cc3e-4349-b7f6-efb11c1ff14f-kube-api-access-t7dl2\") pod \"metallb-operator-webhook-server-65fbf78689-nbmlk\" (UID: \"b6bf63fb-cc3e-4349-b7f6-efb11c1ff14f\") " pod="metallb-system/metallb-operator-webhook-server-65fbf78689-nbmlk" Dec 07 16:14:05 crc kubenswrapper[4716]: I1207 16:14:05.557277 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b6bf63fb-cc3e-4349-b7f6-efb11c1ff14f-webhook-cert\") pod \"metallb-operator-webhook-server-65fbf78689-nbmlk\" (UID: \"b6bf63fb-cc3e-4349-b7f6-efb11c1ff14f\") " pod="metallb-system/metallb-operator-webhook-server-65fbf78689-nbmlk" Dec 07 16:14:05 crc kubenswrapper[4716]: I1207 16:14:05.557329 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b6bf63fb-cc3e-4349-b7f6-efb11c1ff14f-apiservice-cert\") pod \"metallb-operator-webhook-server-65fbf78689-nbmlk\" (UID: \"b6bf63fb-cc3e-4349-b7f6-efb11c1ff14f\") " pod="metallb-system/metallb-operator-webhook-server-65fbf78689-nbmlk" Dec 07 16:14:05 crc kubenswrapper[4716]: I1207 16:14:05.657861 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t7dl2\" (UniqueName: \"kubernetes.io/projected/b6bf63fb-cc3e-4349-b7f6-efb11c1ff14f-kube-api-access-t7dl2\") pod \"metallb-operator-webhook-server-65fbf78689-nbmlk\" (UID: \"b6bf63fb-cc3e-4349-b7f6-efb11c1ff14f\") " pod="metallb-system/metallb-operator-webhook-server-65fbf78689-nbmlk" Dec 07 16:14:05 crc kubenswrapper[4716]: I1207 16:14:05.658104 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b6bf63fb-cc3e-4349-b7f6-efb11c1ff14f-webhook-cert\") pod \"metallb-operator-webhook-server-65fbf78689-nbmlk\" (UID: \"b6bf63fb-cc3e-4349-b7f6-efb11c1ff14f\") " pod="metallb-system/metallb-operator-webhook-server-65fbf78689-nbmlk" Dec 07 16:14:05 crc kubenswrapper[4716]: I1207 16:14:05.658236 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b6bf63fb-cc3e-4349-b7f6-efb11c1ff14f-apiservice-cert\") pod \"metallb-operator-webhook-server-65fbf78689-nbmlk\" (UID: \"b6bf63fb-cc3e-4349-b7f6-efb11c1ff14f\") " pod="metallb-system/metallb-operator-webhook-server-65fbf78689-nbmlk" Dec 07 16:14:05 crc kubenswrapper[4716]: I1207 16:14:05.662432 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b6bf63fb-cc3e-4349-b7f6-efb11c1ff14f-webhook-cert\") pod \"metallb-operator-webhook-server-65fbf78689-nbmlk\" (UID: \"b6bf63fb-cc3e-4349-b7f6-efb11c1ff14f\") " pod="metallb-system/metallb-operator-webhook-server-65fbf78689-nbmlk" Dec 07 16:14:05 crc kubenswrapper[4716]: I1207 16:14:05.666868 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b6bf63fb-cc3e-4349-b7f6-efb11c1ff14f-apiservice-cert\") pod \"metallb-operator-webhook-server-65fbf78689-nbmlk\" (UID: \"b6bf63fb-cc3e-4349-b7f6-efb11c1ff14f\") " pod="metallb-system/metallb-operator-webhook-server-65fbf78689-nbmlk" Dec 07 16:14:05 crc kubenswrapper[4716]: I1207 16:14:05.689029 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t7dl2\" (UniqueName: \"kubernetes.io/projected/b6bf63fb-cc3e-4349-b7f6-efb11c1ff14f-kube-api-access-t7dl2\") pod \"metallb-operator-webhook-server-65fbf78689-nbmlk\" (UID: \"b6bf63fb-cc3e-4349-b7f6-efb11c1ff14f\") " pod="metallb-system/metallb-operator-webhook-server-65fbf78689-nbmlk" Dec 07 16:14:05 crc kubenswrapper[4716]: I1207 16:14:05.826684 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-655c466c6d-8pvn6"] Dec 07 16:14:05 crc kubenswrapper[4716]: I1207 16:14:05.856473 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-65fbf78689-nbmlk" Dec 07 16:14:06 crc kubenswrapper[4716]: I1207 16:14:06.049149 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-65fbf78689-nbmlk"] Dec 07 16:14:06 crc kubenswrapper[4716]: W1207 16:14:06.053966 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb6bf63fb_cc3e_4349_b7f6_efb11c1ff14f.slice/crio-b1ac92a9ee6cb2b7d60353b623b93d46b8c9357139c83c4c38512e1d5177ea4e WatchSource:0}: Error finding container b1ac92a9ee6cb2b7d60353b623b93d46b8c9357139c83c4c38512e1d5177ea4e: Status 404 returned error can't find the container with id b1ac92a9ee6cb2b7d60353b623b93d46b8c9357139c83c4c38512e1d5177ea4e Dec 07 16:14:06 crc kubenswrapper[4716]: I1207 16:14:06.344725 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-65fbf78689-nbmlk" event={"ID":"b6bf63fb-cc3e-4349-b7f6-efb11c1ff14f","Type":"ContainerStarted","Data":"b1ac92a9ee6cb2b7d60353b623b93d46b8c9357139c83c4c38512e1d5177ea4e"} Dec 07 16:14:06 crc kubenswrapper[4716]: I1207 16:14:06.346192 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-655c466c6d-8pvn6" event={"ID":"049c277a-93fc-42db-952b-b1549f6a599c","Type":"ContainerStarted","Data":"c3c305c17165989bb49824226677f635d90deaf406da0201e454a7308ab6bfd4"} Dec 07 16:14:09 crc kubenswrapper[4716]: I1207 16:14:09.367972 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-655c466c6d-8pvn6" event={"ID":"049c277a-93fc-42db-952b-b1549f6a599c","Type":"ContainerStarted","Data":"308c328c1b03d57dd0ecc79152b0f238547114cc595a838838ac662cc8929518"} Dec 07 16:14:09 crc kubenswrapper[4716]: I1207 16:14:09.368550 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-655c466c6d-8pvn6" Dec 07 16:14:09 crc kubenswrapper[4716]: I1207 16:14:09.389418 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-655c466c6d-8pvn6" podStartSLOduration=1.564420217 podStartE2EDuration="4.389399884s" podCreationTimestamp="2025-12-07 16:14:05 +0000 UTC" firstStartedPulling="2025-12-07 16:14:05.839117338 +0000 UTC m=+708.529402250" lastFinishedPulling="2025-12-07 16:14:08.664097005 +0000 UTC m=+711.354381917" observedRunningTime="2025-12-07 16:14:09.387165145 +0000 UTC m=+712.077450057" watchObservedRunningTime="2025-12-07 16:14:09.389399884 +0000 UTC m=+712.079684796" Dec 07 16:14:11 crc kubenswrapper[4716]: I1207 16:14:11.377660 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-65fbf78689-nbmlk" event={"ID":"b6bf63fb-cc3e-4349-b7f6-efb11c1ff14f","Type":"ContainerStarted","Data":"89b954c913eb26396894e06a6415246f9780f1f625885998a7d1ec7b680a18e6"} Dec 07 16:14:11 crc kubenswrapper[4716]: I1207 16:14:11.377982 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-65fbf78689-nbmlk" Dec 07 16:14:11 crc kubenswrapper[4716]: I1207 16:14:11.396997 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-65fbf78689-nbmlk" podStartSLOduration=1.9660724630000002 podStartE2EDuration="6.396984251s" podCreationTimestamp="2025-12-07 16:14:05 +0000 UTC" firstStartedPulling="2025-12-07 16:14:06.057442754 +0000 UTC m=+708.747727666" lastFinishedPulling="2025-12-07 16:14:10.488354522 +0000 UTC m=+713.178639454" observedRunningTime="2025-12-07 16:14:11.393854787 +0000 UTC m=+714.084139689" watchObservedRunningTime="2025-12-07 16:14:11.396984251 +0000 UTC m=+714.087269163" Dec 07 16:14:25 crc kubenswrapper[4716]: I1207 16:14:25.861370 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-65fbf78689-nbmlk" Dec 07 16:14:45 crc kubenswrapper[4716]: I1207 16:14:45.552497 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-655c466c6d-8pvn6" Dec 07 16:14:46 crc kubenswrapper[4716]: I1207 16:14:46.025484 4716 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 07 16:14:46 crc kubenswrapper[4716]: I1207 16:14:46.283386 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-dqzl5"] Dec 07 16:14:46 crc kubenswrapper[4716]: I1207 16:14:46.284247 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-dqzl5" Dec 07 16:14:46 crc kubenswrapper[4716]: I1207 16:14:46.291687 4716 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-d7flh" Dec 07 16:14:46 crc kubenswrapper[4716]: I1207 16:14:46.291869 4716 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Dec 07 16:14:46 crc kubenswrapper[4716]: I1207 16:14:46.292877 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-9zdjv"] Dec 07 16:14:46 crc kubenswrapper[4716]: I1207 16:14:46.294965 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-9zdjv" Dec 07 16:14:46 crc kubenswrapper[4716]: I1207 16:14:46.298040 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Dec 07 16:14:46 crc kubenswrapper[4716]: I1207 16:14:46.298240 4716 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Dec 07 16:14:46 crc kubenswrapper[4716]: I1207 16:14:46.303778 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-dqzl5"] Dec 07 16:14:46 crc kubenswrapper[4716]: I1207 16:14:46.362123 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-xnsb6"] Dec 07 16:14:46 crc kubenswrapper[4716]: I1207 16:14:46.363227 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-xnsb6" Dec 07 16:14:46 crc kubenswrapper[4716]: I1207 16:14:46.365025 4716 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-556tb" Dec 07 16:14:46 crc kubenswrapper[4716]: I1207 16:14:46.365211 4716 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Dec 07 16:14:46 crc kubenswrapper[4716]: I1207 16:14:46.365475 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Dec 07 16:14:46 crc kubenswrapper[4716]: I1207 16:14:46.366711 4716 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Dec 07 16:14:46 crc kubenswrapper[4716]: I1207 16:14:46.394339 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/82d3c78d-32ff-47ed-854a-d3786c962ffd-memberlist\") pod \"speaker-xnsb6\" (UID: \"82d3c78d-32ff-47ed-854a-d3786c962ffd\") " pod="metallb-system/speaker-xnsb6" Dec 07 16:14:46 crc kubenswrapper[4716]: I1207 16:14:46.394385 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vnkrf\" (UniqueName: \"kubernetes.io/projected/5e169ec0-a81a-45eb-b2a6-8ac97d905caa-kube-api-access-vnkrf\") pod \"frr-k8s-9zdjv\" (UID: \"5e169ec0-a81a-45eb-b2a6-8ac97d905caa\") " pod="metallb-system/frr-k8s-9zdjv" Dec 07 16:14:46 crc kubenswrapper[4716]: I1207 16:14:46.394405 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/5e169ec0-a81a-45eb-b2a6-8ac97d905caa-frr-sockets\") pod \"frr-k8s-9zdjv\" (UID: \"5e169ec0-a81a-45eb-b2a6-8ac97d905caa\") " pod="metallb-system/frr-k8s-9zdjv" Dec 07 16:14:46 crc kubenswrapper[4716]: I1207 16:14:46.394422 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/82d3c78d-32ff-47ed-854a-d3786c962ffd-metallb-excludel2\") pod \"speaker-xnsb6\" (UID: \"82d3c78d-32ff-47ed-854a-d3786c962ffd\") " pod="metallb-system/speaker-xnsb6" Dec 07 16:14:46 crc kubenswrapper[4716]: I1207 16:14:46.394439 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/5e169ec0-a81a-45eb-b2a6-8ac97d905caa-reloader\") pod \"frr-k8s-9zdjv\" (UID: \"5e169ec0-a81a-45eb-b2a6-8ac97d905caa\") " pod="metallb-system/frr-k8s-9zdjv" Dec 07 16:14:46 crc kubenswrapper[4716]: I1207 16:14:46.394458 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/5e169ec0-a81a-45eb-b2a6-8ac97d905caa-metrics\") pod \"frr-k8s-9zdjv\" (UID: \"5e169ec0-a81a-45eb-b2a6-8ac97d905caa\") " pod="metallb-system/frr-k8s-9zdjv" Dec 07 16:14:46 crc kubenswrapper[4716]: I1207 16:14:46.394473 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-shmkd\" (UniqueName: \"kubernetes.io/projected/1a4973da-40f0-4277-a6ff-06ff961b4815-kube-api-access-shmkd\") pod \"frr-k8s-webhook-server-7fcb986d4-dqzl5\" (UID: \"1a4973da-40f0-4277-a6ff-06ff961b4815\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-dqzl5" Dec 07 16:14:46 crc kubenswrapper[4716]: I1207 16:14:46.394497 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1a4973da-40f0-4277-a6ff-06ff961b4815-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-dqzl5\" (UID: \"1a4973da-40f0-4277-a6ff-06ff961b4815\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-dqzl5" Dec 07 16:14:46 crc kubenswrapper[4716]: I1207 16:14:46.394523 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/5e169ec0-a81a-45eb-b2a6-8ac97d905caa-frr-conf\") pod \"frr-k8s-9zdjv\" (UID: \"5e169ec0-a81a-45eb-b2a6-8ac97d905caa\") " pod="metallb-system/frr-k8s-9zdjv" Dec 07 16:14:46 crc kubenswrapper[4716]: I1207 16:14:46.394539 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tnr9q\" (UniqueName: \"kubernetes.io/projected/82d3c78d-32ff-47ed-854a-d3786c962ffd-kube-api-access-tnr9q\") pod \"speaker-xnsb6\" (UID: \"82d3c78d-32ff-47ed-854a-d3786c962ffd\") " pod="metallb-system/speaker-xnsb6" Dec 07 16:14:46 crc kubenswrapper[4716]: I1207 16:14:46.394565 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/5e169ec0-a81a-45eb-b2a6-8ac97d905caa-frr-startup\") pod \"frr-k8s-9zdjv\" (UID: \"5e169ec0-a81a-45eb-b2a6-8ac97d905caa\") " pod="metallb-system/frr-k8s-9zdjv" Dec 07 16:14:46 crc kubenswrapper[4716]: I1207 16:14:46.394580 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5e169ec0-a81a-45eb-b2a6-8ac97d905caa-metrics-certs\") pod \"frr-k8s-9zdjv\" (UID: \"5e169ec0-a81a-45eb-b2a6-8ac97d905caa\") " pod="metallb-system/frr-k8s-9zdjv" Dec 07 16:14:46 crc kubenswrapper[4716]: I1207 16:14:46.394606 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/82d3c78d-32ff-47ed-854a-d3786c962ffd-metrics-certs\") pod \"speaker-xnsb6\" (UID: \"82d3c78d-32ff-47ed-854a-d3786c962ffd\") " pod="metallb-system/speaker-xnsb6" Dec 07 16:14:46 crc kubenswrapper[4716]: I1207 16:14:46.394617 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-f8648f98b-t728x"] Dec 07 16:14:46 crc kubenswrapper[4716]: I1207 16:14:46.395674 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-t728x" Dec 07 16:14:46 crc kubenswrapper[4716]: I1207 16:14:46.398408 4716 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Dec 07 16:14:46 crc kubenswrapper[4716]: I1207 16:14:46.414025 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-t728x"] Dec 07 16:14:46 crc kubenswrapper[4716]: I1207 16:14:46.495402 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/5e169ec0-a81a-45eb-b2a6-8ac97d905caa-frr-startup\") pod \"frr-k8s-9zdjv\" (UID: \"5e169ec0-a81a-45eb-b2a6-8ac97d905caa\") " pod="metallb-system/frr-k8s-9zdjv" Dec 07 16:14:46 crc kubenswrapper[4716]: I1207 16:14:46.495447 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5e169ec0-a81a-45eb-b2a6-8ac97d905caa-metrics-certs\") pod \"frr-k8s-9zdjv\" (UID: \"5e169ec0-a81a-45eb-b2a6-8ac97d905caa\") " pod="metallb-system/frr-k8s-9zdjv" Dec 07 16:14:46 crc kubenswrapper[4716]: I1207 16:14:46.495485 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/82d3c78d-32ff-47ed-854a-d3786c962ffd-metrics-certs\") pod \"speaker-xnsb6\" (UID: \"82d3c78d-32ff-47ed-854a-d3786c962ffd\") " pod="metallb-system/speaker-xnsb6" Dec 07 16:14:46 crc kubenswrapper[4716]: I1207 16:14:46.495501 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/82d3c78d-32ff-47ed-854a-d3786c962ffd-memberlist\") pod \"speaker-xnsb6\" (UID: \"82d3c78d-32ff-47ed-854a-d3786c962ffd\") " pod="metallb-system/speaker-xnsb6" Dec 07 16:14:46 crc kubenswrapper[4716]: I1207 16:14:46.495522 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f233fcdd-5c35-4ffb-bb33-7e2249f1c6c0-cert\") pod \"controller-f8648f98b-t728x\" (UID: \"f233fcdd-5c35-4ffb-bb33-7e2249f1c6c0\") " pod="metallb-system/controller-f8648f98b-t728x" Dec 07 16:14:46 crc kubenswrapper[4716]: I1207 16:14:46.495542 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/5e169ec0-a81a-45eb-b2a6-8ac97d905caa-frr-sockets\") pod \"frr-k8s-9zdjv\" (UID: \"5e169ec0-a81a-45eb-b2a6-8ac97d905caa\") " pod="metallb-system/frr-k8s-9zdjv" Dec 07 16:14:46 crc kubenswrapper[4716]: I1207 16:14:46.495559 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vnkrf\" (UniqueName: \"kubernetes.io/projected/5e169ec0-a81a-45eb-b2a6-8ac97d905caa-kube-api-access-vnkrf\") pod \"frr-k8s-9zdjv\" (UID: \"5e169ec0-a81a-45eb-b2a6-8ac97d905caa\") " pod="metallb-system/frr-k8s-9zdjv" Dec 07 16:14:46 crc kubenswrapper[4716]: I1207 16:14:46.495575 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/82d3c78d-32ff-47ed-854a-d3786c962ffd-metallb-excludel2\") pod \"speaker-xnsb6\" (UID: \"82d3c78d-32ff-47ed-854a-d3786c962ffd\") " pod="metallb-system/speaker-xnsb6" Dec 07 16:14:46 crc kubenswrapper[4716]: I1207 16:14:46.495593 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/5e169ec0-a81a-45eb-b2a6-8ac97d905caa-reloader\") pod \"frr-k8s-9zdjv\" (UID: \"5e169ec0-a81a-45eb-b2a6-8ac97d905caa\") " pod="metallb-system/frr-k8s-9zdjv" Dec 07 16:14:46 crc kubenswrapper[4716]: I1207 16:14:46.495610 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-64gzx\" (UniqueName: \"kubernetes.io/projected/f233fcdd-5c35-4ffb-bb33-7e2249f1c6c0-kube-api-access-64gzx\") pod \"controller-f8648f98b-t728x\" (UID: \"f233fcdd-5c35-4ffb-bb33-7e2249f1c6c0\") " pod="metallb-system/controller-f8648f98b-t728x" Dec 07 16:14:46 crc kubenswrapper[4716]: I1207 16:14:46.495628 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/5e169ec0-a81a-45eb-b2a6-8ac97d905caa-metrics\") pod \"frr-k8s-9zdjv\" (UID: \"5e169ec0-a81a-45eb-b2a6-8ac97d905caa\") " pod="metallb-system/frr-k8s-9zdjv" Dec 07 16:14:46 crc kubenswrapper[4716]: I1207 16:14:46.495645 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-shmkd\" (UniqueName: \"kubernetes.io/projected/1a4973da-40f0-4277-a6ff-06ff961b4815-kube-api-access-shmkd\") pod \"frr-k8s-webhook-server-7fcb986d4-dqzl5\" (UID: \"1a4973da-40f0-4277-a6ff-06ff961b4815\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-dqzl5" Dec 07 16:14:46 crc kubenswrapper[4716]: I1207 16:14:46.495667 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f233fcdd-5c35-4ffb-bb33-7e2249f1c6c0-metrics-certs\") pod \"controller-f8648f98b-t728x\" (UID: \"f233fcdd-5c35-4ffb-bb33-7e2249f1c6c0\") " pod="metallb-system/controller-f8648f98b-t728x" Dec 07 16:14:46 crc kubenswrapper[4716]: I1207 16:14:46.495690 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1a4973da-40f0-4277-a6ff-06ff961b4815-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-dqzl5\" (UID: \"1a4973da-40f0-4277-a6ff-06ff961b4815\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-dqzl5" Dec 07 16:14:46 crc kubenswrapper[4716]: I1207 16:14:46.495718 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/5e169ec0-a81a-45eb-b2a6-8ac97d905caa-frr-conf\") pod \"frr-k8s-9zdjv\" (UID: \"5e169ec0-a81a-45eb-b2a6-8ac97d905caa\") " pod="metallb-system/frr-k8s-9zdjv" Dec 07 16:14:46 crc kubenswrapper[4716]: I1207 16:14:46.495734 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tnr9q\" (UniqueName: \"kubernetes.io/projected/82d3c78d-32ff-47ed-854a-d3786c962ffd-kube-api-access-tnr9q\") pod \"speaker-xnsb6\" (UID: \"82d3c78d-32ff-47ed-854a-d3786c962ffd\") " pod="metallb-system/speaker-xnsb6" Dec 07 16:14:46 crc kubenswrapper[4716]: E1207 16:14:46.496325 4716 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Dec 07 16:14:46 crc kubenswrapper[4716]: E1207 16:14:46.496404 4716 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 07 16:14:46 crc kubenswrapper[4716]: E1207 16:14:46.496410 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/82d3c78d-32ff-47ed-854a-d3786c962ffd-metrics-certs podName:82d3c78d-32ff-47ed-854a-d3786c962ffd nodeName:}" failed. No retries permitted until 2025-12-07 16:14:46.996386063 +0000 UTC m=+749.686671055 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/82d3c78d-32ff-47ed-854a-d3786c962ffd-metrics-certs") pod "speaker-xnsb6" (UID: "82d3c78d-32ff-47ed-854a-d3786c962ffd") : secret "speaker-certs-secret" not found Dec 07 16:14:46 crc kubenswrapper[4716]: E1207 16:14:46.496463 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/82d3c78d-32ff-47ed-854a-d3786c962ffd-memberlist podName:82d3c78d-32ff-47ed-854a-d3786c962ffd nodeName:}" failed. No retries permitted until 2025-12-07 16:14:46.996450235 +0000 UTC m=+749.686735147 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/82d3c78d-32ff-47ed-854a-d3786c962ffd-memberlist") pod "speaker-xnsb6" (UID: "82d3c78d-32ff-47ed-854a-d3786c962ffd") : secret "metallb-memberlist" not found Dec 07 16:14:46 crc kubenswrapper[4716]: I1207 16:14:46.496349 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/5e169ec0-a81a-45eb-b2a6-8ac97d905caa-frr-startup\") pod \"frr-k8s-9zdjv\" (UID: \"5e169ec0-a81a-45eb-b2a6-8ac97d905caa\") " pod="metallb-system/frr-k8s-9zdjv" Dec 07 16:14:46 crc kubenswrapper[4716]: I1207 16:14:46.496621 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/5e169ec0-a81a-45eb-b2a6-8ac97d905caa-reloader\") pod \"frr-k8s-9zdjv\" (UID: \"5e169ec0-a81a-45eb-b2a6-8ac97d905caa\") " pod="metallb-system/frr-k8s-9zdjv" Dec 07 16:14:46 crc kubenswrapper[4716]: I1207 16:14:46.496801 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/5e169ec0-a81a-45eb-b2a6-8ac97d905caa-frr-sockets\") pod \"frr-k8s-9zdjv\" (UID: \"5e169ec0-a81a-45eb-b2a6-8ac97d905caa\") " pod="metallb-system/frr-k8s-9zdjv" Dec 07 16:14:46 crc kubenswrapper[4716]: I1207 16:14:46.497028 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/5e169ec0-a81a-45eb-b2a6-8ac97d905caa-frr-conf\") pod \"frr-k8s-9zdjv\" (UID: \"5e169ec0-a81a-45eb-b2a6-8ac97d905caa\") " pod="metallb-system/frr-k8s-9zdjv" Dec 07 16:14:46 crc kubenswrapper[4716]: I1207 16:14:46.497229 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/82d3c78d-32ff-47ed-854a-d3786c962ffd-metallb-excludel2\") pod \"speaker-xnsb6\" (UID: \"82d3c78d-32ff-47ed-854a-d3786c962ffd\") " pod="metallb-system/speaker-xnsb6" Dec 07 16:14:46 crc kubenswrapper[4716]: I1207 16:14:46.497439 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/5e169ec0-a81a-45eb-b2a6-8ac97d905caa-metrics\") pod \"frr-k8s-9zdjv\" (UID: \"5e169ec0-a81a-45eb-b2a6-8ac97d905caa\") " pod="metallb-system/frr-k8s-9zdjv" Dec 07 16:14:46 crc kubenswrapper[4716]: I1207 16:14:46.503752 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5e169ec0-a81a-45eb-b2a6-8ac97d905caa-metrics-certs\") pod \"frr-k8s-9zdjv\" (UID: \"5e169ec0-a81a-45eb-b2a6-8ac97d905caa\") " pod="metallb-system/frr-k8s-9zdjv" Dec 07 16:14:46 crc kubenswrapper[4716]: I1207 16:14:46.506291 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1a4973da-40f0-4277-a6ff-06ff961b4815-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-dqzl5\" (UID: \"1a4973da-40f0-4277-a6ff-06ff961b4815\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-dqzl5" Dec 07 16:14:46 crc kubenswrapper[4716]: I1207 16:14:46.517174 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-shmkd\" (UniqueName: \"kubernetes.io/projected/1a4973da-40f0-4277-a6ff-06ff961b4815-kube-api-access-shmkd\") pod \"frr-k8s-webhook-server-7fcb986d4-dqzl5\" (UID: \"1a4973da-40f0-4277-a6ff-06ff961b4815\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-dqzl5" Dec 07 16:14:46 crc kubenswrapper[4716]: I1207 16:14:46.517740 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vnkrf\" (UniqueName: \"kubernetes.io/projected/5e169ec0-a81a-45eb-b2a6-8ac97d905caa-kube-api-access-vnkrf\") pod \"frr-k8s-9zdjv\" (UID: \"5e169ec0-a81a-45eb-b2a6-8ac97d905caa\") " pod="metallb-system/frr-k8s-9zdjv" Dec 07 16:14:46 crc kubenswrapper[4716]: I1207 16:14:46.519292 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tnr9q\" (UniqueName: \"kubernetes.io/projected/82d3c78d-32ff-47ed-854a-d3786c962ffd-kube-api-access-tnr9q\") pod \"speaker-xnsb6\" (UID: \"82d3c78d-32ff-47ed-854a-d3786c962ffd\") " pod="metallb-system/speaker-xnsb6" Dec 07 16:14:46 crc kubenswrapper[4716]: I1207 16:14:46.596162 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f233fcdd-5c35-4ffb-bb33-7e2249f1c6c0-cert\") pod \"controller-f8648f98b-t728x\" (UID: \"f233fcdd-5c35-4ffb-bb33-7e2249f1c6c0\") " pod="metallb-system/controller-f8648f98b-t728x" Dec 07 16:14:46 crc kubenswrapper[4716]: I1207 16:14:46.596756 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-64gzx\" (UniqueName: \"kubernetes.io/projected/f233fcdd-5c35-4ffb-bb33-7e2249f1c6c0-kube-api-access-64gzx\") pod \"controller-f8648f98b-t728x\" (UID: \"f233fcdd-5c35-4ffb-bb33-7e2249f1c6c0\") " pod="metallb-system/controller-f8648f98b-t728x" Dec 07 16:14:46 crc kubenswrapper[4716]: I1207 16:14:46.596806 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f233fcdd-5c35-4ffb-bb33-7e2249f1c6c0-metrics-certs\") pod \"controller-f8648f98b-t728x\" (UID: \"f233fcdd-5c35-4ffb-bb33-7e2249f1c6c0\") " pod="metallb-system/controller-f8648f98b-t728x" Dec 07 16:14:46 crc kubenswrapper[4716]: I1207 16:14:46.598737 4716 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 07 16:14:46 crc kubenswrapper[4716]: I1207 16:14:46.600922 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-dqzl5" Dec 07 16:14:46 crc kubenswrapper[4716]: I1207 16:14:46.607892 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f233fcdd-5c35-4ffb-bb33-7e2249f1c6c0-metrics-certs\") pod \"controller-f8648f98b-t728x\" (UID: \"f233fcdd-5c35-4ffb-bb33-7e2249f1c6c0\") " pod="metallb-system/controller-f8648f98b-t728x" Dec 07 16:14:46 crc kubenswrapper[4716]: I1207 16:14:46.609868 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f233fcdd-5c35-4ffb-bb33-7e2249f1c6c0-cert\") pod \"controller-f8648f98b-t728x\" (UID: \"f233fcdd-5c35-4ffb-bb33-7e2249f1c6c0\") " pod="metallb-system/controller-f8648f98b-t728x" Dec 07 16:14:46 crc kubenswrapper[4716]: I1207 16:14:46.612607 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-64gzx\" (UniqueName: \"kubernetes.io/projected/f233fcdd-5c35-4ffb-bb33-7e2249f1c6c0-kube-api-access-64gzx\") pod \"controller-f8648f98b-t728x\" (UID: \"f233fcdd-5c35-4ffb-bb33-7e2249f1c6c0\") " pod="metallb-system/controller-f8648f98b-t728x" Dec 07 16:14:46 crc kubenswrapper[4716]: I1207 16:14:46.619384 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-9zdjv" Dec 07 16:14:46 crc kubenswrapper[4716]: I1207 16:14:46.709545 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-t728x" Dec 07 16:14:46 crc kubenswrapper[4716]: I1207 16:14:46.993132 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-t728x"] Dec 07 16:14:47 crc kubenswrapper[4716]: I1207 16:14:47.004500 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/82d3c78d-32ff-47ed-854a-d3786c962ffd-metrics-certs\") pod \"speaker-xnsb6\" (UID: \"82d3c78d-32ff-47ed-854a-d3786c962ffd\") " pod="metallb-system/speaker-xnsb6" Dec 07 16:14:47 crc kubenswrapper[4716]: I1207 16:14:47.004543 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/82d3c78d-32ff-47ed-854a-d3786c962ffd-memberlist\") pod \"speaker-xnsb6\" (UID: \"82d3c78d-32ff-47ed-854a-d3786c962ffd\") " pod="metallb-system/speaker-xnsb6" Dec 07 16:14:47 crc kubenswrapper[4716]: E1207 16:14:47.004658 4716 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 07 16:14:47 crc kubenswrapper[4716]: E1207 16:14:47.004722 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/82d3c78d-32ff-47ed-854a-d3786c962ffd-memberlist podName:82d3c78d-32ff-47ed-854a-d3786c962ffd nodeName:}" failed. No retries permitted until 2025-12-07 16:14:48.004703569 +0000 UTC m=+750.694988481 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/82d3c78d-32ff-47ed-854a-d3786c962ffd-memberlist") pod "speaker-xnsb6" (UID: "82d3c78d-32ff-47ed-854a-d3786c962ffd") : secret "metallb-memberlist" not found Dec 07 16:14:47 crc kubenswrapper[4716]: I1207 16:14:47.010497 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/82d3c78d-32ff-47ed-854a-d3786c962ffd-metrics-certs\") pod \"speaker-xnsb6\" (UID: \"82d3c78d-32ff-47ed-854a-d3786c962ffd\") " pod="metallb-system/speaker-xnsb6" Dec 07 16:14:47 crc kubenswrapper[4716]: I1207 16:14:47.040164 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-dqzl5"] Dec 07 16:14:47 crc kubenswrapper[4716]: I1207 16:14:47.579702 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-dqzl5" event={"ID":"1a4973da-40f0-4277-a6ff-06ff961b4815","Type":"ContainerStarted","Data":"7e4c86dd8b9717df4fbbe7ae95d2cee3d165262aff92c9d9d0b1d8b0977a90a8"} Dec 07 16:14:47 crc kubenswrapper[4716]: I1207 16:14:47.580887 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-9zdjv" event={"ID":"5e169ec0-a81a-45eb-b2a6-8ac97d905caa","Type":"ContainerStarted","Data":"678fa2423116ec7e12a04ce074019e4cf13c4ce4746f849329b3627141014d45"} Dec 07 16:14:47 crc kubenswrapper[4716]: I1207 16:14:47.582702 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-t728x" event={"ID":"f233fcdd-5c35-4ffb-bb33-7e2249f1c6c0","Type":"ContainerStarted","Data":"fb3805041672741d23a8fe98df9dc820194d0475118092b7dae26326b8ad1b90"} Dec 07 16:14:47 crc kubenswrapper[4716]: I1207 16:14:47.582724 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-t728x" event={"ID":"f233fcdd-5c35-4ffb-bb33-7e2249f1c6c0","Type":"ContainerStarted","Data":"178ef532ffa177a2ecdc1e541035f4802719865246488860b607b1c31bce1130"} Dec 07 16:14:47 crc kubenswrapper[4716]: I1207 16:14:47.582733 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-t728x" event={"ID":"f233fcdd-5c35-4ffb-bb33-7e2249f1c6c0","Type":"ContainerStarted","Data":"3bbe82973bf2d87f43c634d99288f1bd0c22bc329ac5c3893abf85384bc4c9a4"} Dec 07 16:14:47 crc kubenswrapper[4716]: I1207 16:14:47.583385 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-f8648f98b-t728x" Dec 07 16:14:47 crc kubenswrapper[4716]: I1207 16:14:47.596577 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-f8648f98b-t728x" podStartSLOduration=1.596561168 podStartE2EDuration="1.596561168s" podCreationTimestamp="2025-12-07 16:14:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:14:47.596069585 +0000 UTC m=+750.286354507" watchObservedRunningTime="2025-12-07 16:14:47.596561168 +0000 UTC m=+750.286846090" Dec 07 16:14:48 crc kubenswrapper[4716]: I1207 16:14:48.017583 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/82d3c78d-32ff-47ed-854a-d3786c962ffd-memberlist\") pod \"speaker-xnsb6\" (UID: \"82d3c78d-32ff-47ed-854a-d3786c962ffd\") " pod="metallb-system/speaker-xnsb6" Dec 07 16:14:48 crc kubenswrapper[4716]: I1207 16:14:48.021735 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/82d3c78d-32ff-47ed-854a-d3786c962ffd-memberlist\") pod \"speaker-xnsb6\" (UID: \"82d3c78d-32ff-47ed-854a-d3786c962ffd\") " pod="metallb-system/speaker-xnsb6" Dec 07 16:14:48 crc kubenswrapper[4716]: I1207 16:14:48.181743 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-xnsb6" Dec 07 16:14:48 crc kubenswrapper[4716]: W1207 16:14:48.226185 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod82d3c78d_32ff_47ed_854a_d3786c962ffd.slice/crio-2ff159b5f5734a1670cd7d69fbab97642f0d901e1a4395e3a4ed82dc9924fa0a WatchSource:0}: Error finding container 2ff159b5f5734a1670cd7d69fbab97642f0d901e1a4395e3a4ed82dc9924fa0a: Status 404 returned error can't find the container with id 2ff159b5f5734a1670cd7d69fbab97642f0d901e1a4395e3a4ed82dc9924fa0a Dec 07 16:14:48 crc kubenswrapper[4716]: I1207 16:14:48.612411 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-xnsb6" event={"ID":"82d3c78d-32ff-47ed-854a-d3786c962ffd","Type":"ContainerStarted","Data":"a44e1cd8ffb65c4ebf4c907da40489dbcf07cbd0ac037f32932d14411e563555"} Dec 07 16:14:48 crc kubenswrapper[4716]: I1207 16:14:48.612719 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-xnsb6" event={"ID":"82d3c78d-32ff-47ed-854a-d3786c962ffd","Type":"ContainerStarted","Data":"2ff159b5f5734a1670cd7d69fbab97642f0d901e1a4395e3a4ed82dc9924fa0a"} Dec 07 16:14:49 crc kubenswrapper[4716]: I1207 16:14:49.622053 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-xnsb6" event={"ID":"82d3c78d-32ff-47ed-854a-d3786c962ffd","Type":"ContainerStarted","Data":"108659c8ee02d220e69b45a96f03ffdd6ebbcb0d665050e30a3951e63b5d34e4"} Dec 07 16:14:49 crc kubenswrapper[4716]: I1207 16:14:49.622170 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-xnsb6" Dec 07 16:14:49 crc kubenswrapper[4716]: I1207 16:14:49.661642 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-xnsb6" podStartSLOduration=3.661612051 podStartE2EDuration="3.661612051s" podCreationTimestamp="2025-12-07 16:14:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:14:49.660975134 +0000 UTC m=+752.351260046" watchObservedRunningTime="2025-12-07 16:14:49.661612051 +0000 UTC m=+752.351896983" Dec 07 16:14:54 crc kubenswrapper[4716]: I1207 16:14:54.660225 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-dqzl5" event={"ID":"1a4973da-40f0-4277-a6ff-06ff961b4815","Type":"ContainerStarted","Data":"83de931ef34fd918837dc85d497dcbb7f729a3dbf260c9266b648de6119d3e14"} Dec 07 16:14:54 crc kubenswrapper[4716]: I1207 16:14:54.660465 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-dqzl5" Dec 07 16:14:54 crc kubenswrapper[4716]: I1207 16:14:54.663120 4716 generic.go:334] "Generic (PLEG): container finished" podID="5e169ec0-a81a-45eb-b2a6-8ac97d905caa" containerID="6240a7473336c13be9cea2f7bf4fa7052b1d2c952943d3c592df0be7817945de" exitCode=0 Dec 07 16:14:54 crc kubenswrapper[4716]: I1207 16:14:54.663166 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-9zdjv" event={"ID":"5e169ec0-a81a-45eb-b2a6-8ac97d905caa","Type":"ContainerDied","Data":"6240a7473336c13be9cea2f7bf4fa7052b1d2c952943d3c592df0be7817945de"} Dec 07 16:14:54 crc kubenswrapper[4716]: I1207 16:14:54.684833 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-dqzl5" podStartSLOduration=1.8886361329999999 podStartE2EDuration="8.68481382s" podCreationTimestamp="2025-12-07 16:14:46 +0000 UTC" firstStartedPulling="2025-12-07 16:14:47.059165591 +0000 UTC m=+749.749450503" lastFinishedPulling="2025-12-07 16:14:53.855343288 +0000 UTC m=+756.545628190" observedRunningTime="2025-12-07 16:14:54.684122462 +0000 UTC m=+757.374407404" watchObservedRunningTime="2025-12-07 16:14:54.68481382 +0000 UTC m=+757.375098722" Dec 07 16:14:55 crc kubenswrapper[4716]: I1207 16:14:55.673424 4716 generic.go:334] "Generic (PLEG): container finished" podID="5e169ec0-a81a-45eb-b2a6-8ac97d905caa" containerID="b618013a85fb62b520dd2ad92801a7559c231774483b0a11c365794dbe6efecc" exitCode=0 Dec 07 16:14:55 crc kubenswrapper[4716]: I1207 16:14:55.673495 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-9zdjv" event={"ID":"5e169ec0-a81a-45eb-b2a6-8ac97d905caa","Type":"ContainerDied","Data":"b618013a85fb62b520dd2ad92801a7559c231774483b0a11c365794dbe6efecc"} Dec 07 16:14:56 crc kubenswrapper[4716]: I1207 16:14:56.685957 4716 generic.go:334] "Generic (PLEG): container finished" podID="5e169ec0-a81a-45eb-b2a6-8ac97d905caa" containerID="34328c710c0a044bc6d4fd8e4ddb766e28074241495a7f56ca3d6dcc130d94e2" exitCode=0 Dec 07 16:14:56 crc kubenswrapper[4716]: I1207 16:14:56.686349 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-9zdjv" event={"ID":"5e169ec0-a81a-45eb-b2a6-8ac97d905caa","Type":"ContainerDied","Data":"34328c710c0a044bc6d4fd8e4ddb766e28074241495a7f56ca3d6dcc130d94e2"} Dec 07 16:14:57 crc kubenswrapper[4716]: I1207 16:14:57.696414 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-9zdjv" event={"ID":"5e169ec0-a81a-45eb-b2a6-8ac97d905caa","Type":"ContainerStarted","Data":"cdc7e1234f58dcd21dd62f28f4afc0f8f42b47e1a7f012c95f30ad4c4e3f97b7"} Dec 07 16:14:57 crc kubenswrapper[4716]: I1207 16:14:57.696474 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-9zdjv" event={"ID":"5e169ec0-a81a-45eb-b2a6-8ac97d905caa","Type":"ContainerStarted","Data":"cc7288e4937d3dcebdc2e3315edc8365e0704a5dd6f694966553d1a151e94ddf"} Dec 07 16:14:57 crc kubenswrapper[4716]: I1207 16:14:57.696494 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-9zdjv" event={"ID":"5e169ec0-a81a-45eb-b2a6-8ac97d905caa","Type":"ContainerStarted","Data":"8e6895d8d7adccb974bb97be55eb06e86229bb0b5ed2e40eca9e2e5ebf36c6ac"} Dec 07 16:14:57 crc kubenswrapper[4716]: I1207 16:14:57.696510 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-9zdjv" event={"ID":"5e169ec0-a81a-45eb-b2a6-8ac97d905caa","Type":"ContainerStarted","Data":"0e1533cc40e851c50693d0c668282a0898e6286a1819aa887b9d9e897e96d6d9"} Dec 07 16:14:57 crc kubenswrapper[4716]: I1207 16:14:57.696528 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-9zdjv" event={"ID":"5e169ec0-a81a-45eb-b2a6-8ac97d905caa","Type":"ContainerStarted","Data":"492b2989a07849efee60b434054224d9d5317848de391fa17595920ed8283ff5"} Dec 07 16:14:57 crc kubenswrapper[4716]: I1207 16:14:57.696542 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-9zdjv" event={"ID":"5e169ec0-a81a-45eb-b2a6-8ac97d905caa","Type":"ContainerStarted","Data":"345d611a3e4e483e93e6e4a078b0b1305199cd67762bfb2984c14f1819ab50b7"} Dec 07 16:14:57 crc kubenswrapper[4716]: I1207 16:14:57.696576 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-9zdjv" Dec 07 16:14:57 crc kubenswrapper[4716]: I1207 16:14:57.720159 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-9zdjv" podStartSLOduration=4.61939003 podStartE2EDuration="11.720141269s" podCreationTimestamp="2025-12-07 16:14:46 +0000 UTC" firstStartedPulling="2025-12-07 16:14:46.774127754 +0000 UTC m=+749.464412666" lastFinishedPulling="2025-12-07 16:14:53.874878993 +0000 UTC m=+756.565163905" observedRunningTime="2025-12-07 16:14:57.718427173 +0000 UTC m=+760.408712125" watchObservedRunningTime="2025-12-07 16:14:57.720141269 +0000 UTC m=+760.410426181" Dec 07 16:14:58 crc kubenswrapper[4716]: I1207 16:14:58.192648 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-xnsb6" Dec 07 16:15:00 crc kubenswrapper[4716]: I1207 16:15:00.148516 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29418735-swmgk"] Dec 07 16:15:00 crc kubenswrapper[4716]: I1207 16:15:00.150552 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29418735-swmgk" Dec 07 16:15:00 crc kubenswrapper[4716]: I1207 16:15:00.152923 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 07 16:15:00 crc kubenswrapper[4716]: I1207 16:15:00.154069 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 07 16:15:00 crc kubenswrapper[4716]: I1207 16:15:00.158291 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29418735-swmgk"] Dec 07 16:15:00 crc kubenswrapper[4716]: I1207 16:15:00.345037 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9bbwc\" (UniqueName: \"kubernetes.io/projected/9e624e76-4467-44cc-b0d0-520ced143ba2-kube-api-access-9bbwc\") pod \"collect-profiles-29418735-swmgk\" (UID: \"9e624e76-4467-44cc-b0d0-520ced143ba2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418735-swmgk" Dec 07 16:15:00 crc kubenswrapper[4716]: I1207 16:15:00.345672 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9e624e76-4467-44cc-b0d0-520ced143ba2-config-volume\") pod \"collect-profiles-29418735-swmgk\" (UID: \"9e624e76-4467-44cc-b0d0-520ced143ba2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418735-swmgk" Dec 07 16:15:00 crc kubenswrapper[4716]: I1207 16:15:00.345917 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9e624e76-4467-44cc-b0d0-520ced143ba2-secret-volume\") pod \"collect-profiles-29418735-swmgk\" (UID: \"9e624e76-4467-44cc-b0d0-520ced143ba2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418735-swmgk" Dec 07 16:15:00 crc kubenswrapper[4716]: I1207 16:15:00.446917 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9bbwc\" (UniqueName: \"kubernetes.io/projected/9e624e76-4467-44cc-b0d0-520ced143ba2-kube-api-access-9bbwc\") pod \"collect-profiles-29418735-swmgk\" (UID: \"9e624e76-4467-44cc-b0d0-520ced143ba2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418735-swmgk" Dec 07 16:15:00 crc kubenswrapper[4716]: I1207 16:15:00.447243 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9e624e76-4467-44cc-b0d0-520ced143ba2-config-volume\") pod \"collect-profiles-29418735-swmgk\" (UID: \"9e624e76-4467-44cc-b0d0-520ced143ba2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418735-swmgk" Dec 07 16:15:00 crc kubenswrapper[4716]: I1207 16:15:00.447357 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9e624e76-4467-44cc-b0d0-520ced143ba2-secret-volume\") pod \"collect-profiles-29418735-swmgk\" (UID: \"9e624e76-4467-44cc-b0d0-520ced143ba2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418735-swmgk" Dec 07 16:15:00 crc kubenswrapper[4716]: I1207 16:15:00.449014 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9e624e76-4467-44cc-b0d0-520ced143ba2-config-volume\") pod \"collect-profiles-29418735-swmgk\" (UID: \"9e624e76-4467-44cc-b0d0-520ced143ba2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418735-swmgk" Dec 07 16:15:00 crc kubenswrapper[4716]: I1207 16:15:00.456758 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9e624e76-4467-44cc-b0d0-520ced143ba2-secret-volume\") pod \"collect-profiles-29418735-swmgk\" (UID: \"9e624e76-4467-44cc-b0d0-520ced143ba2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418735-swmgk" Dec 07 16:15:00 crc kubenswrapper[4716]: I1207 16:15:00.471129 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9bbwc\" (UniqueName: \"kubernetes.io/projected/9e624e76-4467-44cc-b0d0-520ced143ba2-kube-api-access-9bbwc\") pod \"collect-profiles-29418735-swmgk\" (UID: \"9e624e76-4467-44cc-b0d0-520ced143ba2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418735-swmgk" Dec 07 16:15:00 crc kubenswrapper[4716]: I1207 16:15:00.474892 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29418735-swmgk" Dec 07 16:15:00 crc kubenswrapper[4716]: I1207 16:15:00.895789 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29418735-swmgk"] Dec 07 16:15:00 crc kubenswrapper[4716]: W1207 16:15:00.901463 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9e624e76_4467_44cc_b0d0_520ced143ba2.slice/crio-0ee494bf2e50a33395e8a49df65328bb8756d2e8419d57ed623e7754fd3dc3f0 WatchSource:0}: Error finding container 0ee494bf2e50a33395e8a49df65328bb8756d2e8419d57ed623e7754fd3dc3f0: Status 404 returned error can't find the container with id 0ee494bf2e50a33395e8a49df65328bb8756d2e8419d57ed623e7754fd3dc3f0 Dec 07 16:15:01 crc kubenswrapper[4716]: I1207 16:15:01.620562 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-9zdjv" Dec 07 16:15:01 crc kubenswrapper[4716]: I1207 16:15:01.674273 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-9zdjv" Dec 07 16:15:01 crc kubenswrapper[4716]: I1207 16:15:01.722997 4716 generic.go:334] "Generic (PLEG): container finished" podID="9e624e76-4467-44cc-b0d0-520ced143ba2" containerID="c5cf5db4504ee2f9b599aae7d13d4c19913d206ce7bf19a6754a54d7185b83ce" exitCode=0 Dec 07 16:15:01 crc kubenswrapper[4716]: I1207 16:15:01.723119 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29418735-swmgk" event={"ID":"9e624e76-4467-44cc-b0d0-520ced143ba2","Type":"ContainerDied","Data":"c5cf5db4504ee2f9b599aae7d13d4c19913d206ce7bf19a6754a54d7185b83ce"} Dec 07 16:15:01 crc kubenswrapper[4716]: I1207 16:15:01.723161 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29418735-swmgk" event={"ID":"9e624e76-4467-44cc-b0d0-520ced143ba2","Type":"ContainerStarted","Data":"0ee494bf2e50a33395e8a49df65328bb8756d2e8419d57ed623e7754fd3dc3f0"} Dec 07 16:15:02 crc kubenswrapper[4716]: I1207 16:15:02.961201 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29418735-swmgk" Dec 07 16:15:02 crc kubenswrapper[4716]: I1207 16:15:02.987931 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9e624e76-4467-44cc-b0d0-520ced143ba2-secret-volume\") pod \"9e624e76-4467-44cc-b0d0-520ced143ba2\" (UID: \"9e624e76-4467-44cc-b0d0-520ced143ba2\") " Dec 07 16:15:02 crc kubenswrapper[4716]: I1207 16:15:02.988114 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9e624e76-4467-44cc-b0d0-520ced143ba2-config-volume\") pod \"9e624e76-4467-44cc-b0d0-520ced143ba2\" (UID: \"9e624e76-4467-44cc-b0d0-520ced143ba2\") " Dec 07 16:15:02 crc kubenswrapper[4716]: I1207 16:15:02.988185 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9bbwc\" (UniqueName: \"kubernetes.io/projected/9e624e76-4467-44cc-b0d0-520ced143ba2-kube-api-access-9bbwc\") pod \"9e624e76-4467-44cc-b0d0-520ced143ba2\" (UID: \"9e624e76-4467-44cc-b0d0-520ced143ba2\") " Dec 07 16:15:02 crc kubenswrapper[4716]: I1207 16:15:02.988985 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e624e76-4467-44cc-b0d0-520ced143ba2-config-volume" (OuterVolumeSpecName: "config-volume") pod "9e624e76-4467-44cc-b0d0-520ced143ba2" (UID: "9e624e76-4467-44cc-b0d0-520ced143ba2"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:15:02 crc kubenswrapper[4716]: I1207 16:15:02.994449 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e624e76-4467-44cc-b0d0-520ced143ba2-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "9e624e76-4467-44cc-b0d0-520ced143ba2" (UID: "9e624e76-4467-44cc-b0d0-520ced143ba2"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:15:02 crc kubenswrapper[4716]: I1207 16:15:02.994500 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e624e76-4467-44cc-b0d0-520ced143ba2-kube-api-access-9bbwc" (OuterVolumeSpecName: "kube-api-access-9bbwc") pod "9e624e76-4467-44cc-b0d0-520ced143ba2" (UID: "9e624e76-4467-44cc-b0d0-520ced143ba2"). InnerVolumeSpecName "kube-api-access-9bbwc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:15:03 crc kubenswrapper[4716]: I1207 16:15:03.090464 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9bbwc\" (UniqueName: \"kubernetes.io/projected/9e624e76-4467-44cc-b0d0-520ced143ba2-kube-api-access-9bbwc\") on node \"crc\" DevicePath \"\"" Dec 07 16:15:03 crc kubenswrapper[4716]: I1207 16:15:03.090527 4716 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9e624e76-4467-44cc-b0d0-520ced143ba2-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 07 16:15:03 crc kubenswrapper[4716]: I1207 16:15:03.090552 4716 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9e624e76-4467-44cc-b0d0-520ced143ba2-config-volume\") on node \"crc\" DevicePath \"\"" Dec 07 16:15:03 crc kubenswrapper[4716]: I1207 16:15:03.736448 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29418735-swmgk" event={"ID":"9e624e76-4467-44cc-b0d0-520ced143ba2","Type":"ContainerDied","Data":"0ee494bf2e50a33395e8a49df65328bb8756d2e8419d57ed623e7754fd3dc3f0"} Dec 07 16:15:03 crc kubenswrapper[4716]: I1207 16:15:03.736782 4716 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0ee494bf2e50a33395e8a49df65328bb8756d2e8419d57ed623e7754fd3dc3f0" Dec 07 16:15:03 crc kubenswrapper[4716]: I1207 16:15:03.736522 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29418735-swmgk" Dec 07 16:15:04 crc kubenswrapper[4716]: I1207 16:15:04.494952 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-m57m9"] Dec 07 16:15:04 crc kubenswrapper[4716]: E1207 16:15:04.495286 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e624e76-4467-44cc-b0d0-520ced143ba2" containerName="collect-profiles" Dec 07 16:15:04 crc kubenswrapper[4716]: I1207 16:15:04.495298 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e624e76-4467-44cc-b0d0-520ced143ba2" containerName="collect-profiles" Dec 07 16:15:04 crc kubenswrapper[4716]: I1207 16:15:04.495403 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e624e76-4467-44cc-b0d0-520ced143ba2" containerName="collect-profiles" Dec 07 16:15:04 crc kubenswrapper[4716]: I1207 16:15:04.495812 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-m57m9" Dec 07 16:15:04 crc kubenswrapper[4716]: I1207 16:15:04.499422 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Dec 07 16:15:04 crc kubenswrapper[4716]: I1207 16:15:04.499450 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Dec 07 16:15:04 crc kubenswrapper[4716]: I1207 16:15:04.499859 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-h655l" Dec 07 16:15:04 crc kubenswrapper[4716]: I1207 16:15:04.504949 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-m57m9"] Dec 07 16:15:04 crc kubenswrapper[4716]: I1207 16:15:04.510545 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9ssq\" (UniqueName: \"kubernetes.io/projected/aeb52c61-7bf6-4e04-8e0a-c9e4dcd2d4de-kube-api-access-t9ssq\") pod \"openstack-operator-index-m57m9\" (UID: \"aeb52c61-7bf6-4e04-8e0a-c9e4dcd2d4de\") " pod="openstack-operators/openstack-operator-index-m57m9" Dec 07 16:15:04 crc kubenswrapper[4716]: I1207 16:15:04.611463 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9ssq\" (UniqueName: \"kubernetes.io/projected/aeb52c61-7bf6-4e04-8e0a-c9e4dcd2d4de-kube-api-access-t9ssq\") pod \"openstack-operator-index-m57m9\" (UID: \"aeb52c61-7bf6-4e04-8e0a-c9e4dcd2d4de\") " pod="openstack-operators/openstack-operator-index-m57m9" Dec 07 16:15:04 crc kubenswrapper[4716]: I1207 16:15:04.632859 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9ssq\" (UniqueName: \"kubernetes.io/projected/aeb52c61-7bf6-4e04-8e0a-c9e4dcd2d4de-kube-api-access-t9ssq\") pod \"openstack-operator-index-m57m9\" (UID: \"aeb52c61-7bf6-4e04-8e0a-c9e4dcd2d4de\") " pod="openstack-operators/openstack-operator-index-m57m9" Dec 07 16:15:04 crc kubenswrapper[4716]: I1207 16:15:04.822409 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-m57m9" Dec 07 16:15:05 crc kubenswrapper[4716]: I1207 16:15:05.088155 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-m57m9"] Dec 07 16:15:05 crc kubenswrapper[4716]: W1207 16:15:05.095611 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaeb52c61_7bf6_4e04_8e0a_c9e4dcd2d4de.slice/crio-fdb741c411d0b5387540e54b7d50f99e20e072c849ee9c4c742cc093b59c428a WatchSource:0}: Error finding container fdb741c411d0b5387540e54b7d50f99e20e072c849ee9c4c742cc093b59c428a: Status 404 returned error can't find the container with id fdb741c411d0b5387540e54b7d50f99e20e072c849ee9c4c742cc093b59c428a Dec 07 16:15:05 crc kubenswrapper[4716]: I1207 16:15:05.749763 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-m57m9" event={"ID":"aeb52c61-7bf6-4e04-8e0a-c9e4dcd2d4de","Type":"ContainerStarted","Data":"fdb741c411d0b5387540e54b7d50f99e20e072c849ee9c4c742cc093b59c428a"} Dec 07 16:15:06 crc kubenswrapper[4716]: I1207 16:15:06.605418 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-dqzl5" Dec 07 16:15:06 crc kubenswrapper[4716]: I1207 16:15:06.624537 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-9zdjv" Dec 07 16:15:06 crc kubenswrapper[4716]: I1207 16:15:06.722227 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-f8648f98b-t728x" Dec 07 16:15:09 crc kubenswrapper[4716]: I1207 16:15:09.678466 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-m57m9"] Dec 07 16:15:09 crc kubenswrapper[4716]: I1207 16:15:09.777319 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-m57m9" event={"ID":"aeb52c61-7bf6-4e04-8e0a-c9e4dcd2d4de","Type":"ContainerStarted","Data":"8cf6bb9606da09222d0ec3afcd6491b71deed8bf7cd5ec2f0693df2c58cf2148"} Dec 07 16:15:09 crc kubenswrapper[4716]: I1207 16:15:09.789171 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-m57m9" podStartSLOduration=1.464171748 podStartE2EDuration="5.78915126s" podCreationTimestamp="2025-12-07 16:15:04 +0000 UTC" firstStartedPulling="2025-12-07 16:15:05.097385985 +0000 UTC m=+767.787670897" lastFinishedPulling="2025-12-07 16:15:09.422365497 +0000 UTC m=+772.112650409" observedRunningTime="2025-12-07 16:15:09.788867053 +0000 UTC m=+772.479151975" watchObservedRunningTime="2025-12-07 16:15:09.78915126 +0000 UTC m=+772.479436172" Dec 07 16:15:10 crc kubenswrapper[4716]: I1207 16:15:10.292196 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-qcnsg"] Dec 07 16:15:10 crc kubenswrapper[4716]: I1207 16:15:10.293729 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-qcnsg" Dec 07 16:15:10 crc kubenswrapper[4716]: I1207 16:15:10.301231 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-qcnsg"] Dec 07 16:15:10 crc kubenswrapper[4716]: I1207 16:15:10.392524 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67sp7\" (UniqueName: \"kubernetes.io/projected/12e91e68-d5ea-4b85-81ca-88f8d00e06bc-kube-api-access-67sp7\") pod \"openstack-operator-index-qcnsg\" (UID: \"12e91e68-d5ea-4b85-81ca-88f8d00e06bc\") " pod="openstack-operators/openstack-operator-index-qcnsg" Dec 07 16:15:10 crc kubenswrapper[4716]: I1207 16:15:10.493965 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67sp7\" (UniqueName: \"kubernetes.io/projected/12e91e68-d5ea-4b85-81ca-88f8d00e06bc-kube-api-access-67sp7\") pod \"openstack-operator-index-qcnsg\" (UID: \"12e91e68-d5ea-4b85-81ca-88f8d00e06bc\") " pod="openstack-operators/openstack-operator-index-qcnsg" Dec 07 16:15:10 crc kubenswrapper[4716]: I1207 16:15:10.513382 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67sp7\" (UniqueName: \"kubernetes.io/projected/12e91e68-d5ea-4b85-81ca-88f8d00e06bc-kube-api-access-67sp7\") pod \"openstack-operator-index-qcnsg\" (UID: \"12e91e68-d5ea-4b85-81ca-88f8d00e06bc\") " pod="openstack-operators/openstack-operator-index-qcnsg" Dec 07 16:15:10 crc kubenswrapper[4716]: I1207 16:15:10.613747 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-qcnsg" Dec 07 16:15:10 crc kubenswrapper[4716]: I1207 16:15:10.785630 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-m57m9" podUID="aeb52c61-7bf6-4e04-8e0a-c9e4dcd2d4de" containerName="registry-server" containerID="cri-o://8cf6bb9606da09222d0ec3afcd6491b71deed8bf7cd5ec2f0693df2c58cf2148" gracePeriod=2 Dec 07 16:15:11 crc kubenswrapper[4716]: I1207 16:15:11.025847 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-qcnsg"] Dec 07 16:15:11 crc kubenswrapper[4716]: W1207 16:15:11.036173 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod12e91e68_d5ea_4b85_81ca_88f8d00e06bc.slice/crio-4cd46a0e034e33f1b65d24acbba1cd69b4b4c48b76700985c470bbc1b2617a79 WatchSource:0}: Error finding container 4cd46a0e034e33f1b65d24acbba1cd69b4b4c48b76700985c470bbc1b2617a79: Status 404 returned error can't find the container with id 4cd46a0e034e33f1b65d24acbba1cd69b4b4c48b76700985c470bbc1b2617a79 Dec 07 16:15:11 crc kubenswrapper[4716]: I1207 16:15:11.110473 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-m57m9" Dec 07 16:15:11 crc kubenswrapper[4716]: I1207 16:15:11.306901 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t9ssq\" (UniqueName: \"kubernetes.io/projected/aeb52c61-7bf6-4e04-8e0a-c9e4dcd2d4de-kube-api-access-t9ssq\") pod \"aeb52c61-7bf6-4e04-8e0a-c9e4dcd2d4de\" (UID: \"aeb52c61-7bf6-4e04-8e0a-c9e4dcd2d4de\") " Dec 07 16:15:11 crc kubenswrapper[4716]: I1207 16:15:11.313311 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aeb52c61-7bf6-4e04-8e0a-c9e4dcd2d4de-kube-api-access-t9ssq" (OuterVolumeSpecName: "kube-api-access-t9ssq") pod "aeb52c61-7bf6-4e04-8e0a-c9e4dcd2d4de" (UID: "aeb52c61-7bf6-4e04-8e0a-c9e4dcd2d4de"). InnerVolumeSpecName "kube-api-access-t9ssq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:15:11 crc kubenswrapper[4716]: I1207 16:15:11.408320 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t9ssq\" (UniqueName: \"kubernetes.io/projected/aeb52c61-7bf6-4e04-8e0a-c9e4dcd2d4de-kube-api-access-t9ssq\") on node \"crc\" DevicePath \"\"" Dec 07 16:15:11 crc kubenswrapper[4716]: I1207 16:15:11.794899 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-qcnsg" event={"ID":"12e91e68-d5ea-4b85-81ca-88f8d00e06bc","Type":"ContainerStarted","Data":"9887ed9c8c8a5fd08606c43927a864351ca2762af8b4851d1107cd8ca2edb1f4"} Dec 07 16:15:11 crc kubenswrapper[4716]: I1207 16:15:11.795467 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-qcnsg" event={"ID":"12e91e68-d5ea-4b85-81ca-88f8d00e06bc","Type":"ContainerStarted","Data":"4cd46a0e034e33f1b65d24acbba1cd69b4b4c48b76700985c470bbc1b2617a79"} Dec 07 16:15:11 crc kubenswrapper[4716]: I1207 16:15:11.798200 4716 generic.go:334] "Generic (PLEG): container finished" podID="aeb52c61-7bf6-4e04-8e0a-c9e4dcd2d4de" containerID="8cf6bb9606da09222d0ec3afcd6491b71deed8bf7cd5ec2f0693df2c58cf2148" exitCode=0 Dec 07 16:15:11 crc kubenswrapper[4716]: I1207 16:15:11.798251 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-m57m9" event={"ID":"aeb52c61-7bf6-4e04-8e0a-c9e4dcd2d4de","Type":"ContainerDied","Data":"8cf6bb9606da09222d0ec3afcd6491b71deed8bf7cd5ec2f0693df2c58cf2148"} Dec 07 16:15:11 crc kubenswrapper[4716]: I1207 16:15:11.798284 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-m57m9" Dec 07 16:15:11 crc kubenswrapper[4716]: I1207 16:15:11.798313 4716 scope.go:117] "RemoveContainer" containerID="8cf6bb9606da09222d0ec3afcd6491b71deed8bf7cd5ec2f0693df2c58cf2148" Dec 07 16:15:11 crc kubenswrapper[4716]: I1207 16:15:11.798299 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-m57m9" event={"ID":"aeb52c61-7bf6-4e04-8e0a-c9e4dcd2d4de","Type":"ContainerDied","Data":"fdb741c411d0b5387540e54b7d50f99e20e072c849ee9c4c742cc093b59c428a"} Dec 07 16:15:11 crc kubenswrapper[4716]: I1207 16:15:11.817164 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-qcnsg" podStartSLOduration=1.7617402 podStartE2EDuration="1.817139408s" podCreationTimestamp="2025-12-07 16:15:10 +0000 UTC" firstStartedPulling="2025-12-07 16:15:11.041889133 +0000 UTC m=+773.732174045" lastFinishedPulling="2025-12-07 16:15:11.097288331 +0000 UTC m=+773.787573253" observedRunningTime="2025-12-07 16:15:11.813520992 +0000 UTC m=+774.503805914" watchObservedRunningTime="2025-12-07 16:15:11.817139408 +0000 UTC m=+774.507424350" Dec 07 16:15:11 crc kubenswrapper[4716]: I1207 16:15:11.821953 4716 scope.go:117] "RemoveContainer" containerID="8cf6bb9606da09222d0ec3afcd6491b71deed8bf7cd5ec2f0693df2c58cf2148" Dec 07 16:15:11 crc kubenswrapper[4716]: E1207 16:15:11.822401 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8cf6bb9606da09222d0ec3afcd6491b71deed8bf7cd5ec2f0693df2c58cf2148\": container with ID starting with 8cf6bb9606da09222d0ec3afcd6491b71deed8bf7cd5ec2f0693df2c58cf2148 not found: ID does not exist" containerID="8cf6bb9606da09222d0ec3afcd6491b71deed8bf7cd5ec2f0693df2c58cf2148" Dec 07 16:15:11 crc kubenswrapper[4716]: I1207 16:15:11.822466 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8cf6bb9606da09222d0ec3afcd6491b71deed8bf7cd5ec2f0693df2c58cf2148"} err="failed to get container status \"8cf6bb9606da09222d0ec3afcd6491b71deed8bf7cd5ec2f0693df2c58cf2148\": rpc error: code = NotFound desc = could not find container \"8cf6bb9606da09222d0ec3afcd6491b71deed8bf7cd5ec2f0693df2c58cf2148\": container with ID starting with 8cf6bb9606da09222d0ec3afcd6491b71deed8bf7cd5ec2f0693df2c58cf2148 not found: ID does not exist" Dec 07 16:15:11 crc kubenswrapper[4716]: I1207 16:15:11.832196 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-m57m9"] Dec 07 16:15:11 crc kubenswrapper[4716]: I1207 16:15:11.835502 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-m57m9"] Dec 07 16:15:13 crc kubenswrapper[4716]: I1207 16:15:13.667179 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aeb52c61-7bf6-4e04-8e0a-c9e4dcd2d4de" path="/var/lib/kubelet/pods/aeb52c61-7bf6-4e04-8e0a-c9e4dcd2d4de/volumes" Dec 07 16:15:20 crc kubenswrapper[4716]: I1207 16:15:20.614542 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-qcnsg" Dec 07 16:15:20 crc kubenswrapper[4716]: I1207 16:15:20.616201 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-qcnsg" Dec 07 16:15:20 crc kubenswrapper[4716]: I1207 16:15:20.640978 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-qcnsg" Dec 07 16:15:20 crc kubenswrapper[4716]: I1207 16:15:20.899714 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-qcnsg" Dec 07 16:15:22 crc kubenswrapper[4716]: I1207 16:15:22.761323 4716 patch_prober.go:28] interesting pod/machine-config-daemon-zcxxp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 16:15:22 crc kubenswrapper[4716]: I1207 16:15:22.761672 4716 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 16:15:34 crc kubenswrapper[4716]: I1207 16:15:34.775372 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/4cb94d7f59637b410c6ea06aeac57a499ffab903391b8415798aca23986vcrp"] Dec 07 16:15:34 crc kubenswrapper[4716]: E1207 16:15:34.776559 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aeb52c61-7bf6-4e04-8e0a-c9e4dcd2d4de" containerName="registry-server" Dec 07 16:15:34 crc kubenswrapper[4716]: I1207 16:15:34.776579 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="aeb52c61-7bf6-4e04-8e0a-c9e4dcd2d4de" containerName="registry-server" Dec 07 16:15:34 crc kubenswrapper[4716]: I1207 16:15:34.776925 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="aeb52c61-7bf6-4e04-8e0a-c9e4dcd2d4de" containerName="registry-server" Dec 07 16:15:34 crc kubenswrapper[4716]: I1207 16:15:34.780592 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/4cb94d7f59637b410c6ea06aeac57a499ffab903391b8415798aca23986vcrp" Dec 07 16:15:34 crc kubenswrapper[4716]: I1207 16:15:34.806556 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-z7khq" Dec 07 16:15:34 crc kubenswrapper[4716]: I1207 16:15:34.832140 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/4cb94d7f59637b410c6ea06aeac57a499ffab903391b8415798aca23986vcrp"] Dec 07 16:15:34 crc kubenswrapper[4716]: I1207 16:15:34.947253 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7w2w7\" (UniqueName: \"kubernetes.io/projected/5bdf0591-fc57-4545-8ab4-e89de972df50-kube-api-access-7w2w7\") pod \"4cb94d7f59637b410c6ea06aeac57a499ffab903391b8415798aca23986vcrp\" (UID: \"5bdf0591-fc57-4545-8ab4-e89de972df50\") " pod="openstack-operators/4cb94d7f59637b410c6ea06aeac57a499ffab903391b8415798aca23986vcrp" Dec 07 16:15:34 crc kubenswrapper[4716]: I1207 16:15:34.947333 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5bdf0591-fc57-4545-8ab4-e89de972df50-bundle\") pod \"4cb94d7f59637b410c6ea06aeac57a499ffab903391b8415798aca23986vcrp\" (UID: \"5bdf0591-fc57-4545-8ab4-e89de972df50\") " pod="openstack-operators/4cb94d7f59637b410c6ea06aeac57a499ffab903391b8415798aca23986vcrp" Dec 07 16:15:34 crc kubenswrapper[4716]: I1207 16:15:34.947443 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5bdf0591-fc57-4545-8ab4-e89de972df50-util\") pod \"4cb94d7f59637b410c6ea06aeac57a499ffab903391b8415798aca23986vcrp\" (UID: \"5bdf0591-fc57-4545-8ab4-e89de972df50\") " pod="openstack-operators/4cb94d7f59637b410c6ea06aeac57a499ffab903391b8415798aca23986vcrp" Dec 07 16:15:35 crc kubenswrapper[4716]: I1207 16:15:35.048371 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5bdf0591-fc57-4545-8ab4-e89de972df50-util\") pod \"4cb94d7f59637b410c6ea06aeac57a499ffab903391b8415798aca23986vcrp\" (UID: \"5bdf0591-fc57-4545-8ab4-e89de972df50\") " pod="openstack-operators/4cb94d7f59637b410c6ea06aeac57a499ffab903391b8415798aca23986vcrp" Dec 07 16:15:35 crc kubenswrapper[4716]: I1207 16:15:35.048438 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7w2w7\" (UniqueName: \"kubernetes.io/projected/5bdf0591-fc57-4545-8ab4-e89de972df50-kube-api-access-7w2w7\") pod \"4cb94d7f59637b410c6ea06aeac57a499ffab903391b8415798aca23986vcrp\" (UID: \"5bdf0591-fc57-4545-8ab4-e89de972df50\") " pod="openstack-operators/4cb94d7f59637b410c6ea06aeac57a499ffab903391b8415798aca23986vcrp" Dec 07 16:15:35 crc kubenswrapper[4716]: I1207 16:15:35.048469 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5bdf0591-fc57-4545-8ab4-e89de972df50-bundle\") pod \"4cb94d7f59637b410c6ea06aeac57a499ffab903391b8415798aca23986vcrp\" (UID: \"5bdf0591-fc57-4545-8ab4-e89de972df50\") " pod="openstack-operators/4cb94d7f59637b410c6ea06aeac57a499ffab903391b8415798aca23986vcrp" Dec 07 16:15:35 crc kubenswrapper[4716]: I1207 16:15:35.049033 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5bdf0591-fc57-4545-8ab4-e89de972df50-bundle\") pod \"4cb94d7f59637b410c6ea06aeac57a499ffab903391b8415798aca23986vcrp\" (UID: \"5bdf0591-fc57-4545-8ab4-e89de972df50\") " pod="openstack-operators/4cb94d7f59637b410c6ea06aeac57a499ffab903391b8415798aca23986vcrp" Dec 07 16:15:35 crc kubenswrapper[4716]: I1207 16:15:35.049544 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5bdf0591-fc57-4545-8ab4-e89de972df50-util\") pod \"4cb94d7f59637b410c6ea06aeac57a499ffab903391b8415798aca23986vcrp\" (UID: \"5bdf0591-fc57-4545-8ab4-e89de972df50\") " pod="openstack-operators/4cb94d7f59637b410c6ea06aeac57a499ffab903391b8415798aca23986vcrp" Dec 07 16:15:35 crc kubenswrapper[4716]: I1207 16:15:35.076998 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7w2w7\" (UniqueName: \"kubernetes.io/projected/5bdf0591-fc57-4545-8ab4-e89de972df50-kube-api-access-7w2w7\") pod \"4cb94d7f59637b410c6ea06aeac57a499ffab903391b8415798aca23986vcrp\" (UID: \"5bdf0591-fc57-4545-8ab4-e89de972df50\") " pod="openstack-operators/4cb94d7f59637b410c6ea06aeac57a499ffab903391b8415798aca23986vcrp" Dec 07 16:15:35 crc kubenswrapper[4716]: I1207 16:15:35.126492 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/4cb94d7f59637b410c6ea06aeac57a499ffab903391b8415798aca23986vcrp" Dec 07 16:15:35 crc kubenswrapper[4716]: I1207 16:15:35.423043 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/4cb94d7f59637b410c6ea06aeac57a499ffab903391b8415798aca23986vcrp"] Dec 07 16:15:35 crc kubenswrapper[4716]: I1207 16:15:35.967815 4716 generic.go:334] "Generic (PLEG): container finished" podID="5bdf0591-fc57-4545-8ab4-e89de972df50" containerID="ff3f0488998a972acdb758f5049d0b0323c825fd0a29cca2410e728c2c4aef6d" exitCode=0 Dec 07 16:15:35 crc kubenswrapper[4716]: I1207 16:15:35.967883 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/4cb94d7f59637b410c6ea06aeac57a499ffab903391b8415798aca23986vcrp" event={"ID":"5bdf0591-fc57-4545-8ab4-e89de972df50","Type":"ContainerDied","Data":"ff3f0488998a972acdb758f5049d0b0323c825fd0a29cca2410e728c2c4aef6d"} Dec 07 16:15:35 crc kubenswrapper[4716]: I1207 16:15:35.968361 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/4cb94d7f59637b410c6ea06aeac57a499ffab903391b8415798aca23986vcrp" event={"ID":"5bdf0591-fc57-4545-8ab4-e89de972df50","Type":"ContainerStarted","Data":"a98b5babe29bb98209c8eed5ef0617b7ab5bad61bb578ae40f1be5e700b8e125"} Dec 07 16:15:36 crc kubenswrapper[4716]: I1207 16:15:36.975204 4716 generic.go:334] "Generic (PLEG): container finished" podID="5bdf0591-fc57-4545-8ab4-e89de972df50" containerID="efe67c933641225986ea0e8ea6ae689b87ea6ccf9dbf29ea554dd1ee10435923" exitCode=0 Dec 07 16:15:36 crc kubenswrapper[4716]: I1207 16:15:36.975252 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/4cb94d7f59637b410c6ea06aeac57a499ffab903391b8415798aca23986vcrp" event={"ID":"5bdf0591-fc57-4545-8ab4-e89de972df50","Type":"ContainerDied","Data":"efe67c933641225986ea0e8ea6ae689b87ea6ccf9dbf29ea554dd1ee10435923"} Dec 07 16:15:37 crc kubenswrapper[4716]: E1207 16:15:37.327437 4716 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5bdf0591_fc57_4545_8ab4_e89de972df50.slice/crio-85b71af4893f82bdd7fde6e04a51c3a793a9f6f4e375fe20dda23dc0dd6d9d81.scope\": RecentStats: unable to find data in memory cache]" Dec 07 16:15:37 crc kubenswrapper[4716]: I1207 16:15:37.985311 4716 generic.go:334] "Generic (PLEG): container finished" podID="5bdf0591-fc57-4545-8ab4-e89de972df50" containerID="85b71af4893f82bdd7fde6e04a51c3a793a9f6f4e375fe20dda23dc0dd6d9d81" exitCode=0 Dec 07 16:15:37 crc kubenswrapper[4716]: I1207 16:15:37.985410 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/4cb94d7f59637b410c6ea06aeac57a499ffab903391b8415798aca23986vcrp" event={"ID":"5bdf0591-fc57-4545-8ab4-e89de972df50","Type":"ContainerDied","Data":"85b71af4893f82bdd7fde6e04a51c3a793a9f6f4e375fe20dda23dc0dd6d9d81"} Dec 07 16:15:39 crc kubenswrapper[4716]: I1207 16:15:39.285522 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/4cb94d7f59637b410c6ea06aeac57a499ffab903391b8415798aca23986vcrp" Dec 07 16:15:39 crc kubenswrapper[4716]: I1207 16:15:39.439122 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5bdf0591-fc57-4545-8ab4-e89de972df50-util\") pod \"5bdf0591-fc57-4545-8ab4-e89de972df50\" (UID: \"5bdf0591-fc57-4545-8ab4-e89de972df50\") " Dec 07 16:15:39 crc kubenswrapper[4716]: I1207 16:15:39.439273 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7w2w7\" (UniqueName: \"kubernetes.io/projected/5bdf0591-fc57-4545-8ab4-e89de972df50-kube-api-access-7w2w7\") pod \"5bdf0591-fc57-4545-8ab4-e89de972df50\" (UID: \"5bdf0591-fc57-4545-8ab4-e89de972df50\") " Dec 07 16:15:39 crc kubenswrapper[4716]: I1207 16:15:39.439305 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5bdf0591-fc57-4545-8ab4-e89de972df50-bundle\") pod \"5bdf0591-fc57-4545-8ab4-e89de972df50\" (UID: \"5bdf0591-fc57-4545-8ab4-e89de972df50\") " Dec 07 16:15:39 crc kubenswrapper[4716]: I1207 16:15:39.440124 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5bdf0591-fc57-4545-8ab4-e89de972df50-bundle" (OuterVolumeSpecName: "bundle") pod "5bdf0591-fc57-4545-8ab4-e89de972df50" (UID: "5bdf0591-fc57-4545-8ab4-e89de972df50"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:15:39 crc kubenswrapper[4716]: I1207 16:15:39.449907 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5bdf0591-fc57-4545-8ab4-e89de972df50-kube-api-access-7w2w7" (OuterVolumeSpecName: "kube-api-access-7w2w7") pod "5bdf0591-fc57-4545-8ab4-e89de972df50" (UID: "5bdf0591-fc57-4545-8ab4-e89de972df50"). InnerVolumeSpecName "kube-api-access-7w2w7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:15:39 crc kubenswrapper[4716]: I1207 16:15:39.458926 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5bdf0591-fc57-4545-8ab4-e89de972df50-util" (OuterVolumeSpecName: "util") pod "5bdf0591-fc57-4545-8ab4-e89de972df50" (UID: "5bdf0591-fc57-4545-8ab4-e89de972df50"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:15:39 crc kubenswrapper[4716]: I1207 16:15:39.541368 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7w2w7\" (UniqueName: \"kubernetes.io/projected/5bdf0591-fc57-4545-8ab4-e89de972df50-kube-api-access-7w2w7\") on node \"crc\" DevicePath \"\"" Dec 07 16:15:39 crc kubenswrapper[4716]: I1207 16:15:39.541434 4716 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5bdf0591-fc57-4545-8ab4-e89de972df50-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 16:15:39 crc kubenswrapper[4716]: I1207 16:15:39.541445 4716 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5bdf0591-fc57-4545-8ab4-e89de972df50-util\") on node \"crc\" DevicePath \"\"" Dec 07 16:15:40 crc kubenswrapper[4716]: I1207 16:15:40.005047 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/4cb94d7f59637b410c6ea06aeac57a499ffab903391b8415798aca23986vcrp" event={"ID":"5bdf0591-fc57-4545-8ab4-e89de972df50","Type":"ContainerDied","Data":"a98b5babe29bb98209c8eed5ef0617b7ab5bad61bb578ae40f1be5e700b8e125"} Dec 07 16:15:40 crc kubenswrapper[4716]: I1207 16:15:40.005166 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/4cb94d7f59637b410c6ea06aeac57a499ffab903391b8415798aca23986vcrp" Dec 07 16:15:40 crc kubenswrapper[4716]: I1207 16:15:40.005188 4716 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a98b5babe29bb98209c8eed5ef0617b7ab5bad61bb578ae40f1be5e700b8e125" Dec 07 16:15:48 crc kubenswrapper[4716]: I1207 16:15:48.810983 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-64d95d5f5f-6t4rp"] Dec 07 16:15:48 crc kubenswrapper[4716]: E1207 16:15:48.811900 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bdf0591-fc57-4545-8ab4-e89de972df50" containerName="util" Dec 07 16:15:48 crc kubenswrapper[4716]: I1207 16:15:48.811919 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bdf0591-fc57-4545-8ab4-e89de972df50" containerName="util" Dec 07 16:15:48 crc kubenswrapper[4716]: E1207 16:15:48.811930 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bdf0591-fc57-4545-8ab4-e89de972df50" containerName="pull" Dec 07 16:15:48 crc kubenswrapper[4716]: I1207 16:15:48.811937 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bdf0591-fc57-4545-8ab4-e89de972df50" containerName="pull" Dec 07 16:15:48 crc kubenswrapper[4716]: E1207 16:15:48.811952 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bdf0591-fc57-4545-8ab4-e89de972df50" containerName="extract" Dec 07 16:15:48 crc kubenswrapper[4716]: I1207 16:15:48.811961 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bdf0591-fc57-4545-8ab4-e89de972df50" containerName="extract" Dec 07 16:15:48 crc kubenswrapper[4716]: I1207 16:15:48.812157 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="5bdf0591-fc57-4545-8ab4-e89de972df50" containerName="extract" Dec 07 16:15:48 crc kubenswrapper[4716]: I1207 16:15:48.812607 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-64d95d5f5f-6t4rp" Dec 07 16:15:48 crc kubenswrapper[4716]: I1207 16:15:48.815661 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-f5fhs" Dec 07 16:15:48 crc kubenswrapper[4716]: I1207 16:15:48.886990 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jlf8k\" (UniqueName: \"kubernetes.io/projected/93975489-d46c-4ea7-a11b-65202433b51f-kube-api-access-jlf8k\") pod \"openstack-operator-controller-operator-64d95d5f5f-6t4rp\" (UID: \"93975489-d46c-4ea7-a11b-65202433b51f\") " pod="openstack-operators/openstack-operator-controller-operator-64d95d5f5f-6t4rp" Dec 07 16:15:48 crc kubenswrapper[4716]: I1207 16:15:48.920918 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-64d95d5f5f-6t4rp"] Dec 07 16:15:48 crc kubenswrapper[4716]: I1207 16:15:48.988031 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jlf8k\" (UniqueName: \"kubernetes.io/projected/93975489-d46c-4ea7-a11b-65202433b51f-kube-api-access-jlf8k\") pod \"openstack-operator-controller-operator-64d95d5f5f-6t4rp\" (UID: \"93975489-d46c-4ea7-a11b-65202433b51f\") " pod="openstack-operators/openstack-operator-controller-operator-64d95d5f5f-6t4rp" Dec 07 16:15:50 crc kubenswrapper[4716]: I1207 16:15:50.258577 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jlf8k\" (UniqueName: \"kubernetes.io/projected/93975489-d46c-4ea7-a11b-65202433b51f-kube-api-access-jlf8k\") pod \"openstack-operator-controller-operator-64d95d5f5f-6t4rp\" (UID: \"93975489-d46c-4ea7-a11b-65202433b51f\") " pod="openstack-operators/openstack-operator-controller-operator-64d95d5f5f-6t4rp" Dec 07 16:15:50 crc kubenswrapper[4716]: I1207 16:15:50.337814 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-64d95d5f5f-6t4rp" Dec 07 16:15:50 crc kubenswrapper[4716]: I1207 16:15:50.617203 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-64d95d5f5f-6t4rp"] Dec 07 16:15:50 crc kubenswrapper[4716]: W1207 16:15:50.627722 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod93975489_d46c_4ea7_a11b_65202433b51f.slice/crio-bb98da6840330e0594fc74d5676fa4658dff712fa147518fdcc1ae71fd5757d1 WatchSource:0}: Error finding container bb98da6840330e0594fc74d5676fa4658dff712fa147518fdcc1ae71fd5757d1: Status 404 returned error can't find the container with id bb98da6840330e0594fc74d5676fa4658dff712fa147518fdcc1ae71fd5757d1 Dec 07 16:15:51 crc kubenswrapper[4716]: I1207 16:15:51.082623 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-64d95d5f5f-6t4rp" event={"ID":"93975489-d46c-4ea7-a11b-65202433b51f","Type":"ContainerStarted","Data":"bb98da6840330e0594fc74d5676fa4658dff712fa147518fdcc1ae71fd5757d1"} Dec 07 16:15:52 crc kubenswrapper[4716]: I1207 16:15:52.761251 4716 patch_prober.go:28] interesting pod/machine-config-daemon-zcxxp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 16:15:52 crc kubenswrapper[4716]: I1207 16:15:52.761509 4716 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 16:15:55 crc kubenswrapper[4716]: I1207 16:15:55.109021 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-64d95d5f5f-6t4rp" event={"ID":"93975489-d46c-4ea7-a11b-65202433b51f","Type":"ContainerStarted","Data":"affb50cb887192898433191e280a2fd81c4285ffd7249639aa44ea33d3028867"} Dec 07 16:15:55 crc kubenswrapper[4716]: I1207 16:15:55.109588 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-64d95d5f5f-6t4rp" Dec 07 16:15:55 crc kubenswrapper[4716]: I1207 16:15:55.148281 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-64d95d5f5f-6t4rp" podStartSLOduration=3.798547115 podStartE2EDuration="7.14825498s" podCreationTimestamp="2025-12-07 16:15:48 +0000 UTC" firstStartedPulling="2025-12-07 16:15:50.63277865 +0000 UTC m=+813.323063552" lastFinishedPulling="2025-12-07 16:15:53.982486515 +0000 UTC m=+816.672771417" observedRunningTime="2025-12-07 16:15:55.141205278 +0000 UTC m=+817.831490190" watchObservedRunningTime="2025-12-07 16:15:55.14825498 +0000 UTC m=+817.838539912" Dec 07 16:16:00 crc kubenswrapper[4716]: I1207 16:16:00.341635 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-64d95d5f5f-6t4rp" Dec 07 16:16:22 crc kubenswrapper[4716]: I1207 16:16:22.761368 4716 patch_prober.go:28] interesting pod/machine-config-daemon-zcxxp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 16:16:22 crc kubenswrapper[4716]: I1207 16:16:22.761895 4716 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 16:16:22 crc kubenswrapper[4716]: I1207 16:16:22.761940 4716 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" Dec 07 16:16:22 crc kubenswrapper[4716]: I1207 16:16:22.762519 4716 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8ec0c9fb49e1eaea181d9e297f07686725c05f58dac3a79c7a2ee3f00f979908"} pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 07 16:16:22 crc kubenswrapper[4716]: I1207 16:16:22.762576 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" containerName="machine-config-daemon" containerID="cri-o://8ec0c9fb49e1eaea181d9e297f07686725c05f58dac3a79c7a2ee3f00f979908" gracePeriod=600 Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.320777 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-m957c"] Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.322383 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-m957c" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.329715 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-pp94z" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.337299 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-m957c"] Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.346396 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6c677c69b-4nrlz"] Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.348054 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-4nrlz" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.351477 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-697fb699cf-zts6p"] Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.353309 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-zts6p" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.355355 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-46qd7" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.355647 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-fxw8p" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.372431 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6c677c69b-4nrlz"] Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.378338 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-697fb699cf-zts6p"] Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.395680 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-xqnzw"] Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.396864 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-xqnzw" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.400978 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-5697bb5779-rcmz5"] Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.402357 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-rcmz5" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.402672 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-m9vkp" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.409064 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-tpk2s" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.420146 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-xqnzw"] Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.426712 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zrw6n\" (UniqueName: \"kubernetes.io/projected/8765c8b2-a50b-4417-adbe-1174dcdfe172-kube-api-access-zrw6n\") pod \"barbican-operator-controller-manager-7d9dfd778-m957c\" (UID: \"8765c8b2-a50b-4417-adbe-1174dcdfe172\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-m957c" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.438011 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-nx5mv"] Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.438919 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-nx5mv" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.442495 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-kw949" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.442831 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-5697bb5779-rcmz5"] Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.457137 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-78d48bff9d-pxs5r"] Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.458262 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-pxs5r" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.463410 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-r8bm4" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.463754 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.480426 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-nx5mv"] Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.498156 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-967d97867-hx48c"] Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.499408 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-967d97867-hx48c" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.502169 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-q2lcz" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.528035 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-967d97867-hx48c"] Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.528184 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zrw6n\" (UniqueName: \"kubernetes.io/projected/8765c8b2-a50b-4417-adbe-1174dcdfe172-kube-api-access-zrw6n\") pod \"barbican-operator-controller-manager-7d9dfd778-m957c\" (UID: \"8765c8b2-a50b-4417-adbe-1174dcdfe172\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-m957c" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.528224 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cg6g9\" (UniqueName: \"kubernetes.io/projected/79562d20-950f-428e-ac3a-f78979053266-kube-api-access-cg6g9\") pod \"designate-operator-controller-manager-697fb699cf-zts6p\" (UID: \"79562d20-950f-428e-ac3a-f78979053266\") " pod="openstack-operators/designate-operator-controller-manager-697fb699cf-zts6p" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.528257 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9dqf\" (UniqueName: \"kubernetes.io/projected/62368956-27d2-41fa-a2b2-0fb49d869f11-kube-api-access-h9dqf\") pod \"cinder-operator-controller-manager-6c677c69b-4nrlz\" (UID: \"62368956-27d2-41fa-a2b2-0fb49d869f11\") " pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-4nrlz" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.528289 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-47x2b\" (UniqueName: \"kubernetes.io/projected/6f24594b-ebe7-4518-a067-45891924abe5-kube-api-access-47x2b\") pod \"glance-operator-controller-manager-5697bb5779-rcmz5\" (UID: \"6f24594b-ebe7-4518-a067-45891924abe5\") " pod="openstack-operators/glance-operator-controller-manager-5697bb5779-rcmz5" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.528314 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ln9p9\" (UniqueName: \"kubernetes.io/projected/d9d5c8b8-240b-4c88-ad6e-b04881842f0b-kube-api-access-ln9p9\") pod \"heat-operator-controller-manager-5f64f6f8bb-xqnzw\" (UID: \"d9d5c8b8-240b-4c88-ad6e-b04881842f0b\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-xqnzw" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.537052 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-78d48bff9d-pxs5r"] Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.554302 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-wctdp"] Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.555373 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-wctdp" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.585966 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zrw6n\" (UniqueName: \"kubernetes.io/projected/8765c8b2-a50b-4417-adbe-1174dcdfe172-kube-api-access-zrw6n\") pod \"barbican-operator-controller-manager-7d9dfd778-m957c\" (UID: \"8765c8b2-a50b-4417-adbe-1174dcdfe172\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-m957c" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.587139 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-mcshx" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.594859 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-5b5fd79c9c-kbzpp"] Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.596872 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-kbzpp" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.606624 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-jjvmd" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.620991 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-wctdp"] Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.630919 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5b5fd79c9c-kbzpp"] Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.631400 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q6ldw\" (UniqueName: \"kubernetes.io/projected/d3ef7831-ac99-410d-90bf-42d87042964b-kube-api-access-q6ldw\") pod \"manila-operator-controller-manager-5b5fd79c9c-kbzpp\" (UID: \"d3ef7831-ac99-410d-90bf-42d87042964b\") " pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-kbzpp" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.631465 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/01686458-debd-419a-90ec-cd27cc6953ec-cert\") pod \"infra-operator-controller-manager-78d48bff9d-pxs5r\" (UID: \"01686458-debd-419a-90ec-cd27cc6953ec\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-pxs5r" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.631488 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rz4wr\" (UniqueName: \"kubernetes.io/projected/01686458-debd-419a-90ec-cd27cc6953ec-kube-api-access-rz4wr\") pod \"infra-operator-controller-manager-78d48bff9d-pxs5r\" (UID: \"01686458-debd-419a-90ec-cd27cc6953ec\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-pxs5r" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.631516 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rl89l\" (UniqueName: \"kubernetes.io/projected/20067e63-712d-4f94-9019-627219d4299f-kube-api-access-rl89l\") pod \"horizon-operator-controller-manager-68c6d99b8f-nx5mv\" (UID: \"20067e63-712d-4f94-9019-627219d4299f\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-nx5mv" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.631545 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cg6g9\" (UniqueName: \"kubernetes.io/projected/79562d20-950f-428e-ac3a-f78979053266-kube-api-access-cg6g9\") pod \"designate-operator-controller-manager-697fb699cf-zts6p\" (UID: \"79562d20-950f-428e-ac3a-f78979053266\") " pod="openstack-operators/designate-operator-controller-manager-697fb699cf-zts6p" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.631575 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9dqf\" (UniqueName: \"kubernetes.io/projected/62368956-27d2-41fa-a2b2-0fb49d869f11-kube-api-access-h9dqf\") pod \"cinder-operator-controller-manager-6c677c69b-4nrlz\" (UID: \"62368956-27d2-41fa-a2b2-0fb49d869f11\") " pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-4nrlz" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.631607 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-47x2b\" (UniqueName: \"kubernetes.io/projected/6f24594b-ebe7-4518-a067-45891924abe5-kube-api-access-47x2b\") pod \"glance-operator-controller-manager-5697bb5779-rcmz5\" (UID: \"6f24594b-ebe7-4518-a067-45891924abe5\") " pod="openstack-operators/glance-operator-controller-manager-5697bb5779-rcmz5" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.631637 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tghx4\" (UniqueName: \"kubernetes.io/projected/2babb856-a5c2-4b60-a6bc-00d3f510f014-kube-api-access-tghx4\") pod \"keystone-operator-controller-manager-7765d96ddf-wctdp\" (UID: \"2babb856-a5c2-4b60-a6bc-00d3f510f014\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-wctdp" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.631669 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ln9p9\" (UniqueName: \"kubernetes.io/projected/d9d5c8b8-240b-4c88-ad6e-b04881842f0b-kube-api-access-ln9p9\") pod \"heat-operator-controller-manager-5f64f6f8bb-xqnzw\" (UID: \"d9d5c8b8-240b-4c88-ad6e-b04881842f0b\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-xqnzw" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.631697 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2s4nk\" (UniqueName: \"kubernetes.io/projected/a929d511-1d07-429d-b302-8843a6834f52-kube-api-access-2s4nk\") pod \"ironic-operator-controller-manager-967d97867-hx48c\" (UID: \"a929d511-1d07-429d-b302-8843a6834f52\") " pod="openstack-operators/ironic-operator-controller-manager-967d97867-hx48c" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.638895 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-79c8c4686c-9dsmr"] Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.640239 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-srq27"] Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.641300 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-srq27" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.641825 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-9dsmr" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.654118 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-2zbw5" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.656570 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-m957c" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.678745 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cg6g9\" (UniqueName: \"kubernetes.io/projected/79562d20-950f-428e-ac3a-f78979053266-kube-api-access-cg6g9\") pod \"designate-operator-controller-manager-697fb699cf-zts6p\" (UID: \"79562d20-950f-428e-ac3a-f78979053266\") " pod="openstack-operators/designate-operator-controller-manager-697fb699cf-zts6p" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.679145 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-ww867" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.675070 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-79c8c4686c-9dsmr"] Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.691066 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-srq27"] Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.691111 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-z7lcv"] Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.692176 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-z7lcv" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.698193 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-zts6p" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.699291 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-jkzmh" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.707240 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ln9p9\" (UniqueName: \"kubernetes.io/projected/d9d5c8b8-240b-4c88-ad6e-b04881842f0b-kube-api-access-ln9p9\") pod \"heat-operator-controller-manager-5f64f6f8bb-xqnzw\" (UID: \"d9d5c8b8-240b-4c88-ad6e-b04881842f0b\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-xqnzw" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.710182 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-z7lcv"] Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.711616 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9dqf\" (UniqueName: \"kubernetes.io/projected/62368956-27d2-41fa-a2b2-0fb49d869f11-kube-api-access-h9dqf\") pod \"cinder-operator-controller-manager-6c677c69b-4nrlz\" (UID: \"62368956-27d2-41fa-a2b2-0fb49d869f11\") " pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-4nrlz" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.717449 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-xqnzw" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.723107 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-47x2b\" (UniqueName: \"kubernetes.io/projected/6f24594b-ebe7-4518-a067-45891924abe5-kube-api-access-47x2b\") pod \"glance-operator-controller-manager-5697bb5779-rcmz5\" (UID: \"6f24594b-ebe7-4518-a067-45891924abe5\") " pod="openstack-operators/glance-operator-controller-manager-5697bb5779-rcmz5" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.725226 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-kkdgl"] Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.726304 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-kkdgl" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.733605 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-kkdgl"] Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.734556 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-rcmz5" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.734699 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rl89l\" (UniqueName: \"kubernetes.io/projected/20067e63-712d-4f94-9019-627219d4299f-kube-api-access-rl89l\") pod \"horizon-operator-controller-manager-68c6d99b8f-nx5mv\" (UID: \"20067e63-712d-4f94-9019-627219d4299f\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-nx5mv" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.734782 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tghx4\" (UniqueName: \"kubernetes.io/projected/2babb856-a5c2-4b60-a6bc-00d3f510f014-kube-api-access-tghx4\") pod \"keystone-operator-controller-manager-7765d96ddf-wctdp\" (UID: \"2babb856-a5c2-4b60-a6bc-00d3f510f014\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-wctdp" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.734824 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2s4nk\" (UniqueName: \"kubernetes.io/projected/a929d511-1d07-429d-b302-8843a6834f52-kube-api-access-2s4nk\") pod \"ironic-operator-controller-manager-967d97867-hx48c\" (UID: \"a929d511-1d07-429d-b302-8843a6834f52\") " pod="openstack-operators/ironic-operator-controller-manager-967d97867-hx48c" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.734870 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q6ldw\" (UniqueName: \"kubernetes.io/projected/d3ef7831-ac99-410d-90bf-42d87042964b-kube-api-access-q6ldw\") pod \"manila-operator-controller-manager-5b5fd79c9c-kbzpp\" (UID: \"d3ef7831-ac99-410d-90bf-42d87042964b\") " pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-kbzpp" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.734888 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/01686458-debd-419a-90ec-cd27cc6953ec-cert\") pod \"infra-operator-controller-manager-78d48bff9d-pxs5r\" (UID: \"01686458-debd-419a-90ec-cd27cc6953ec\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-pxs5r" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.734903 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rz4wr\" (UniqueName: \"kubernetes.io/projected/01686458-debd-419a-90ec-cd27cc6953ec-kube-api-access-rz4wr\") pod \"infra-operator-controller-manager-78d48bff9d-pxs5r\" (UID: \"01686458-debd-419a-90ec-cd27cc6953ec\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-pxs5r" Dec 07 16:16:23 crc kubenswrapper[4716]: E1207 16:16:23.735345 4716 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 07 16:16:23 crc kubenswrapper[4716]: E1207 16:16:23.735419 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/01686458-debd-419a-90ec-cd27cc6953ec-cert podName:01686458-debd-419a-90ec-cd27cc6953ec nodeName:}" failed. No retries permitted until 2025-12-07 16:16:24.235397876 +0000 UTC m=+846.925682788 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/01686458-debd-419a-90ec-cd27cc6953ec-cert") pod "infra-operator-controller-manager-78d48bff9d-pxs5r" (UID: "01686458-debd-419a-90ec-cd27cc6953ec") : secret "infra-operator-webhook-server-cert" not found Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.745336 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-wft2h" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.752051 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fdkpkx"] Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.753971 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fdkpkx" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.766327 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-h95ls"] Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.767282 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-h95ls" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.788857 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fdkpkx"] Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.792976 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-lgmbb" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.793998 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.797621 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-w4kkh" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.800520 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rl89l\" (UniqueName: \"kubernetes.io/projected/20067e63-712d-4f94-9019-627219d4299f-kube-api-access-rl89l\") pod \"horizon-operator-controller-manager-68c6d99b8f-nx5mv\" (UID: \"20067e63-712d-4f94-9019-627219d4299f\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-nx5mv" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.810305 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-8xrch"] Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.811283 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-8xrch" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.816691 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rz4wr\" (UniqueName: \"kubernetes.io/projected/01686458-debd-419a-90ec-cd27cc6953ec-kube-api-access-rz4wr\") pod \"infra-operator-controller-manager-78d48bff9d-pxs5r\" (UID: \"01686458-debd-419a-90ec-cd27cc6953ec\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-pxs5r" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.821206 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-5mj84" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.828859 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tghx4\" (UniqueName: \"kubernetes.io/projected/2babb856-a5c2-4b60-a6bc-00d3f510f014-kube-api-access-tghx4\") pod \"keystone-operator-controller-manager-7765d96ddf-wctdp\" (UID: \"2babb856-a5c2-4b60-a6bc-00d3f510f014\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-wctdp" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.828920 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-9d58d64bc-p2hbf"] Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.829920 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-p2hbf" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.839280 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-h95ls"] Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.839799 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c5lvc\" (UniqueName: \"kubernetes.io/projected/34e0d69c-2f36-4ba6-a4e3-114cabe016c2-kube-api-access-c5lvc\") pod \"openstack-baremetal-operator-controller-manager-84b575879fdkpkx\" (UID: \"34e0d69c-2f36-4ba6-a4e3-114cabe016c2\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fdkpkx" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.839840 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k28gb\" (UniqueName: \"kubernetes.io/projected/0f05e35d-f818-4f65-bac6-f6a22006627e-kube-api-access-k28gb\") pod \"placement-operator-controller-manager-78f8948974-8xrch\" (UID: \"0f05e35d-f818-4f65-bac6-f6a22006627e\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-8xrch" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.839881 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zd9gl\" (UniqueName: \"kubernetes.io/projected/5be39543-890e-479c-a041-d864922e038f-kube-api-access-zd9gl\") pod \"mariadb-operator-controller-manager-79c8c4686c-9dsmr\" (UID: \"5be39543-890e-479c-a041-d864922e038f\") " pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-9dsmr" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.839919 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pm9xg\" (UniqueName: \"kubernetes.io/projected/69c2e54f-9af6-44e8-abdd-226a29b64da6-kube-api-access-pm9xg\") pod \"nova-operator-controller-manager-697bc559fc-z7lcv\" (UID: \"69c2e54f-9af6-44e8-abdd-226a29b64da6\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-z7lcv" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.839937 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/34e0d69c-2f36-4ba6-a4e3-114cabe016c2-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879fdkpkx\" (UID: \"34e0d69c-2f36-4ba6-a4e3-114cabe016c2\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fdkpkx" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.839958 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mj8z2\" (UniqueName: \"kubernetes.io/projected/57c9705a-9bf1-45b4-aefb-d224a9d72a4a-kube-api-access-mj8z2\") pod \"octavia-operator-controller-manager-998648c74-kkdgl\" (UID: \"57c9705a-9bf1-45b4-aefb-d224a9d72a4a\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-kkdgl" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.839984 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kj2xt\" (UniqueName: \"kubernetes.io/projected/d9d5a1f4-edb7-4630-8907-5d29c4678f24-kube-api-access-kj2xt\") pod \"swift-operator-controller-manager-9d58d64bc-p2hbf\" (UID: \"d9d5a1f4-edb7-4630-8907-5d29c4678f24\") " pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-p2hbf" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.840005 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqcsd\" (UniqueName: \"kubernetes.io/projected/1bf3b0b2-ea38-4fa3-b07c-4467dc111476-kube-api-access-cqcsd\") pod \"ovn-operator-controller-manager-b6456fdb6-h95ls\" (UID: \"1bf3b0b2-ea38-4fa3-b07c-4467dc111476\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-h95ls" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.840041 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cmqd2\" (UniqueName: \"kubernetes.io/projected/463e6bed-ce60-4e35-98a3-cd837b9066f2-kube-api-access-cmqd2\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-srq27\" (UID: \"463e6bed-ce60-4e35-98a3-cd837b9066f2\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-srq27" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.850204 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-9d58d64bc-p2hbf"] Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.851497 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-8xrch"] Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.853429 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2s4nk\" (UniqueName: \"kubernetes.io/projected/a929d511-1d07-429d-b302-8843a6834f52-kube-api-access-2s4nk\") pod \"ironic-operator-controller-manager-967d97867-hx48c\" (UID: \"a929d511-1d07-429d-b302-8843a6834f52\") " pod="openstack-operators/ironic-operator-controller-manager-967d97867-hx48c" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.853883 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-29mxb" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.856774 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q6ldw\" (UniqueName: \"kubernetes.io/projected/d3ef7831-ac99-410d-90bf-42d87042964b-kube-api-access-q6ldw\") pod \"manila-operator-controller-manager-5b5fd79c9c-kbzpp\" (UID: \"d3ef7831-ac99-410d-90bf-42d87042964b\") " pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-kbzpp" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.864594 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-58d5ff84df-ggcr7"] Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.865427 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-58d5ff84df-ggcr7"] Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.865498 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-ggcr7" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.876198 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-jnn7r" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.897611 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-4dnl8"] Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.898713 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-4dnl8" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.901398 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-fbq86" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.919180 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-4dnl8"] Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.919584 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-wctdp" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.935097 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-kbzpp" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.940239 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-667bd8d554-8zwtk"] Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.983056 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cmqd2\" (UniqueName: \"kubernetes.io/projected/463e6bed-ce60-4e35-98a3-cd837b9066f2-kube-api-access-cmqd2\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-srq27\" (UID: \"463e6bed-ce60-4e35-98a3-cd837b9066f2\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-srq27" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.985792 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-4nrlz" Dec 07 16:16:23 crc kubenswrapper[4716]: I1207 16:16:23.990799 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-8zwtk" Dec 07 16:16:24 crc kubenswrapper[4716]: I1207 16:16:23.996019 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c5lvc\" (UniqueName: \"kubernetes.io/projected/34e0d69c-2f36-4ba6-a4e3-114cabe016c2-kube-api-access-c5lvc\") pod \"openstack-baremetal-operator-controller-manager-84b575879fdkpkx\" (UID: \"34e0d69c-2f36-4ba6-a4e3-114cabe016c2\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fdkpkx" Dec 07 16:16:24 crc kubenswrapper[4716]: I1207 16:16:23.996071 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k28gb\" (UniqueName: \"kubernetes.io/projected/0f05e35d-f818-4f65-bac6-f6a22006627e-kube-api-access-k28gb\") pod \"placement-operator-controller-manager-78f8948974-8xrch\" (UID: \"0f05e35d-f818-4f65-bac6-f6a22006627e\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-8xrch" Dec 07 16:16:24 crc kubenswrapper[4716]: I1207 16:16:23.996138 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zd9gl\" (UniqueName: \"kubernetes.io/projected/5be39543-890e-479c-a041-d864922e038f-kube-api-access-zd9gl\") pod \"mariadb-operator-controller-manager-79c8c4686c-9dsmr\" (UID: \"5be39543-890e-479c-a041-d864922e038f\") " pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-9dsmr" Dec 07 16:16:24 crc kubenswrapper[4716]: I1207 16:16:23.996187 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pm9xg\" (UniqueName: \"kubernetes.io/projected/69c2e54f-9af6-44e8-abdd-226a29b64da6-kube-api-access-pm9xg\") pod \"nova-operator-controller-manager-697bc559fc-z7lcv\" (UID: \"69c2e54f-9af6-44e8-abdd-226a29b64da6\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-z7lcv" Dec 07 16:16:24 crc kubenswrapper[4716]: I1207 16:16:23.996212 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/34e0d69c-2f36-4ba6-a4e3-114cabe016c2-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879fdkpkx\" (UID: \"34e0d69c-2f36-4ba6-a4e3-114cabe016c2\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fdkpkx" Dec 07 16:16:24 crc kubenswrapper[4716]: I1207 16:16:23.996240 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mj8z2\" (UniqueName: \"kubernetes.io/projected/57c9705a-9bf1-45b4-aefb-d224a9d72a4a-kube-api-access-mj8z2\") pod \"octavia-operator-controller-manager-998648c74-kkdgl\" (UID: \"57c9705a-9bf1-45b4-aefb-d224a9d72a4a\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-kkdgl" Dec 07 16:16:24 crc kubenswrapper[4716]: I1207 16:16:23.996265 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kj2xt\" (UniqueName: \"kubernetes.io/projected/d9d5a1f4-edb7-4630-8907-5d29c4678f24-kube-api-access-kj2xt\") pod \"swift-operator-controller-manager-9d58d64bc-p2hbf\" (UID: \"d9d5a1f4-edb7-4630-8907-5d29c4678f24\") " pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-p2hbf" Dec 07 16:16:24 crc kubenswrapper[4716]: I1207 16:16:23.996298 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqcsd\" (UniqueName: \"kubernetes.io/projected/1bf3b0b2-ea38-4fa3-b07c-4467dc111476-kube-api-access-cqcsd\") pod \"ovn-operator-controller-manager-b6456fdb6-h95ls\" (UID: \"1bf3b0b2-ea38-4fa3-b07c-4467dc111476\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-h95ls" Dec 07 16:16:24 crc kubenswrapper[4716]: E1207 16:16:23.997192 4716 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 07 16:16:24 crc kubenswrapper[4716]: E1207 16:16:23.997237 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/34e0d69c-2f36-4ba6-a4e3-114cabe016c2-cert podName:34e0d69c-2f36-4ba6-a4e3-114cabe016c2 nodeName:}" failed. No retries permitted until 2025-12-07 16:16:24.497222021 +0000 UTC m=+847.187506933 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/34e0d69c-2f36-4ba6-a4e3-114cabe016c2-cert") pod "openstack-baremetal-operator-controller-manager-84b575879fdkpkx" (UID: "34e0d69c-2f36-4ba6-a4e3-114cabe016c2") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 07 16:16:24 crc kubenswrapper[4716]: I1207 16:16:24.002180 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-2w8ww" Dec 07 16:16:24 crc kubenswrapper[4716]: I1207 16:16:24.017224 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-667bd8d554-8zwtk"] Dec 07 16:16:24 crc kubenswrapper[4716]: I1207 16:16:24.023833 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cmqd2\" (UniqueName: \"kubernetes.io/projected/463e6bed-ce60-4e35-98a3-cd837b9066f2-kube-api-access-cmqd2\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-srq27\" (UID: \"463e6bed-ce60-4e35-98a3-cd837b9066f2\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-srq27" Dec 07 16:16:24 crc kubenswrapper[4716]: I1207 16:16:24.070903 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-nx5mv" Dec 07 16:16:24 crc kubenswrapper[4716]: I1207 16:16:24.077624 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mj8z2\" (UniqueName: \"kubernetes.io/projected/57c9705a-9bf1-45b4-aefb-d224a9d72a4a-kube-api-access-mj8z2\") pod \"octavia-operator-controller-manager-998648c74-kkdgl\" (UID: \"57c9705a-9bf1-45b4-aefb-d224a9d72a4a\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-kkdgl" Dec 07 16:16:24 crc kubenswrapper[4716]: I1207 16:16:24.080416 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kj2xt\" (UniqueName: \"kubernetes.io/projected/d9d5a1f4-edb7-4630-8907-5d29c4678f24-kube-api-access-kj2xt\") pod \"swift-operator-controller-manager-9d58d64bc-p2hbf\" (UID: \"d9d5a1f4-edb7-4630-8907-5d29c4678f24\") " pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-p2hbf" Dec 07 16:16:24 crc kubenswrapper[4716]: I1207 16:16:24.081377 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqcsd\" (UniqueName: \"kubernetes.io/projected/1bf3b0b2-ea38-4fa3-b07c-4467dc111476-kube-api-access-cqcsd\") pod \"ovn-operator-controller-manager-b6456fdb6-h95ls\" (UID: \"1bf3b0b2-ea38-4fa3-b07c-4467dc111476\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-h95ls" Dec 07 16:16:24 crc kubenswrapper[4716]: I1207 16:16:24.092618 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-h95ls" Dec 07 16:16:24 crc kubenswrapper[4716]: I1207 16:16:24.096679 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k28gb\" (UniqueName: \"kubernetes.io/projected/0f05e35d-f818-4f65-bac6-f6a22006627e-kube-api-access-k28gb\") pod \"placement-operator-controller-manager-78f8948974-8xrch\" (UID: \"0f05e35d-f818-4f65-bac6-f6a22006627e\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-8xrch" Dec 07 16:16:24 crc kubenswrapper[4716]: I1207 16:16:24.102135 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8gskr\" (UniqueName: \"kubernetes.io/projected/2ef47713-a73c-4c63-8fa0-ffc530832285-kube-api-access-8gskr\") pod \"watcher-operator-controller-manager-667bd8d554-8zwtk\" (UID: \"2ef47713-a73c-4c63-8fa0-ffc530832285\") " pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-8zwtk" Dec 07 16:16:24 crc kubenswrapper[4716]: I1207 16:16:24.102191 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zsmst\" (UniqueName: \"kubernetes.io/projected/e4c2f2f6-3285-4f48-a428-b2065f33c046-kube-api-access-zsmst\") pod \"telemetry-operator-controller-manager-58d5ff84df-ggcr7\" (UID: \"e4c2f2f6-3285-4f48-a428-b2065f33c046\") " pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-ggcr7" Dec 07 16:16:24 crc kubenswrapper[4716]: I1207 16:16:24.102237 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5q6j7\" (UniqueName: \"kubernetes.io/projected/0c6a49cd-1cfe-49c8-a068-f85c735b46b1-kube-api-access-5q6j7\") pod \"test-operator-controller-manager-5854674fcc-4dnl8\" (UID: \"0c6a49cd-1cfe-49c8-a068-f85c735b46b1\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-4dnl8" Dec 07 16:16:24 crc kubenswrapper[4716]: I1207 16:16:24.105693 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c5lvc\" (UniqueName: \"kubernetes.io/projected/34e0d69c-2f36-4ba6-a4e3-114cabe016c2-kube-api-access-c5lvc\") pod \"openstack-baremetal-operator-controller-manager-84b575879fdkpkx\" (UID: \"34e0d69c-2f36-4ba6-a4e3-114cabe016c2\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fdkpkx" Dec 07 16:16:24 crc kubenswrapper[4716]: I1207 16:16:24.115656 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zd9gl\" (UniqueName: \"kubernetes.io/projected/5be39543-890e-479c-a041-d864922e038f-kube-api-access-zd9gl\") pod \"mariadb-operator-controller-manager-79c8c4686c-9dsmr\" (UID: \"5be39543-890e-479c-a041-d864922e038f\") " pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-9dsmr" Dec 07 16:16:24 crc kubenswrapper[4716]: I1207 16:16:24.116663 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pm9xg\" (UniqueName: \"kubernetes.io/projected/69c2e54f-9af6-44e8-abdd-226a29b64da6-kube-api-access-pm9xg\") pod \"nova-operator-controller-manager-697bc559fc-z7lcv\" (UID: \"69c2e54f-9af6-44e8-abdd-226a29b64da6\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-z7lcv" Dec 07 16:16:24 crc kubenswrapper[4716]: I1207 16:16:24.121516 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-p2hbf" Dec 07 16:16:24 crc kubenswrapper[4716]: I1207 16:16:24.130124 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-967d97867-hx48c" Dec 07 16:16:24 crc kubenswrapper[4716]: I1207 16:16:24.155383 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-kkdgl" Dec 07 16:16:24 crc kubenswrapper[4716]: I1207 16:16:24.155791 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7db7d7886d-z6n25"] Dec 07 16:16:24 crc kubenswrapper[4716]: I1207 16:16:24.156673 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-7db7d7886d-z6n25" Dec 07 16:16:24 crc kubenswrapper[4716]: I1207 16:16:24.160429 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-7bdkt" Dec 07 16:16:24 crc kubenswrapper[4716]: I1207 16:16:24.160614 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Dec 07 16:16:24 crc kubenswrapper[4716]: I1207 16:16:24.160714 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Dec 07 16:16:24 crc kubenswrapper[4716]: I1207 16:16:24.178168 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7db7d7886d-z6n25"] Dec 07 16:16:24 crc kubenswrapper[4716]: I1207 16:16:24.196135 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-7mjmm"] Dec 07 16:16:24 crc kubenswrapper[4716]: I1207 16:16:24.197181 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-7mjmm" Dec 07 16:16:24 crc kubenswrapper[4716]: I1207 16:16:24.200415 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-7mjmm"] Dec 07 16:16:24 crc kubenswrapper[4716]: I1207 16:16:24.206682 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-m78tr" Dec 07 16:16:24 crc kubenswrapper[4716]: I1207 16:16:24.223722 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5csl5\" (UniqueName: \"kubernetes.io/projected/9d840ba3-d5ef-48fc-87b2-6af4d164c9ba-kube-api-access-5csl5\") pod \"openstack-operator-controller-manager-7db7d7886d-z6n25\" (UID: \"9d840ba3-d5ef-48fc-87b2-6af4d164c9ba\") " pod="openstack-operators/openstack-operator-controller-manager-7db7d7886d-z6n25" Dec 07 16:16:24 crc kubenswrapper[4716]: I1207 16:16:24.223770 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/9d840ba3-d5ef-48fc-87b2-6af4d164c9ba-webhook-certs\") pod \"openstack-operator-controller-manager-7db7d7886d-z6n25\" (UID: \"9d840ba3-d5ef-48fc-87b2-6af4d164c9ba\") " pod="openstack-operators/openstack-operator-controller-manager-7db7d7886d-z6n25" Dec 07 16:16:24 crc kubenswrapper[4716]: I1207 16:16:24.223803 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8gskr\" (UniqueName: \"kubernetes.io/projected/2ef47713-a73c-4c63-8fa0-ffc530832285-kube-api-access-8gskr\") pod \"watcher-operator-controller-manager-667bd8d554-8zwtk\" (UID: \"2ef47713-a73c-4c63-8fa0-ffc530832285\") " pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-8zwtk" Dec 07 16:16:24 crc kubenswrapper[4716]: I1207 16:16:24.223825 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zsmst\" (UniqueName: \"kubernetes.io/projected/e4c2f2f6-3285-4f48-a428-b2065f33c046-kube-api-access-zsmst\") pod \"telemetry-operator-controller-manager-58d5ff84df-ggcr7\" (UID: \"e4c2f2f6-3285-4f48-a428-b2065f33c046\") " pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-ggcr7" Dec 07 16:16:24 crc kubenswrapper[4716]: I1207 16:16:24.223910 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hnrd\" (UniqueName: \"kubernetes.io/projected/79aa5696-34a1-45cf-b965-56f64fe63f4e-kube-api-access-4hnrd\") pod \"rabbitmq-cluster-operator-manager-668c99d594-7mjmm\" (UID: \"79aa5696-34a1-45cf-b965-56f64fe63f4e\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-7mjmm" Dec 07 16:16:24 crc kubenswrapper[4716]: I1207 16:16:24.223977 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5q6j7\" (UniqueName: \"kubernetes.io/projected/0c6a49cd-1cfe-49c8-a068-f85c735b46b1-kube-api-access-5q6j7\") pod \"test-operator-controller-manager-5854674fcc-4dnl8\" (UID: \"0c6a49cd-1cfe-49c8-a068-f85c735b46b1\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-4dnl8" Dec 07 16:16:24 crc kubenswrapper[4716]: I1207 16:16:24.224047 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9d840ba3-d5ef-48fc-87b2-6af4d164c9ba-metrics-certs\") pod \"openstack-operator-controller-manager-7db7d7886d-z6n25\" (UID: \"9d840ba3-d5ef-48fc-87b2-6af4d164c9ba\") " pod="openstack-operators/openstack-operator-controller-manager-7db7d7886d-z6n25" Dec 07 16:16:24 crc kubenswrapper[4716]: I1207 16:16:24.246127 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zsmst\" (UniqueName: \"kubernetes.io/projected/e4c2f2f6-3285-4f48-a428-b2065f33c046-kube-api-access-zsmst\") pod \"telemetry-operator-controller-manager-58d5ff84df-ggcr7\" (UID: \"e4c2f2f6-3285-4f48-a428-b2065f33c046\") " pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-ggcr7" Dec 07 16:16:24 crc kubenswrapper[4716]: I1207 16:16:24.255389 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8gskr\" (UniqueName: \"kubernetes.io/projected/2ef47713-a73c-4c63-8fa0-ffc530832285-kube-api-access-8gskr\") pod \"watcher-operator-controller-manager-667bd8d554-8zwtk\" (UID: \"2ef47713-a73c-4c63-8fa0-ffc530832285\") " pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-8zwtk" Dec 07 16:16:24 crc kubenswrapper[4716]: I1207 16:16:24.255423 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5q6j7\" (UniqueName: \"kubernetes.io/projected/0c6a49cd-1cfe-49c8-a068-f85c735b46b1-kube-api-access-5q6j7\") pod \"test-operator-controller-manager-5854674fcc-4dnl8\" (UID: \"0c6a49cd-1cfe-49c8-a068-f85c735b46b1\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-4dnl8" Dec 07 16:16:24 crc kubenswrapper[4716]: I1207 16:16:24.255998 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-4dnl8" Dec 07 16:16:24 crc kubenswrapper[4716]: I1207 16:16:24.265629 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-srq27" Dec 07 16:16:24 crc kubenswrapper[4716]: I1207 16:16:24.275797 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-8zwtk" Dec 07 16:16:24 crc kubenswrapper[4716]: I1207 16:16:24.286976 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-9dsmr" Dec 07 16:16:24 crc kubenswrapper[4716]: I1207 16:16:24.324989 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hnrd\" (UniqueName: \"kubernetes.io/projected/79aa5696-34a1-45cf-b965-56f64fe63f4e-kube-api-access-4hnrd\") pod \"rabbitmq-cluster-operator-manager-668c99d594-7mjmm\" (UID: \"79aa5696-34a1-45cf-b965-56f64fe63f4e\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-7mjmm" Dec 07 16:16:24 crc kubenswrapper[4716]: I1207 16:16:24.325063 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/01686458-debd-419a-90ec-cd27cc6953ec-cert\") pod \"infra-operator-controller-manager-78d48bff9d-pxs5r\" (UID: \"01686458-debd-419a-90ec-cd27cc6953ec\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-pxs5r" Dec 07 16:16:24 crc kubenswrapper[4716]: I1207 16:16:24.325148 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9d840ba3-d5ef-48fc-87b2-6af4d164c9ba-metrics-certs\") pod \"openstack-operator-controller-manager-7db7d7886d-z6n25\" (UID: \"9d840ba3-d5ef-48fc-87b2-6af4d164c9ba\") " pod="openstack-operators/openstack-operator-controller-manager-7db7d7886d-z6n25" Dec 07 16:16:24 crc kubenswrapper[4716]: I1207 16:16:24.325215 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5csl5\" (UniqueName: \"kubernetes.io/projected/9d840ba3-d5ef-48fc-87b2-6af4d164c9ba-kube-api-access-5csl5\") pod \"openstack-operator-controller-manager-7db7d7886d-z6n25\" (UID: \"9d840ba3-d5ef-48fc-87b2-6af4d164c9ba\") " pod="openstack-operators/openstack-operator-controller-manager-7db7d7886d-z6n25" Dec 07 16:16:24 crc kubenswrapper[4716]: I1207 16:16:24.325238 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/9d840ba3-d5ef-48fc-87b2-6af4d164c9ba-webhook-certs\") pod \"openstack-operator-controller-manager-7db7d7886d-z6n25\" (UID: \"9d840ba3-d5ef-48fc-87b2-6af4d164c9ba\") " pod="openstack-operators/openstack-operator-controller-manager-7db7d7886d-z6n25" Dec 07 16:16:24 crc kubenswrapper[4716]: E1207 16:16:24.326509 4716 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 07 16:16:24 crc kubenswrapper[4716]: E1207 16:16:24.326626 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9d840ba3-d5ef-48fc-87b2-6af4d164c9ba-metrics-certs podName:9d840ba3-d5ef-48fc-87b2-6af4d164c9ba nodeName:}" failed. No retries permitted until 2025-12-07 16:16:24.826591307 +0000 UTC m=+847.516876409 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9d840ba3-d5ef-48fc-87b2-6af4d164c9ba-metrics-certs") pod "openstack-operator-controller-manager-7db7d7886d-z6n25" (UID: "9d840ba3-d5ef-48fc-87b2-6af4d164c9ba") : secret "metrics-server-cert" not found Dec 07 16:16:24 crc kubenswrapper[4716]: E1207 16:16:24.326823 4716 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 07 16:16:24 crc kubenswrapper[4716]: E1207 16:16:24.326870 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9d840ba3-d5ef-48fc-87b2-6af4d164c9ba-webhook-certs podName:9d840ba3-d5ef-48fc-87b2-6af4d164c9ba nodeName:}" failed. No retries permitted until 2025-12-07 16:16:24.826853375 +0000 UTC m=+847.517138287 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/9d840ba3-d5ef-48fc-87b2-6af4d164c9ba-webhook-certs") pod "openstack-operator-controller-manager-7db7d7886d-z6n25" (UID: "9d840ba3-d5ef-48fc-87b2-6af4d164c9ba") : secret "webhook-server-cert" not found Dec 07 16:16:24 crc kubenswrapper[4716]: E1207 16:16:24.327266 4716 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 07 16:16:24 crc kubenswrapper[4716]: E1207 16:16:24.327305 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/01686458-debd-419a-90ec-cd27cc6953ec-cert podName:01686458-debd-419a-90ec-cd27cc6953ec nodeName:}" failed. No retries permitted until 2025-12-07 16:16:25.327293028 +0000 UTC m=+848.017578160 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/01686458-debd-419a-90ec-cd27cc6953ec-cert") pod "infra-operator-controller-manager-78d48bff9d-pxs5r" (UID: "01686458-debd-419a-90ec-cd27cc6953ec") : secret "infra-operator-webhook-server-cert" not found Dec 07 16:16:24 crc kubenswrapper[4716]: I1207 16:16:24.337630 4716 generic.go:334] "Generic (PLEG): container finished" podID="c15b59eb-565d-4556-a4ce-75afdf159dc8" containerID="8ec0c9fb49e1eaea181d9e297f07686725c05f58dac3a79c7a2ee3f00f979908" exitCode=0 Dec 07 16:16:24 crc kubenswrapper[4716]: I1207 16:16:24.337686 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" event={"ID":"c15b59eb-565d-4556-a4ce-75afdf159dc8","Type":"ContainerDied","Data":"8ec0c9fb49e1eaea181d9e297f07686725c05f58dac3a79c7a2ee3f00f979908"} Dec 07 16:16:24 crc kubenswrapper[4716]: I1207 16:16:24.337725 4716 scope.go:117] "RemoveContainer" containerID="def7373f410cd12d2a355b52c07eee66525502687b1341f3cbd5919e77d0a06e" Dec 07 16:16:24 crc kubenswrapper[4716]: I1207 16:16:24.352593 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5csl5\" (UniqueName: \"kubernetes.io/projected/9d840ba3-d5ef-48fc-87b2-6af4d164c9ba-kube-api-access-5csl5\") pod \"openstack-operator-controller-manager-7db7d7886d-z6n25\" (UID: \"9d840ba3-d5ef-48fc-87b2-6af4d164c9ba\") " pod="openstack-operators/openstack-operator-controller-manager-7db7d7886d-z6n25" Dec 07 16:16:24 crc kubenswrapper[4716]: I1207 16:16:24.353108 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hnrd\" (UniqueName: \"kubernetes.io/projected/79aa5696-34a1-45cf-b965-56f64fe63f4e-kube-api-access-4hnrd\") pod \"rabbitmq-cluster-operator-manager-668c99d594-7mjmm\" (UID: \"79aa5696-34a1-45cf-b965-56f64fe63f4e\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-7mjmm" Dec 07 16:16:24 crc kubenswrapper[4716]: I1207 16:16:24.394176 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-8xrch" Dec 07 16:16:24 crc kubenswrapper[4716]: I1207 16:16:24.405466 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-z7lcv" Dec 07 16:16:24 crc kubenswrapper[4716]: I1207 16:16:24.529423 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/34e0d69c-2f36-4ba6-a4e3-114cabe016c2-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879fdkpkx\" (UID: \"34e0d69c-2f36-4ba6-a4e3-114cabe016c2\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fdkpkx" Dec 07 16:16:24 crc kubenswrapper[4716]: E1207 16:16:24.529590 4716 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 07 16:16:24 crc kubenswrapper[4716]: E1207 16:16:24.529644 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/34e0d69c-2f36-4ba6-a4e3-114cabe016c2-cert podName:34e0d69c-2f36-4ba6-a4e3-114cabe016c2 nodeName:}" failed. No retries permitted until 2025-12-07 16:16:25.529625805 +0000 UTC m=+848.219910707 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/34e0d69c-2f36-4ba6-a4e3-114cabe016c2-cert") pod "openstack-baremetal-operator-controller-manager-84b575879fdkpkx" (UID: "34e0d69c-2f36-4ba6-a4e3-114cabe016c2") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 07 16:16:24 crc kubenswrapper[4716]: I1207 16:16:24.538770 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-ggcr7" Dec 07 16:16:24 crc kubenswrapper[4716]: I1207 16:16:24.559211 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-697fb699cf-zts6p"] Dec 07 16:16:24 crc kubenswrapper[4716]: I1207 16:16:24.624568 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-7mjmm" Dec 07 16:16:24 crc kubenswrapper[4716]: I1207 16:16:24.667132 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-xqnzw"] Dec 07 16:16:24 crc kubenswrapper[4716]: I1207 16:16:24.684654 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-m957c"] Dec 07 16:16:24 crc kubenswrapper[4716]: I1207 16:16:24.834950 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/9d840ba3-d5ef-48fc-87b2-6af4d164c9ba-webhook-certs\") pod \"openstack-operator-controller-manager-7db7d7886d-z6n25\" (UID: \"9d840ba3-d5ef-48fc-87b2-6af4d164c9ba\") " pod="openstack-operators/openstack-operator-controller-manager-7db7d7886d-z6n25" Dec 07 16:16:24 crc kubenswrapper[4716]: I1207 16:16:24.835053 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9d840ba3-d5ef-48fc-87b2-6af4d164c9ba-metrics-certs\") pod \"openstack-operator-controller-manager-7db7d7886d-z6n25\" (UID: \"9d840ba3-d5ef-48fc-87b2-6af4d164c9ba\") " pod="openstack-operators/openstack-operator-controller-manager-7db7d7886d-z6n25" Dec 07 16:16:24 crc kubenswrapper[4716]: E1207 16:16:24.835232 4716 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 07 16:16:24 crc kubenswrapper[4716]: E1207 16:16:24.835280 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9d840ba3-d5ef-48fc-87b2-6af4d164c9ba-metrics-certs podName:9d840ba3-d5ef-48fc-87b2-6af4d164c9ba nodeName:}" failed. No retries permitted until 2025-12-07 16:16:25.83526523 +0000 UTC m=+848.525550142 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9d840ba3-d5ef-48fc-87b2-6af4d164c9ba-metrics-certs") pod "openstack-operator-controller-manager-7db7d7886d-z6n25" (UID: "9d840ba3-d5ef-48fc-87b2-6af4d164c9ba") : secret "metrics-server-cert" not found Dec 07 16:16:24 crc kubenswrapper[4716]: E1207 16:16:24.835324 4716 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 07 16:16:24 crc kubenswrapper[4716]: E1207 16:16:24.835344 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9d840ba3-d5ef-48fc-87b2-6af4d164c9ba-webhook-certs podName:9d840ba3-d5ef-48fc-87b2-6af4d164c9ba nodeName:}" failed. No retries permitted until 2025-12-07 16:16:25.835337382 +0000 UTC m=+848.525622294 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/9d840ba3-d5ef-48fc-87b2-6af4d164c9ba-webhook-certs") pod "openstack-operator-controller-manager-7db7d7886d-z6n25" (UID: "9d840ba3-d5ef-48fc-87b2-6af4d164c9ba") : secret "webhook-server-cert" not found Dec 07 16:16:24 crc kubenswrapper[4716]: I1207 16:16:24.839566 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-wctdp"] Dec 07 16:16:24 crc kubenswrapper[4716]: I1207 16:16:24.852465 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-5697bb5779-rcmz5"] Dec 07 16:16:24 crc kubenswrapper[4716]: I1207 16:16:24.924232 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6c677c69b-4nrlz"] Dec 07 16:16:24 crc kubenswrapper[4716]: I1207 16:16:24.927744 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-nx5mv"] Dec 07 16:16:24 crc kubenswrapper[4716]: W1207 16:16:24.928305 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod62368956_27d2_41fa_a2b2_0fb49d869f11.slice/crio-7a44141ca5b9c4d8276a672020bd2799513a94f4a92b3b2694e2be1e7b176d11 WatchSource:0}: Error finding container 7a44141ca5b9c4d8276a672020bd2799513a94f4a92b3b2694e2be1e7b176d11: Status 404 returned error can't find the container with id 7a44141ca5b9c4d8276a672020bd2799513a94f4a92b3b2694e2be1e7b176d11 Dec 07 16:16:24 crc kubenswrapper[4716]: W1207 16:16:24.929353 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod20067e63_712d_4f94_9019_627219d4299f.slice/crio-8f5090e59c0ca4bb8f239daa7c346a202441292de7113070213a3b095bdb283b WatchSource:0}: Error finding container 8f5090e59c0ca4bb8f239daa7c346a202441292de7113070213a3b095bdb283b: Status 404 returned error can't find the container with id 8f5090e59c0ca4bb8f239daa7c346a202441292de7113070213a3b095bdb283b Dec 07 16:16:25 crc kubenswrapper[4716]: I1207 16:16:25.099802 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-kkdgl"] Dec 07 16:16:25 crc kubenswrapper[4716]: W1207 16:16:25.106659 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda929d511_1d07_429d_b302_8843a6834f52.slice/crio-b0197ff74d00b67875c83bc613133b5817c6df30f12dbb042cb890535e7e0529 WatchSource:0}: Error finding container b0197ff74d00b67875c83bc613133b5817c6df30f12dbb042cb890535e7e0529: Status 404 returned error can't find the container with id b0197ff74d00b67875c83bc613133b5817c6df30f12dbb042cb890535e7e0529 Dec 07 16:16:25 crc kubenswrapper[4716]: I1207 16:16:25.111615 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-h95ls"] Dec 07 16:16:25 crc kubenswrapper[4716]: I1207 16:16:25.119268 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-967d97867-hx48c"] Dec 07 16:16:25 crc kubenswrapper[4716]: I1207 16:16:25.125921 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-9d58d64bc-p2hbf"] Dec 07 16:16:25 crc kubenswrapper[4716]: I1207 16:16:25.134489 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-79c8c4686c-9dsmr"] Dec 07 16:16:25 crc kubenswrapper[4716]: I1207 16:16:25.142022 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5b5fd79c9c-kbzpp"] Dec 07 16:16:25 crc kubenswrapper[4716]: E1207 16:16:25.148306 4716 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/manila-operator@sha256:44126f9c6b1d2bf752ddf989e20a4fc4cc1c07723d4fcb78465ccb2f55da6b3a,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-q6ldw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-5b5fd79c9c-kbzpp_openstack-operators(d3ef7831-ac99-410d-90bf-42d87042964b): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 07 16:16:25 crc kubenswrapper[4716]: E1207 16:16:25.151382 4716 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-q6ldw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-5b5fd79c9c-kbzpp_openstack-operators(d3ef7831-ac99-410d-90bf-42d87042964b): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 07 16:16:25 crc kubenswrapper[4716]: E1207 16:16:25.152653 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-kbzpp" podUID="d3ef7831-ac99-410d-90bf-42d87042964b" Dec 07 16:16:25 crc kubenswrapper[4716]: I1207 16:16:25.246611 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-srq27"] Dec 07 16:16:25 crc kubenswrapper[4716]: W1207 16:16:25.250340 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod463e6bed_ce60_4e35_98a3_cd837b9066f2.slice/crio-b9e0039a1ecd085c88c65e911aacccde7f4c0104ba0160c34a5476a6b1f5d86a WatchSource:0}: Error finding container b9e0039a1ecd085c88c65e911aacccde7f4c0104ba0160c34a5476a6b1f5d86a: Status 404 returned error can't find the container with id b9e0039a1ecd085c88c65e911aacccde7f4c0104ba0160c34a5476a6b1f5d86a Dec 07 16:16:25 crc kubenswrapper[4716]: I1207 16:16:25.255445 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-667bd8d554-8zwtk"] Dec 07 16:16:25 crc kubenswrapper[4716]: E1207 16:16:25.261575 4716 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-pm9xg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-697bc559fc-z7lcv_openstack-operators(69c2e54f-9af6-44e8-abdd-226a29b64da6): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 07 16:16:25 crc kubenswrapper[4716]: E1207 16:16:25.264219 4716 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:6b3e0302608a2e70f9b5ae9167f6fbf59264f226d9db99d48f70466ab2f216b8,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8gskr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-667bd8d554-8zwtk_openstack-operators(2ef47713-a73c-4c63-8fa0-ffc530832285): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 07 16:16:25 crc kubenswrapper[4716]: E1207 16:16:25.264550 4716 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-pm9xg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-697bc559fc-z7lcv_openstack-operators(69c2e54f-9af6-44e8-abdd-226a29b64da6): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 07 16:16:25 crc kubenswrapper[4716]: E1207 16:16:25.265996 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-z7lcv" podUID="69c2e54f-9af6-44e8-abdd-226a29b64da6" Dec 07 16:16:25 crc kubenswrapper[4716]: E1207 16:16:25.266144 4716 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8gskr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-667bd8d554-8zwtk_openstack-operators(2ef47713-a73c-4c63-8fa0-ffc530832285): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 07 16:16:25 crc kubenswrapper[4716]: E1207 16:16:25.268650 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-8zwtk" podUID="2ef47713-a73c-4c63-8fa0-ffc530832285" Dec 07 16:16:25 crc kubenswrapper[4716]: E1207 16:16:25.272242 4716 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:f27e732ec1faee765461bf137d9be81278b2fa39675019a73622755e1e610b6f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zsmst,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-58d5ff84df-ggcr7_openstack-operators(e4c2f2f6-3285-4f48-a428-b2065f33c046): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 07 16:16:25 crc kubenswrapper[4716]: I1207 16:16:25.280945 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-z7lcv"] Dec 07 16:16:25 crc kubenswrapper[4716]: E1207 16:16:25.283996 4716 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zsmst,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-58d5ff84df-ggcr7_openstack-operators(e4c2f2f6-3285-4f48-a428-b2065f33c046): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 07 16:16:25 crc kubenswrapper[4716]: E1207 16:16:25.287394 4716 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-k28gb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-8xrch_openstack-operators(0f05e35d-f818-4f65-bac6-f6a22006627e): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 07 16:16:25 crc kubenswrapper[4716]: E1207 16:16:25.288372 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-ggcr7" podUID="e4c2f2f6-3285-4f48-a428-b2065f33c046" Dec 07 16:16:25 crc kubenswrapper[4716]: E1207 16:16:25.289267 4716 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-k28gb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-8xrch_openstack-operators(0f05e35d-f818-4f65-bac6-f6a22006627e): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 07 16:16:25 crc kubenswrapper[4716]: I1207 16:16:25.292861 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-58d5ff84df-ggcr7"] Dec 07 16:16:25 crc kubenswrapper[4716]: E1207 16:16:25.296223 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-8xrch" podUID="0f05e35d-f818-4f65-bac6-f6a22006627e" Dec 07 16:16:25 crc kubenswrapper[4716]: E1207 16:16:25.301098 4716 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5q6j7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-4dnl8_openstack-operators(0c6a49cd-1cfe-49c8-a068-f85c735b46b1): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 07 16:16:25 crc kubenswrapper[4716]: I1207 16:16:25.301885 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-8xrch"] Dec 07 16:16:25 crc kubenswrapper[4716]: E1207 16:16:25.303595 4716 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5q6j7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-4dnl8_openstack-operators(0c6a49cd-1cfe-49c8-a068-f85c735b46b1): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 07 16:16:25 crc kubenswrapper[4716]: E1207 16:16:25.304797 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-4dnl8" podUID="0c6a49cd-1cfe-49c8-a068-f85c735b46b1" Dec 07 16:16:25 crc kubenswrapper[4716]: I1207 16:16:25.310492 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-4dnl8"] Dec 07 16:16:25 crc kubenswrapper[4716]: I1207 16:16:25.340979 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/01686458-debd-419a-90ec-cd27cc6953ec-cert\") pod \"infra-operator-controller-manager-78d48bff9d-pxs5r\" (UID: \"01686458-debd-419a-90ec-cd27cc6953ec\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-pxs5r" Dec 07 16:16:25 crc kubenswrapper[4716]: E1207 16:16:25.341155 4716 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 07 16:16:25 crc kubenswrapper[4716]: E1207 16:16:25.341212 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/01686458-debd-419a-90ec-cd27cc6953ec-cert podName:01686458-debd-419a-90ec-cd27cc6953ec nodeName:}" failed. No retries permitted until 2025-12-07 16:16:27.341193116 +0000 UTC m=+850.031478028 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/01686458-debd-419a-90ec-cd27cc6953ec-cert") pod "infra-operator-controller-manager-78d48bff9d-pxs5r" (UID: "01686458-debd-419a-90ec-cd27cc6953ec") : secret "infra-operator-webhook-server-cert" not found Dec 07 16:16:25 crc kubenswrapper[4716]: I1207 16:16:25.347940 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-967d97867-hx48c" event={"ID":"a929d511-1d07-429d-b302-8843a6834f52","Type":"ContainerStarted","Data":"b0197ff74d00b67875c83bc613133b5817c6df30f12dbb042cb890535e7e0529"} Dec 07 16:16:25 crc kubenswrapper[4716]: I1207 16:16:25.349541 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-4nrlz" event={"ID":"62368956-27d2-41fa-a2b2-0fb49d869f11","Type":"ContainerStarted","Data":"7a44141ca5b9c4d8276a672020bd2799513a94f4a92b3b2694e2be1e7b176d11"} Dec 07 16:16:25 crc kubenswrapper[4716]: I1207 16:16:25.351581 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-z7lcv" event={"ID":"69c2e54f-9af6-44e8-abdd-226a29b64da6","Type":"ContainerStarted","Data":"0128c4a82d37ed166e94964bc2fa690b3139a2a5ab0bb3d5cb7f6ef1f5aec935"} Dec 07 16:16:25 crc kubenswrapper[4716]: I1207 16:16:25.352736 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-srq27" event={"ID":"463e6bed-ce60-4e35-98a3-cd837b9066f2","Type":"ContainerStarted","Data":"b9e0039a1ecd085c88c65e911aacccde7f4c0104ba0160c34a5476a6b1f5d86a"} Dec 07 16:16:25 crc kubenswrapper[4716]: I1207 16:16:25.354227 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-m957c" event={"ID":"8765c8b2-a50b-4417-adbe-1174dcdfe172","Type":"ContainerStarted","Data":"b5353d7e2569f43224f0039b7e4a7cd13bc6052147a014e8b74820ce45fdb3a3"} Dec 07 16:16:25 crc kubenswrapper[4716]: E1207 16:16:25.354830 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-z7lcv" podUID="69c2e54f-9af6-44e8-abdd-226a29b64da6" Dec 07 16:16:25 crc kubenswrapper[4716]: I1207 16:16:25.356581 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-9dsmr" event={"ID":"5be39543-890e-479c-a041-d864922e038f","Type":"ContainerStarted","Data":"75f7d5c53ff118efda16e66da84d8b88dac1bd042469a3b26a9901af9d1d51cb"} Dec 07 16:16:25 crc kubenswrapper[4716]: I1207 16:16:25.357480 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-rcmz5" event={"ID":"6f24594b-ebe7-4518-a067-45891924abe5","Type":"ContainerStarted","Data":"ac27b67b8f410071cfffa1fdfa037d47e52139bc5e00edb3eb07e89cafd1ca6c"} Dec 07 16:16:25 crc kubenswrapper[4716]: I1207 16:16:25.358330 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-wctdp" event={"ID":"2babb856-a5c2-4b60-a6bc-00d3f510f014","Type":"ContainerStarted","Data":"971f79f80f0c8f4e1acfb04fa8ce4570a251ea7990c10ad99d77058d901456a5"} Dec 07 16:16:25 crc kubenswrapper[4716]: I1207 16:16:25.359611 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-4dnl8" event={"ID":"0c6a49cd-1cfe-49c8-a068-f85c735b46b1","Type":"ContainerStarted","Data":"17cd685ff8d4b57fca8d9dce6ba4237fce2c978181e7d0dcb4f38d0b6cdb7d36"} Dec 07 16:16:25 crc kubenswrapper[4716]: I1207 16:16:25.360728 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-kbzpp" event={"ID":"d3ef7831-ac99-410d-90bf-42d87042964b","Type":"ContainerStarted","Data":"c5f9b87c915ece9b87d1e5fa50014e7b33a65d8d1e14971e8e8c9088e2a971c6"} Dec 07 16:16:25 crc kubenswrapper[4716]: E1207 16:16:25.361453 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-4dnl8" podUID="0c6a49cd-1cfe-49c8-a068-f85c735b46b1" Dec 07 16:16:25 crc kubenswrapper[4716]: E1207 16:16:25.363269 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:44126f9c6b1d2bf752ddf989e20a4fc4cc1c07723d4fcb78465ccb2f55da6b3a\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-kbzpp" podUID="d3ef7831-ac99-410d-90bf-42d87042964b" Dec 07 16:16:25 crc kubenswrapper[4716]: I1207 16:16:25.417694 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-nx5mv" event={"ID":"20067e63-712d-4f94-9019-627219d4299f","Type":"ContainerStarted","Data":"8f5090e59c0ca4bb8f239daa7c346a202441292de7113070213a3b095bdb283b"} Dec 07 16:16:25 crc kubenswrapper[4716]: I1207 16:16:25.426345 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" event={"ID":"c15b59eb-565d-4556-a4ce-75afdf159dc8","Type":"ContainerStarted","Data":"73e652154907931afdb244fb07e6dd85483e6b06afe1a76ebcc2b8bf8b2c7310"} Dec 07 16:16:25 crc kubenswrapper[4716]: I1207 16:16:25.427542 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-zts6p" event={"ID":"79562d20-950f-428e-ac3a-f78979053266","Type":"ContainerStarted","Data":"35c38f76cf348325cadd6100f4e1c73b0bebdea6ddeda2d39c292705ebfa011c"} Dec 07 16:16:25 crc kubenswrapper[4716]: I1207 16:16:25.428969 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-8zwtk" event={"ID":"2ef47713-a73c-4c63-8fa0-ffc530832285","Type":"ContainerStarted","Data":"118db2ef3eb06fe601211f59cbf66de6b54572a6429f8e5502321e1fd9b2e404"} Dec 07 16:16:25 crc kubenswrapper[4716]: I1207 16:16:25.432438 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-8xrch" event={"ID":"0f05e35d-f818-4f65-bac6-f6a22006627e","Type":"ContainerStarted","Data":"fba107916662d8e3b59de3d6aac1889546e28dd94932bc68e339620fced1ad4b"} Dec 07 16:16:25 crc kubenswrapper[4716]: E1207 16:16:25.433025 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:6b3e0302608a2e70f9b5ae9167f6fbf59264f226d9db99d48f70466ab2f216b8\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-8zwtk" podUID="2ef47713-a73c-4c63-8fa0-ffc530832285" Dec 07 16:16:25 crc kubenswrapper[4716]: E1207 16:16:25.437015 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-8xrch" podUID="0f05e35d-f818-4f65-bac6-f6a22006627e" Dec 07 16:16:25 crc kubenswrapper[4716]: I1207 16:16:25.438235 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-h95ls" event={"ID":"1bf3b0b2-ea38-4fa3-b07c-4467dc111476","Type":"ContainerStarted","Data":"006f689d4ae94a0446c5a48c22f514be502484862235c6caf0ac9efbcac690c8"} Dec 07 16:16:25 crc kubenswrapper[4716]: I1207 16:16:25.439219 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-p2hbf" event={"ID":"d9d5a1f4-edb7-4630-8907-5d29c4678f24","Type":"ContainerStarted","Data":"b4df742367d05322f225a4b82f80bd0545cb756b8564dd9949cebf6a11f1612b"} Dec 07 16:16:25 crc kubenswrapper[4716]: I1207 16:16:25.451476 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-7mjmm"] Dec 07 16:16:25 crc kubenswrapper[4716]: I1207 16:16:25.454573 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-kkdgl" event={"ID":"57c9705a-9bf1-45b4-aefb-d224a9d72a4a","Type":"ContainerStarted","Data":"d74a9a0ebe17d15a842224feae92a4d17ef5fb54cf27cab3cd50d518db65e642"} Dec 07 16:16:25 crc kubenswrapper[4716]: I1207 16:16:25.465386 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-ggcr7" event={"ID":"e4c2f2f6-3285-4f48-a428-b2065f33c046","Type":"ContainerStarted","Data":"2bdaa00fcedc0519d60c345e57f6ee0f3ce48ac551ec6619ad73a22d3e839d65"} Dec 07 16:16:25 crc kubenswrapper[4716]: I1207 16:16:25.466660 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-xqnzw" event={"ID":"d9d5c8b8-240b-4c88-ad6e-b04881842f0b","Type":"ContainerStarted","Data":"43f3c193e04651bdab171fc2b2fa7d3cb2c111a260dbdf30d43077ae859adbd1"} Dec 07 16:16:25 crc kubenswrapper[4716]: E1207 16:16:25.468067 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:f27e732ec1faee765461bf137d9be81278b2fa39675019a73622755e1e610b6f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-ggcr7" podUID="e4c2f2f6-3285-4f48-a428-b2065f33c046" Dec 07 16:16:25 crc kubenswrapper[4716]: I1207 16:16:25.545495 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/34e0d69c-2f36-4ba6-a4e3-114cabe016c2-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879fdkpkx\" (UID: \"34e0d69c-2f36-4ba6-a4e3-114cabe016c2\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fdkpkx" Dec 07 16:16:25 crc kubenswrapper[4716]: E1207 16:16:25.545756 4716 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 07 16:16:25 crc kubenswrapper[4716]: E1207 16:16:25.545824 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/34e0d69c-2f36-4ba6-a4e3-114cabe016c2-cert podName:34e0d69c-2f36-4ba6-a4e3-114cabe016c2 nodeName:}" failed. No retries permitted until 2025-12-07 16:16:27.545805716 +0000 UTC m=+850.236090628 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/34e0d69c-2f36-4ba6-a4e3-114cabe016c2-cert") pod "openstack-baremetal-operator-controller-manager-84b575879fdkpkx" (UID: "34e0d69c-2f36-4ba6-a4e3-114cabe016c2") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 07 16:16:25 crc kubenswrapper[4716]: I1207 16:16:25.850360 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9d840ba3-d5ef-48fc-87b2-6af4d164c9ba-metrics-certs\") pod \"openstack-operator-controller-manager-7db7d7886d-z6n25\" (UID: \"9d840ba3-d5ef-48fc-87b2-6af4d164c9ba\") " pod="openstack-operators/openstack-operator-controller-manager-7db7d7886d-z6n25" Dec 07 16:16:25 crc kubenswrapper[4716]: I1207 16:16:25.850795 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/9d840ba3-d5ef-48fc-87b2-6af4d164c9ba-webhook-certs\") pod \"openstack-operator-controller-manager-7db7d7886d-z6n25\" (UID: \"9d840ba3-d5ef-48fc-87b2-6af4d164c9ba\") " pod="openstack-operators/openstack-operator-controller-manager-7db7d7886d-z6n25" Dec 07 16:16:25 crc kubenswrapper[4716]: E1207 16:16:25.850529 4716 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 07 16:16:25 crc kubenswrapper[4716]: E1207 16:16:25.850864 4716 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 07 16:16:25 crc kubenswrapper[4716]: E1207 16:16:25.850928 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9d840ba3-d5ef-48fc-87b2-6af4d164c9ba-metrics-certs podName:9d840ba3-d5ef-48fc-87b2-6af4d164c9ba nodeName:}" failed. No retries permitted until 2025-12-07 16:16:27.850909846 +0000 UTC m=+850.541194758 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9d840ba3-d5ef-48fc-87b2-6af4d164c9ba-metrics-certs") pod "openstack-operator-controller-manager-7db7d7886d-z6n25" (UID: "9d840ba3-d5ef-48fc-87b2-6af4d164c9ba") : secret "metrics-server-cert" not found Dec 07 16:16:25 crc kubenswrapper[4716]: E1207 16:16:25.850946 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9d840ba3-d5ef-48fc-87b2-6af4d164c9ba-webhook-certs podName:9d840ba3-d5ef-48fc-87b2-6af4d164c9ba nodeName:}" failed. No retries permitted until 2025-12-07 16:16:27.850937957 +0000 UTC m=+850.541222869 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/9d840ba3-d5ef-48fc-87b2-6af4d164c9ba-webhook-certs") pod "openstack-operator-controller-manager-7db7d7886d-z6n25" (UID: "9d840ba3-d5ef-48fc-87b2-6af4d164c9ba") : secret "webhook-server-cert" not found Dec 07 16:16:26 crc kubenswrapper[4716]: I1207 16:16:26.484273 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-7mjmm" event={"ID":"79aa5696-34a1-45cf-b965-56f64fe63f4e","Type":"ContainerStarted","Data":"389482b68c90233653ad53234b774891675b050119bebefdf9cbbe3276f9d5ae"} Dec 07 16:16:26 crc kubenswrapper[4716]: E1207 16:16:26.487847 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-z7lcv" podUID="69c2e54f-9af6-44e8-abdd-226a29b64da6" Dec 07 16:16:26 crc kubenswrapper[4716]: E1207 16:16:26.488976 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:44126f9c6b1d2bf752ddf989e20a4fc4cc1c07723d4fcb78465ccb2f55da6b3a\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-kbzpp" podUID="d3ef7831-ac99-410d-90bf-42d87042964b" Dec 07 16:16:26 crc kubenswrapper[4716]: E1207 16:16:26.489110 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-8xrch" podUID="0f05e35d-f818-4f65-bac6-f6a22006627e" Dec 07 16:16:26 crc kubenswrapper[4716]: E1207 16:16:26.489277 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-4dnl8" podUID="0c6a49cd-1cfe-49c8-a068-f85c735b46b1" Dec 07 16:16:26 crc kubenswrapper[4716]: E1207 16:16:26.489337 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:6b3e0302608a2e70f9b5ae9167f6fbf59264f226d9db99d48f70466ab2f216b8\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-8zwtk" podUID="2ef47713-a73c-4c63-8fa0-ffc530832285" Dec 07 16:16:26 crc kubenswrapper[4716]: E1207 16:16:26.489778 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:f27e732ec1faee765461bf137d9be81278b2fa39675019a73622755e1e610b6f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-ggcr7" podUID="e4c2f2f6-3285-4f48-a428-b2065f33c046" Dec 07 16:16:27 crc kubenswrapper[4716]: I1207 16:16:27.373769 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/01686458-debd-419a-90ec-cd27cc6953ec-cert\") pod \"infra-operator-controller-manager-78d48bff9d-pxs5r\" (UID: \"01686458-debd-419a-90ec-cd27cc6953ec\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-pxs5r" Dec 07 16:16:27 crc kubenswrapper[4716]: E1207 16:16:27.373968 4716 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 07 16:16:27 crc kubenswrapper[4716]: E1207 16:16:27.374016 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/01686458-debd-419a-90ec-cd27cc6953ec-cert podName:01686458-debd-419a-90ec-cd27cc6953ec nodeName:}" failed. No retries permitted until 2025-12-07 16:16:31.374001161 +0000 UTC m=+854.064286073 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/01686458-debd-419a-90ec-cd27cc6953ec-cert") pod "infra-operator-controller-manager-78d48bff9d-pxs5r" (UID: "01686458-debd-419a-90ec-cd27cc6953ec") : secret "infra-operator-webhook-server-cert" not found Dec 07 16:16:27 crc kubenswrapper[4716]: I1207 16:16:27.576669 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/34e0d69c-2f36-4ba6-a4e3-114cabe016c2-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879fdkpkx\" (UID: \"34e0d69c-2f36-4ba6-a4e3-114cabe016c2\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fdkpkx" Dec 07 16:16:27 crc kubenswrapper[4716]: E1207 16:16:27.576908 4716 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 07 16:16:27 crc kubenswrapper[4716]: E1207 16:16:27.577001 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/34e0d69c-2f36-4ba6-a4e3-114cabe016c2-cert podName:34e0d69c-2f36-4ba6-a4e3-114cabe016c2 nodeName:}" failed. No retries permitted until 2025-12-07 16:16:31.576982686 +0000 UTC m=+854.267267598 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/34e0d69c-2f36-4ba6-a4e3-114cabe016c2-cert") pod "openstack-baremetal-operator-controller-manager-84b575879fdkpkx" (UID: "34e0d69c-2f36-4ba6-a4e3-114cabe016c2") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 07 16:16:27 crc kubenswrapper[4716]: I1207 16:16:27.881323 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/9d840ba3-d5ef-48fc-87b2-6af4d164c9ba-webhook-certs\") pod \"openstack-operator-controller-manager-7db7d7886d-z6n25\" (UID: \"9d840ba3-d5ef-48fc-87b2-6af4d164c9ba\") " pod="openstack-operators/openstack-operator-controller-manager-7db7d7886d-z6n25" Dec 07 16:16:27 crc kubenswrapper[4716]: I1207 16:16:27.881536 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9d840ba3-d5ef-48fc-87b2-6af4d164c9ba-metrics-certs\") pod \"openstack-operator-controller-manager-7db7d7886d-z6n25\" (UID: \"9d840ba3-d5ef-48fc-87b2-6af4d164c9ba\") " pod="openstack-operators/openstack-operator-controller-manager-7db7d7886d-z6n25" Dec 07 16:16:27 crc kubenswrapper[4716]: E1207 16:16:27.881546 4716 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 07 16:16:27 crc kubenswrapper[4716]: E1207 16:16:27.881605 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9d840ba3-d5ef-48fc-87b2-6af4d164c9ba-webhook-certs podName:9d840ba3-d5ef-48fc-87b2-6af4d164c9ba nodeName:}" failed. No retries permitted until 2025-12-07 16:16:31.881588312 +0000 UTC m=+854.571873234 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/9d840ba3-d5ef-48fc-87b2-6af4d164c9ba-webhook-certs") pod "openstack-operator-controller-manager-7db7d7886d-z6n25" (UID: "9d840ba3-d5ef-48fc-87b2-6af4d164c9ba") : secret "webhook-server-cert" not found Dec 07 16:16:27 crc kubenswrapper[4716]: E1207 16:16:27.881664 4716 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 07 16:16:27 crc kubenswrapper[4716]: E1207 16:16:27.881702 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9d840ba3-d5ef-48fc-87b2-6af4d164c9ba-metrics-certs podName:9d840ba3-d5ef-48fc-87b2-6af4d164c9ba nodeName:}" failed. No retries permitted until 2025-12-07 16:16:31.881690085 +0000 UTC m=+854.571974997 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9d840ba3-d5ef-48fc-87b2-6af4d164c9ba-metrics-certs") pod "openstack-operator-controller-manager-7db7d7886d-z6n25" (UID: "9d840ba3-d5ef-48fc-87b2-6af4d164c9ba") : secret "metrics-server-cert" not found Dec 07 16:16:31 crc kubenswrapper[4716]: I1207 16:16:31.426462 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/01686458-debd-419a-90ec-cd27cc6953ec-cert\") pod \"infra-operator-controller-manager-78d48bff9d-pxs5r\" (UID: \"01686458-debd-419a-90ec-cd27cc6953ec\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-pxs5r" Dec 07 16:16:31 crc kubenswrapper[4716]: E1207 16:16:31.426629 4716 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 07 16:16:31 crc kubenswrapper[4716]: E1207 16:16:31.427586 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/01686458-debd-419a-90ec-cd27cc6953ec-cert podName:01686458-debd-419a-90ec-cd27cc6953ec nodeName:}" failed. No retries permitted until 2025-12-07 16:16:39.427554535 +0000 UTC m=+862.117839447 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/01686458-debd-419a-90ec-cd27cc6953ec-cert") pod "infra-operator-controller-manager-78d48bff9d-pxs5r" (UID: "01686458-debd-419a-90ec-cd27cc6953ec") : secret "infra-operator-webhook-server-cert" not found Dec 07 16:16:31 crc kubenswrapper[4716]: I1207 16:16:31.631373 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/34e0d69c-2f36-4ba6-a4e3-114cabe016c2-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879fdkpkx\" (UID: \"34e0d69c-2f36-4ba6-a4e3-114cabe016c2\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fdkpkx" Dec 07 16:16:31 crc kubenswrapper[4716]: E1207 16:16:31.631533 4716 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 07 16:16:31 crc kubenswrapper[4716]: E1207 16:16:31.631595 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/34e0d69c-2f36-4ba6-a4e3-114cabe016c2-cert podName:34e0d69c-2f36-4ba6-a4e3-114cabe016c2 nodeName:}" failed. No retries permitted until 2025-12-07 16:16:39.631576399 +0000 UTC m=+862.321861311 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/34e0d69c-2f36-4ba6-a4e3-114cabe016c2-cert") pod "openstack-baremetal-operator-controller-manager-84b575879fdkpkx" (UID: "34e0d69c-2f36-4ba6-a4e3-114cabe016c2") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 07 16:16:31 crc kubenswrapper[4716]: I1207 16:16:31.935740 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9d840ba3-d5ef-48fc-87b2-6af4d164c9ba-metrics-certs\") pod \"openstack-operator-controller-manager-7db7d7886d-z6n25\" (UID: \"9d840ba3-d5ef-48fc-87b2-6af4d164c9ba\") " pod="openstack-operators/openstack-operator-controller-manager-7db7d7886d-z6n25" Dec 07 16:16:31 crc kubenswrapper[4716]: I1207 16:16:31.935864 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/9d840ba3-d5ef-48fc-87b2-6af4d164c9ba-webhook-certs\") pod \"openstack-operator-controller-manager-7db7d7886d-z6n25\" (UID: \"9d840ba3-d5ef-48fc-87b2-6af4d164c9ba\") " pod="openstack-operators/openstack-operator-controller-manager-7db7d7886d-z6n25" Dec 07 16:16:31 crc kubenswrapper[4716]: E1207 16:16:31.935971 4716 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 07 16:16:31 crc kubenswrapper[4716]: E1207 16:16:31.936062 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9d840ba3-d5ef-48fc-87b2-6af4d164c9ba-metrics-certs podName:9d840ba3-d5ef-48fc-87b2-6af4d164c9ba nodeName:}" failed. No retries permitted until 2025-12-07 16:16:39.936043202 +0000 UTC m=+862.626328114 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9d840ba3-d5ef-48fc-87b2-6af4d164c9ba-metrics-certs") pod "openstack-operator-controller-manager-7db7d7886d-z6n25" (UID: "9d840ba3-d5ef-48fc-87b2-6af4d164c9ba") : secret "metrics-server-cert" not found Dec 07 16:16:31 crc kubenswrapper[4716]: E1207 16:16:31.936128 4716 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 07 16:16:31 crc kubenswrapper[4716]: E1207 16:16:31.936261 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9d840ba3-d5ef-48fc-87b2-6af4d164c9ba-webhook-certs podName:9d840ba3-d5ef-48fc-87b2-6af4d164c9ba nodeName:}" failed. No retries permitted until 2025-12-07 16:16:39.936233597 +0000 UTC m=+862.626518509 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/9d840ba3-d5ef-48fc-87b2-6af4d164c9ba-webhook-certs") pod "openstack-operator-controller-manager-7db7d7886d-z6n25" (UID: "9d840ba3-d5ef-48fc-87b2-6af4d164c9ba") : secret "webhook-server-cert" not found Dec 07 16:16:37 crc kubenswrapper[4716]: E1207 16:16:37.063323 4716 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/designate-operator@sha256:900050d3501c0785b227db34b89883efe68247816e5c7427cacb74f8aa10605a" Dec 07 16:16:37 crc kubenswrapper[4716]: E1207 16:16:37.064480 4716 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/designate-operator@sha256:900050d3501c0785b227db34b89883efe68247816e5c7427cacb74f8aa10605a,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-cg6g9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod designate-operator-controller-manager-697fb699cf-zts6p_openstack-operators(79562d20-950f-428e-ac3a-f78979053266): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 07 16:16:37 crc kubenswrapper[4716]: E1207 16:16:37.699560 4716 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/swift-operator@sha256:3aa109bb973253ae9dcf339b9b65abbd1176cdb4be672c93e538a5f113816991" Dec 07 16:16:37 crc kubenswrapper[4716]: E1207 16:16:37.699857 4716 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:3aa109bb973253ae9dcf339b9b65abbd1176cdb4be672c93e538a5f113816991,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-kj2xt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-9d58d64bc-p2hbf_openstack-operators(d9d5a1f4-edb7-4630-8907-5d29c4678f24): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 07 16:16:38 crc kubenswrapper[4716]: E1207 16:16:38.127270 4716 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2" Dec 07 16:16:38 crc kubenswrapper[4716]: E1207 16:16:38.127533 4716 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4hnrd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-7mjmm_openstack-operators(79aa5696-34a1-45cf-b965-56f64fe63f4e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 07 16:16:38 crc kubenswrapper[4716]: E1207 16:16:38.130923 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-7mjmm" podUID="79aa5696-34a1-45cf-b965-56f64fe63f4e" Dec 07 16:16:38 crc kubenswrapper[4716]: E1207 16:16:38.591255 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-7mjmm" podUID="79aa5696-34a1-45cf-b965-56f64fe63f4e" Dec 07 16:16:38 crc kubenswrapper[4716]: E1207 16:16:38.667682 4716 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7" Dec 07 16:16:38 crc kubenswrapper[4716]: E1207 16:16:38.667970 4716 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-tghx4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7765d96ddf-wctdp_openstack-operators(2babb856-a5c2-4b60-a6bc-00d3f510f014): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 07 16:16:39 crc kubenswrapper[4716]: I1207 16:16:39.468287 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/01686458-debd-419a-90ec-cd27cc6953ec-cert\") pod \"infra-operator-controller-manager-78d48bff9d-pxs5r\" (UID: \"01686458-debd-419a-90ec-cd27cc6953ec\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-pxs5r" Dec 07 16:16:39 crc kubenswrapper[4716]: E1207 16:16:39.468465 4716 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 07 16:16:39 crc kubenswrapper[4716]: E1207 16:16:39.468733 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/01686458-debd-419a-90ec-cd27cc6953ec-cert podName:01686458-debd-419a-90ec-cd27cc6953ec nodeName:}" failed. No retries permitted until 2025-12-07 16:16:55.468707077 +0000 UTC m=+878.158992059 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/01686458-debd-419a-90ec-cd27cc6953ec-cert") pod "infra-operator-controller-manager-78d48bff9d-pxs5r" (UID: "01686458-debd-419a-90ec-cd27cc6953ec") : secret "infra-operator-webhook-server-cert" not found Dec 07 16:16:39 crc kubenswrapper[4716]: I1207 16:16:39.602727 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-srq27" event={"ID":"463e6bed-ce60-4e35-98a3-cd837b9066f2","Type":"ContainerStarted","Data":"2facf9203a41fe5a64c528e51c4dbeb5506ed14a363ba4393751e24d9f117a9b"} Dec 07 16:16:39 crc kubenswrapper[4716]: I1207 16:16:39.605698 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-9dsmr" event={"ID":"5be39543-890e-479c-a041-d864922e038f","Type":"ContainerStarted","Data":"e0840197c9fff2bb57e54a6eefec90dc79663b94fa234b0e5d11280bf80beb9f"} Dec 07 16:16:39 crc kubenswrapper[4716]: I1207 16:16:39.607577 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-h95ls" event={"ID":"1bf3b0b2-ea38-4fa3-b07c-4467dc111476","Type":"ContainerStarted","Data":"4466051ff51d6f482738fcc54dba929f7e09b7a5abaed854ce73f59b6f5fd786"} Dec 07 16:16:39 crc kubenswrapper[4716]: I1207 16:16:39.609609 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-967d97867-hx48c" event={"ID":"a929d511-1d07-429d-b302-8843a6834f52","Type":"ContainerStarted","Data":"bea467fc6b3739165cd70690a377088bdd2410a753c010b9d75ad84a456aaafa"} Dec 07 16:16:39 crc kubenswrapper[4716]: I1207 16:16:39.611024 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-rcmz5" event={"ID":"6f24594b-ebe7-4518-a067-45891924abe5","Type":"ContainerStarted","Data":"cd6922b5d189f6769d9eee29975afdbc7e96c6e8f61e697c97c7174c900010d6"} Dec 07 16:16:39 crc kubenswrapper[4716]: I1207 16:16:39.612514 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-kkdgl" event={"ID":"57c9705a-9bf1-45b4-aefb-d224a9d72a4a","Type":"ContainerStarted","Data":"e43770c5a4a1120e2b5c8c0518ef6d18a83d2e601ae2cefc918e07152005337f"} Dec 07 16:16:39 crc kubenswrapper[4716]: I1207 16:16:39.613754 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-xqnzw" event={"ID":"d9d5c8b8-240b-4c88-ad6e-b04881842f0b","Type":"ContainerStarted","Data":"50881fa6a58eabf3294c0c75960f1e6afc19281946f765fc37764f98bf4b5f43"} Dec 07 16:16:39 crc kubenswrapper[4716]: I1207 16:16:39.681742 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/34e0d69c-2f36-4ba6-a4e3-114cabe016c2-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879fdkpkx\" (UID: \"34e0d69c-2f36-4ba6-a4e3-114cabe016c2\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fdkpkx" Dec 07 16:16:39 crc kubenswrapper[4716]: E1207 16:16:39.681924 4716 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 07 16:16:39 crc kubenswrapper[4716]: E1207 16:16:39.681980 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/34e0d69c-2f36-4ba6-a4e3-114cabe016c2-cert podName:34e0d69c-2f36-4ba6-a4e3-114cabe016c2 nodeName:}" failed. No retries permitted until 2025-12-07 16:16:55.681961728 +0000 UTC m=+878.372246640 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/34e0d69c-2f36-4ba6-a4e3-114cabe016c2-cert") pod "openstack-baremetal-operator-controller-manager-84b575879fdkpkx" (UID: "34e0d69c-2f36-4ba6-a4e3-114cabe016c2") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 07 16:16:39 crc kubenswrapper[4716]: I1207 16:16:39.985889 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9d840ba3-d5ef-48fc-87b2-6af4d164c9ba-metrics-certs\") pod \"openstack-operator-controller-manager-7db7d7886d-z6n25\" (UID: \"9d840ba3-d5ef-48fc-87b2-6af4d164c9ba\") " pod="openstack-operators/openstack-operator-controller-manager-7db7d7886d-z6n25" Dec 07 16:16:39 crc kubenswrapper[4716]: I1207 16:16:39.985978 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/9d840ba3-d5ef-48fc-87b2-6af4d164c9ba-webhook-certs\") pod \"openstack-operator-controller-manager-7db7d7886d-z6n25\" (UID: \"9d840ba3-d5ef-48fc-87b2-6af4d164c9ba\") " pod="openstack-operators/openstack-operator-controller-manager-7db7d7886d-z6n25" Dec 07 16:16:39 crc kubenswrapper[4716]: E1207 16:16:39.986051 4716 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 07 16:16:39 crc kubenswrapper[4716]: E1207 16:16:39.986132 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9d840ba3-d5ef-48fc-87b2-6af4d164c9ba-metrics-certs podName:9d840ba3-d5ef-48fc-87b2-6af4d164c9ba nodeName:}" failed. No retries permitted until 2025-12-07 16:16:55.986115832 +0000 UTC m=+878.676400804 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9d840ba3-d5ef-48fc-87b2-6af4d164c9ba-metrics-certs") pod "openstack-operator-controller-manager-7db7d7886d-z6n25" (UID: "9d840ba3-d5ef-48fc-87b2-6af4d164c9ba") : secret "metrics-server-cert" not found Dec 07 16:16:39 crc kubenswrapper[4716]: E1207 16:16:39.986251 4716 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 07 16:16:39 crc kubenswrapper[4716]: E1207 16:16:39.986306 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9d840ba3-d5ef-48fc-87b2-6af4d164c9ba-webhook-certs podName:9d840ba3-d5ef-48fc-87b2-6af4d164c9ba nodeName:}" failed. No retries permitted until 2025-12-07 16:16:55.986291336 +0000 UTC m=+878.676576248 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/9d840ba3-d5ef-48fc-87b2-6af4d164c9ba-webhook-certs") pod "openstack-operator-controller-manager-7db7d7886d-z6n25" (UID: "9d840ba3-d5ef-48fc-87b2-6af4d164c9ba") : secret "webhook-server-cert" not found Dec 07 16:16:46 crc kubenswrapper[4716]: I1207 16:16:46.683850 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-nx5mv" event={"ID":"20067e63-712d-4f94-9019-627219d4299f","Type":"ContainerStarted","Data":"7eb67913ed4f6efb5228fd00750227bbd910d2f4f4b5d53be3cec1ab9915ce65"} Dec 07 16:16:46 crc kubenswrapper[4716]: I1207 16:16:46.697576 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-4nrlz" event={"ID":"62368956-27d2-41fa-a2b2-0fb49d869f11","Type":"ContainerStarted","Data":"7f995a14953972d407afdc318635f03473ad5e80752311526f1c46a3e2b795e2"} Dec 07 16:16:47 crc kubenswrapper[4716]: I1207 16:16:47.720537 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-m957c" event={"ID":"8765c8b2-a50b-4417-adbe-1174dcdfe172","Type":"ContainerStarted","Data":"bc3860c8f08768bb4d08109313e6fc5d3948536664a2909679c8fe57b6da51b4"} Dec 07 16:16:51 crc kubenswrapper[4716]: I1207 16:16:51.748820 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-4dnl8" event={"ID":"0c6a49cd-1cfe-49c8-a068-f85c735b46b1","Type":"ContainerStarted","Data":"edd40dc0ca4dd04c896d30fc979eea595388f8538b94db058beffa7c4a37b9ec"} Dec 07 16:16:52 crc kubenswrapper[4716]: E1207 16:16:52.208167 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-p2hbf" podUID="d9d5a1f4-edb7-4630-8907-5d29c4678f24" Dec 07 16:16:52 crc kubenswrapper[4716]: E1207 16:16:52.238155 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-zts6p" podUID="79562d20-950f-428e-ac3a-f78979053266" Dec 07 16:16:52 crc kubenswrapper[4716]: E1207 16:16:52.238433 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-wctdp" podUID="2babb856-a5c2-4b60-a6bc-00d3f510f014" Dec 07 16:16:52 crc kubenswrapper[4716]: I1207 16:16:52.758973 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-p2hbf" event={"ID":"d9d5a1f4-edb7-4630-8907-5d29c4678f24","Type":"ContainerStarted","Data":"fed8ca2d7f8182f0214ece51ad0f20ef43e3a8c08b78054027224c751a129ff3"} Dec 07 16:16:52 crc kubenswrapper[4716]: I1207 16:16:52.762454 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-4nrlz" event={"ID":"62368956-27d2-41fa-a2b2-0fb49d869f11","Type":"ContainerStarted","Data":"ab56d522db6e97fcbf85eed8478e9e2a67c99a041905462822d165ff0dd34472"} Dec 07 16:16:52 crc kubenswrapper[4716]: I1207 16:16:52.763337 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-4nrlz" Dec 07 16:16:52 crc kubenswrapper[4716]: I1207 16:16:52.770007 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-z7lcv" event={"ID":"69c2e54f-9af6-44e8-abdd-226a29b64da6","Type":"ContainerStarted","Data":"5ee9f3078467bc081e539ca16b58d530c39458d23641e41f61174d1e69483435"} Dec 07 16:16:52 crc kubenswrapper[4716]: I1207 16:16:52.772209 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-4nrlz" Dec 07 16:16:52 crc kubenswrapper[4716]: I1207 16:16:52.783454 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-zts6p" event={"ID":"79562d20-950f-428e-ac3a-f78979053266","Type":"ContainerStarted","Data":"e78df20599236df56697ef00a02b53845e28a0e0486768a5689b58630e36fadb"} Dec 07 16:16:52 crc kubenswrapper[4716]: I1207 16:16:52.793219 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-wctdp" event={"ID":"2babb856-a5c2-4b60-a6bc-00d3f510f014","Type":"ContainerStarted","Data":"6683d6927274380d94e303297b55abb6c19dbb02bc942a73fc609b57ed2ac328"} Dec 07 16:16:52 crc kubenswrapper[4716]: I1207 16:16:52.815528 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-nx5mv" event={"ID":"20067e63-712d-4f94-9019-627219d4299f","Type":"ContainerStarted","Data":"343ce05f64db34767ce2f0768a623e524c4e14e350cb1a19f197a81898beb07e"} Dec 07 16:16:52 crc kubenswrapper[4716]: I1207 16:16:52.816851 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-nx5mv" Dec 07 16:16:52 crc kubenswrapper[4716]: I1207 16:16:52.825422 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-nx5mv" Dec 07 16:16:52 crc kubenswrapper[4716]: I1207 16:16:52.826274 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-4nrlz" podStartSLOduration=2.975989482 podStartE2EDuration="29.826261397s" podCreationTimestamp="2025-12-07 16:16:23 +0000 UTC" firstStartedPulling="2025-12-07 16:16:24.935849032 +0000 UTC m=+847.626133944" lastFinishedPulling="2025-12-07 16:16:51.786120947 +0000 UTC m=+874.476405859" observedRunningTime="2025-12-07 16:16:52.824406365 +0000 UTC m=+875.514691277" watchObservedRunningTime="2025-12-07 16:16:52.826261397 +0000 UTC m=+875.516546309" Dec 07 16:16:52 crc kubenswrapper[4716]: I1207 16:16:52.872684 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-m957c" event={"ID":"8765c8b2-a50b-4417-adbe-1174dcdfe172","Type":"ContainerStarted","Data":"448cfa39496411472f0a693556e254fafe17d0872387bc324f3e67f51b0be184"} Dec 07 16:16:52 crc kubenswrapper[4716]: I1207 16:16:52.873991 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-m957c" Dec 07 16:16:52 crc kubenswrapper[4716]: I1207 16:16:52.886723 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-m957c" Dec 07 16:16:52 crc kubenswrapper[4716]: I1207 16:16:52.894650 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-nx5mv" podStartSLOduration=3.102704281 podStartE2EDuration="29.894634782s" podCreationTimestamp="2025-12-07 16:16:23 +0000 UTC" firstStartedPulling="2025-12-07 16:16:24.935811981 +0000 UTC m=+847.626096893" lastFinishedPulling="2025-12-07 16:16:51.727742482 +0000 UTC m=+874.418027394" observedRunningTime="2025-12-07 16:16:52.89384847 +0000 UTC m=+875.584133382" watchObservedRunningTime="2025-12-07 16:16:52.894634782 +0000 UTC m=+875.584919694" Dec 07 16:16:52 crc kubenswrapper[4716]: I1207 16:16:52.903898 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-srq27" event={"ID":"463e6bed-ce60-4e35-98a3-cd837b9066f2","Type":"ContainerStarted","Data":"0e48cbbb5112fdd57eb2bcd7fd40c7bde5ed49193e8efcd71ef44e7e6f5d52e8"} Dec 07 16:16:52 crc kubenswrapper[4716]: I1207 16:16:52.904342 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-srq27" Dec 07 16:16:52 crc kubenswrapper[4716]: I1207 16:16:52.910519 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-srq27" Dec 07 16:16:52 crc kubenswrapper[4716]: I1207 16:16:52.929814 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-8xrch" event={"ID":"0f05e35d-f818-4f65-bac6-f6a22006627e","Type":"ContainerStarted","Data":"6270965904ed6824600c29db25c00672731f27b5f8761ace1269290a2e58e67e"} Dec 07 16:16:52 crc kubenswrapper[4716]: I1207 16:16:52.953686 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-9dsmr" event={"ID":"5be39543-890e-479c-a041-d864922e038f","Type":"ContainerStarted","Data":"900738e705d4097e70096cf06c4d9b5fa66e590942a4ad16eb2168eb7f9167ed"} Dec 07 16:16:52 crc kubenswrapper[4716]: I1207 16:16:52.954538 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-9dsmr" Dec 07 16:16:52 crc kubenswrapper[4716]: I1207 16:16:52.961358 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-9dsmr" Dec 07 16:16:52 crc kubenswrapper[4716]: I1207 16:16:52.962888 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-kbzpp" event={"ID":"d3ef7831-ac99-410d-90bf-42d87042964b","Type":"ContainerStarted","Data":"72a641a003bf04984e3ba1da8296a8bc8e4410fcb6c3f267ad7ffbfa719db3b2"} Dec 07 16:16:52 crc kubenswrapper[4716]: I1207 16:16:52.962931 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-kbzpp" event={"ID":"d3ef7831-ac99-410d-90bf-42d87042964b","Type":"ContainerStarted","Data":"a690dc2087c3b6d561609676448f61809454542563818f3b9c1db355e07f317a"} Dec 07 16:16:52 crc kubenswrapper[4716]: I1207 16:16:52.963598 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-kbzpp" Dec 07 16:16:52 crc kubenswrapper[4716]: I1207 16:16:52.979168 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-ggcr7" event={"ID":"e4c2f2f6-3285-4f48-a428-b2065f33c046","Type":"ContainerStarted","Data":"0a0916b18a8753183e0c86d5e44fde040270046508b49fe82178fc7c9f39a91e"} Dec 07 16:16:52 crc kubenswrapper[4716]: I1207 16:16:52.979416 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-ggcr7" Dec 07 16:16:52 crc kubenswrapper[4716]: I1207 16:16:52.985012 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-h95ls" event={"ID":"1bf3b0b2-ea38-4fa3-b07c-4467dc111476","Type":"ContainerStarted","Data":"dffccf1432289bffd70bd351176f5eb5a8e910d7ce76345d7fe91717e939efba"} Dec 07 16:16:52 crc kubenswrapper[4716]: I1207 16:16:52.985795 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-h95ls" Dec 07 16:16:52 crc kubenswrapper[4716]: I1207 16:16:52.988412 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-h95ls" Dec 07 16:16:52 crc kubenswrapper[4716]: I1207 16:16:52.994880 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-m957c" podStartSLOduration=2.9382695500000002 podStartE2EDuration="29.994863674s" podCreationTimestamp="2025-12-07 16:16:23 +0000 UTC" firstStartedPulling="2025-12-07 16:16:24.717650883 +0000 UTC m=+847.407935795" lastFinishedPulling="2025-12-07 16:16:51.774245007 +0000 UTC m=+874.464529919" observedRunningTime="2025-12-07 16:16:52.970775503 +0000 UTC m=+875.661060415" watchObservedRunningTime="2025-12-07 16:16:52.994863674 +0000 UTC m=+875.685148586" Dec 07 16:16:53 crc kubenswrapper[4716]: I1207 16:16:53.021387 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-srq27" podStartSLOduration=3.585919304 podStartE2EDuration="30.021372152s" podCreationTimestamp="2025-12-07 16:16:23 +0000 UTC" firstStartedPulling="2025-12-07 16:16:25.257614587 +0000 UTC m=+847.947899499" lastFinishedPulling="2025-12-07 16:16:51.693067435 +0000 UTC m=+874.383352347" observedRunningTime="2025-12-07 16:16:53.020361045 +0000 UTC m=+875.710645957" watchObservedRunningTime="2025-12-07 16:16:53.021372152 +0000 UTC m=+875.711657064" Dec 07 16:16:53 crc kubenswrapper[4716]: I1207 16:16:53.024652 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-967d97867-hx48c" event={"ID":"a929d511-1d07-429d-b302-8843a6834f52","Type":"ContainerStarted","Data":"004248504a592c9ebf2dedfd1f3a421b9c8d639221a3952ad751e6619cc22ba8"} Dec 07 16:16:53 crc kubenswrapper[4716]: I1207 16:16:53.025332 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-967d97867-hx48c" Dec 07 16:16:53 crc kubenswrapper[4716]: I1207 16:16:53.061499 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-rcmz5" event={"ID":"6f24594b-ebe7-4518-a067-45891924abe5","Type":"ContainerStarted","Data":"1cb17b62bab845f7b1ebd0821bb20abf05f27a948621f6f3fa0085b4defb30a7"} Dec 07 16:16:53 crc kubenswrapper[4716]: I1207 16:16:53.063892 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-967d97867-hx48c" Dec 07 16:16:53 crc kubenswrapper[4716]: I1207 16:16:53.070137 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-rcmz5" Dec 07 16:16:53 crc kubenswrapper[4716]: I1207 16:16:53.075923 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-kkdgl" event={"ID":"57c9705a-9bf1-45b4-aefb-d224a9d72a4a","Type":"ContainerStarted","Data":"85ad47856ff49ec800034b80ccfa296a9fe9287021817248d292fde2da0bd0f0"} Dec 07 16:16:53 crc kubenswrapper[4716]: I1207 16:16:53.076972 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-998648c74-kkdgl" Dec 07 16:16:53 crc kubenswrapper[4716]: I1207 16:16:53.088269 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-8zwtk" event={"ID":"2ef47713-a73c-4c63-8fa0-ffc530832285","Type":"ContainerStarted","Data":"0482d98524df5ecfb7d3fd27a87cd410f7cf4c1a547fb41e04331a44cb1cdcfe"} Dec 07 16:16:53 crc kubenswrapper[4716]: I1207 16:16:53.089169 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-8zwtk" Dec 07 16:16:53 crc kubenswrapper[4716]: I1207 16:16:53.089367 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-rcmz5" Dec 07 16:16:53 crc kubenswrapper[4716]: I1207 16:16:53.093345 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-998648c74-kkdgl" Dec 07 16:16:53 crc kubenswrapper[4716]: I1207 16:16:53.101336 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-kbzpp" podStartSLOduration=3.644998201 podStartE2EDuration="30.10130701s" podCreationTimestamp="2025-12-07 16:16:23 +0000 UTC" firstStartedPulling="2025-12-07 16:16:25.148127757 +0000 UTC m=+847.838412669" lastFinishedPulling="2025-12-07 16:16:51.604436566 +0000 UTC m=+874.294721478" observedRunningTime="2025-12-07 16:16:53.044277801 +0000 UTC m=+875.734562713" watchObservedRunningTime="2025-12-07 16:16:53.10130701 +0000 UTC m=+875.791591932" Dec 07 16:16:53 crc kubenswrapper[4716]: I1207 16:16:53.104105 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-4dnl8" event={"ID":"0c6a49cd-1cfe-49c8-a068-f85c735b46b1","Type":"ContainerStarted","Data":"8578a5d04bdeec3f671e9992fdacbac6e00bf840d5709cab9602591da1f16fa6"} Dec 07 16:16:53 crc kubenswrapper[4716]: I1207 16:16:53.105478 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5854674fcc-4dnl8" Dec 07 16:16:53 crc kubenswrapper[4716]: I1207 16:16:53.130330 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-9dsmr" podStartSLOduration=3.55494159 podStartE2EDuration="30.130306247s" podCreationTimestamp="2025-12-07 16:16:23 +0000 UTC" firstStartedPulling="2025-12-07 16:16:25.125278739 +0000 UTC m=+847.815563661" lastFinishedPulling="2025-12-07 16:16:51.700643406 +0000 UTC m=+874.390928318" observedRunningTime="2025-12-07 16:16:53.089050739 +0000 UTC m=+875.779335651" watchObservedRunningTime="2025-12-07 16:16:53.130306247 +0000 UTC m=+875.820591159" Dec 07 16:16:53 crc kubenswrapper[4716]: I1207 16:16:53.154202 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-ggcr7" podStartSLOduration=3.800169844 podStartE2EDuration="30.154182783s" podCreationTimestamp="2025-12-07 16:16:23 +0000 UTC" firstStartedPulling="2025-12-07 16:16:25.272024118 +0000 UTC m=+847.962309030" lastFinishedPulling="2025-12-07 16:16:51.626037057 +0000 UTC m=+874.316321969" observedRunningTime="2025-12-07 16:16:53.114619701 +0000 UTC m=+875.804904613" watchObservedRunningTime="2025-12-07 16:16:53.154182783 +0000 UTC m=+875.844467695" Dec 07 16:16:53 crc kubenswrapper[4716]: I1207 16:16:53.156135 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-h95ls" podStartSLOduration=3.587993732 podStartE2EDuration="30.156128578s" podCreationTimestamp="2025-12-07 16:16:23 +0000 UTC" firstStartedPulling="2025-12-07 16:16:25.115716553 +0000 UTC m=+847.806001465" lastFinishedPulling="2025-12-07 16:16:51.683851399 +0000 UTC m=+874.374136311" observedRunningTime="2025-12-07 16:16:53.152199008 +0000 UTC m=+875.842483920" watchObservedRunningTime="2025-12-07 16:16:53.156128578 +0000 UTC m=+875.846413490" Dec 07 16:16:53 crc kubenswrapper[4716]: I1207 16:16:53.230753 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-967d97867-hx48c" podStartSLOduration=3.622216704 podStartE2EDuration="30.230729315s" podCreationTimestamp="2025-12-07 16:16:23 +0000 UTC" firstStartedPulling="2025-12-07 16:16:25.119229371 +0000 UTC m=+847.809514283" lastFinishedPulling="2025-12-07 16:16:51.727741982 +0000 UTC m=+874.418026894" observedRunningTime="2025-12-07 16:16:53.196155393 +0000 UTC m=+875.886440305" watchObservedRunningTime="2025-12-07 16:16:53.230729315 +0000 UTC m=+875.921014227" Dec 07 16:16:53 crc kubenswrapper[4716]: I1207 16:16:53.250433 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-998648c74-kkdgl" podStartSLOduration=3.653983181 podStartE2EDuration="30.250410534s" podCreationTimestamp="2025-12-07 16:16:23 +0000 UTC" firstStartedPulling="2025-12-07 16:16:25.105912371 +0000 UTC m=+847.796197283" lastFinishedPulling="2025-12-07 16:16:51.702339724 +0000 UTC m=+874.392624636" observedRunningTime="2025-12-07 16:16:53.228304898 +0000 UTC m=+875.918589810" watchObservedRunningTime="2025-12-07 16:16:53.250410534 +0000 UTC m=+875.940695446" Dec 07 16:16:53 crc kubenswrapper[4716]: I1207 16:16:53.270151 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-8zwtk" podStartSLOduration=3.909168039 podStartE2EDuration="30.270130533s" podCreationTimestamp="2025-12-07 16:16:23 +0000 UTC" firstStartedPulling="2025-12-07 16:16:25.263987404 +0000 UTC m=+847.954272306" lastFinishedPulling="2025-12-07 16:16:51.624949888 +0000 UTC m=+874.315234800" observedRunningTime="2025-12-07 16:16:53.260536916 +0000 UTC m=+875.950821818" watchObservedRunningTime="2025-12-07 16:16:53.270130533 +0000 UTC m=+875.960415445" Dec 07 16:16:53 crc kubenswrapper[4716]: I1207 16:16:53.296162 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-rcmz5" podStartSLOduration=3.430743672 podStartE2EDuration="30.296142579s" podCreationTimestamp="2025-12-07 16:16:23 +0000 UTC" firstStartedPulling="2025-12-07 16:16:24.864295428 +0000 UTC m=+847.554580340" lastFinishedPulling="2025-12-07 16:16:51.729694335 +0000 UTC m=+874.419979247" observedRunningTime="2025-12-07 16:16:53.283189897 +0000 UTC m=+875.973474809" watchObservedRunningTime="2025-12-07 16:16:53.296142579 +0000 UTC m=+875.986427491" Dec 07 16:16:54 crc kubenswrapper[4716]: I1207 16:16:54.111816 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-ggcr7" event={"ID":"e4c2f2f6-3285-4f48-a428-b2065f33c046","Type":"ContainerStarted","Data":"35ae9301259f24ecfcaca5036b569bb4dbca6c9fb22da6aeb6dc30a1982261b0"} Dec 07 16:16:54 crc kubenswrapper[4716]: I1207 16:16:54.114892 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-8xrch" event={"ID":"0f05e35d-f818-4f65-bac6-f6a22006627e","Type":"ContainerStarted","Data":"fc8aa85b77c131316f871409054511e83c26af99ef21b53b464cb99ceedbbad1"} Dec 07 16:16:54 crc kubenswrapper[4716]: I1207 16:16:54.114959 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-78f8948974-8xrch" Dec 07 16:16:54 crc kubenswrapper[4716]: I1207 16:16:54.116568 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-p2hbf" event={"ID":"d9d5a1f4-edb7-4630-8907-5d29c4678f24","Type":"ContainerStarted","Data":"a0cfd4e8fc200c09ac96d5a5cc5e4fd4d995ea7e6b0d81fad35059b7686e10df"} Dec 07 16:16:54 crc kubenswrapper[4716]: I1207 16:16:54.116644 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-p2hbf" Dec 07 16:16:54 crc kubenswrapper[4716]: I1207 16:16:54.118774 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-8zwtk" event={"ID":"2ef47713-a73c-4c63-8fa0-ffc530832285","Type":"ContainerStarted","Data":"24fe012c5edfe51e1f70ba8c4d93882cb167e7c871281ea3ec420ca2399e074e"} Dec 07 16:16:54 crc kubenswrapper[4716]: I1207 16:16:54.124319 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-7mjmm" event={"ID":"79aa5696-34a1-45cf-b965-56f64fe63f4e","Type":"ContainerStarted","Data":"4dd21b779dfa9239a677a0634d4f4acca340baaf036fca0967c9e41056f7242d"} Dec 07 16:16:54 crc kubenswrapper[4716]: I1207 16:16:54.127685 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-xqnzw" event={"ID":"d9d5c8b8-240b-4c88-ad6e-b04881842f0b","Type":"ContainerStarted","Data":"5ddad8f9f3264b87bbb0aec4af931e25736a3edf15faefd47a942e019c142b84"} Dec 07 16:16:54 crc kubenswrapper[4716]: I1207 16:16:54.127947 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-xqnzw" Dec 07 16:16:54 crc kubenswrapper[4716]: I1207 16:16:54.133525 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-z7lcv" event={"ID":"69c2e54f-9af6-44e8-abdd-226a29b64da6","Type":"ContainerStarted","Data":"81b43e9a6a1c628d137cb250ac91a3893c9be24e01d1656d47768a111c1ecb07"} Dec 07 16:16:54 crc kubenswrapper[4716]: I1207 16:16:54.133610 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-xqnzw" Dec 07 16:16:54 crc kubenswrapper[4716]: I1207 16:16:54.133902 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-z7lcv" Dec 07 16:16:54 crc kubenswrapper[4716]: I1207 16:16:54.137821 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-zts6p" event={"ID":"79562d20-950f-428e-ac3a-f78979053266","Type":"ContainerStarted","Data":"bb46ad63b2fdb18c51bd574ce05cf118570cfcba9cf010ec4a2af6951bb7cbc0"} Dec 07 16:16:54 crc kubenswrapper[4716]: I1207 16:16:54.137929 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-zts6p" Dec 07 16:16:54 crc kubenswrapper[4716]: I1207 16:16:54.139580 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-wctdp" event={"ID":"2babb856-a5c2-4b60-a6bc-00d3f510f014","Type":"ContainerStarted","Data":"56b3f25a5d99654ae3bd2578612164709ecbb1685418b1663d75d85ebc622228"} Dec 07 16:16:54 crc kubenswrapper[4716]: I1207 16:16:54.145308 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5854674fcc-4dnl8" podStartSLOduration=9.628367221 podStartE2EDuration="31.145291816s" podCreationTimestamp="2025-12-07 16:16:23 +0000 UTC" firstStartedPulling="2025-12-07 16:16:25.300952504 +0000 UTC m=+847.991237416" lastFinishedPulling="2025-12-07 16:16:46.817877099 +0000 UTC m=+869.508162011" observedRunningTime="2025-12-07 16:16:53.318741198 +0000 UTC m=+876.009026110" watchObservedRunningTime="2025-12-07 16:16:54.145291816 +0000 UTC m=+876.835576728" Dec 07 16:16:54 crc kubenswrapper[4716]: I1207 16:16:54.146534 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-78f8948974-8xrch" podStartSLOduration=4.807771175 podStartE2EDuration="31.14653026s" podCreationTimestamp="2025-12-07 16:16:23 +0000 UTC" firstStartedPulling="2025-12-07 16:16:25.287224291 +0000 UTC m=+847.977509193" lastFinishedPulling="2025-12-07 16:16:51.625983346 +0000 UTC m=+874.316268278" observedRunningTime="2025-12-07 16:16:54.144120673 +0000 UTC m=+876.834405605" watchObservedRunningTime="2025-12-07 16:16:54.14653026 +0000 UTC m=+876.836815172" Dec 07 16:16:54 crc kubenswrapper[4716]: I1207 16:16:54.159501 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-xqnzw" podStartSLOduration=4.176507839 podStartE2EDuration="31.159482162s" podCreationTimestamp="2025-12-07 16:16:23 +0000 UTC" firstStartedPulling="2025-12-07 16:16:24.715424101 +0000 UTC m=+847.405709013" lastFinishedPulling="2025-12-07 16:16:51.698398424 +0000 UTC m=+874.388683336" observedRunningTime="2025-12-07 16:16:54.158104553 +0000 UTC m=+876.848389465" watchObservedRunningTime="2025-12-07 16:16:54.159482162 +0000 UTC m=+876.849767064" Dec 07 16:16:54 crc kubenswrapper[4716]: I1207 16:16:54.187475 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-p2hbf" podStartSLOduration=3.086771938 podStartE2EDuration="31.187456551s" podCreationTimestamp="2025-12-07 16:16:23 +0000 UTC" firstStartedPulling="2025-12-07 16:16:25.138601841 +0000 UTC m=+847.828886753" lastFinishedPulling="2025-12-07 16:16:53.239286454 +0000 UTC m=+875.929571366" observedRunningTime="2025-12-07 16:16:54.181491805 +0000 UTC m=+876.871776717" watchObservedRunningTime="2025-12-07 16:16:54.187456551 +0000 UTC m=+876.877741473" Dec 07 16:16:54 crc kubenswrapper[4716]: I1207 16:16:54.212039 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-wctdp" podStartSLOduration=2.737107956 podStartE2EDuration="31.212022975s" podCreationTimestamp="2025-12-07 16:16:23 +0000 UTC" firstStartedPulling="2025-12-07 16:16:24.87152793 +0000 UTC m=+847.561812842" lastFinishedPulling="2025-12-07 16:16:53.346442949 +0000 UTC m=+876.036727861" observedRunningTime="2025-12-07 16:16:54.20823529 +0000 UTC m=+876.898520202" watchObservedRunningTime="2025-12-07 16:16:54.212022975 +0000 UTC m=+876.902307887" Dec 07 16:16:54 crc kubenswrapper[4716]: I1207 16:16:54.229245 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-zts6p" podStartSLOduration=2.495909337 podStartE2EDuration="31.229229015s" podCreationTimestamp="2025-12-07 16:16:23 +0000 UTC" firstStartedPulling="2025-12-07 16:16:24.612899915 +0000 UTC m=+847.303184827" lastFinishedPulling="2025-12-07 16:16:53.346219593 +0000 UTC m=+876.036504505" observedRunningTime="2025-12-07 16:16:54.226439337 +0000 UTC m=+876.916724249" watchObservedRunningTime="2025-12-07 16:16:54.229229015 +0000 UTC m=+876.919513927" Dec 07 16:16:54 crc kubenswrapper[4716]: I1207 16:16:54.239617 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-7mjmm" podStartSLOduration=3.523024702 podStartE2EDuration="30.239571392s" podCreationTimestamp="2025-12-07 16:16:24 +0000 UTC" firstStartedPulling="2025-12-07 16:16:25.481584797 +0000 UTC m=+848.171869729" lastFinishedPulling="2025-12-07 16:16:52.198131507 +0000 UTC m=+874.888416419" observedRunningTime="2025-12-07 16:16:54.237709371 +0000 UTC m=+876.927994283" watchObservedRunningTime="2025-12-07 16:16:54.239571392 +0000 UTC m=+876.929856304" Dec 07 16:16:54 crc kubenswrapper[4716]: I1207 16:16:54.257896 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-z7lcv" podStartSLOduration=4.909695226 podStartE2EDuration="31.257877043s" podCreationTimestamp="2025-12-07 16:16:23 +0000 UTC" firstStartedPulling="2025-12-07 16:16:25.261364971 +0000 UTC m=+847.951649883" lastFinishedPulling="2025-12-07 16:16:51.609546798 +0000 UTC m=+874.299831700" observedRunningTime="2025-12-07 16:16:54.254143878 +0000 UTC m=+876.944428790" watchObservedRunningTime="2025-12-07 16:16:54.257877043 +0000 UTC m=+876.948161955" Dec 07 16:16:55 crc kubenswrapper[4716]: I1207 16:16:55.146879 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-wctdp" Dec 07 16:16:55 crc kubenswrapper[4716]: I1207 16:16:55.535577 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/01686458-debd-419a-90ec-cd27cc6953ec-cert\") pod \"infra-operator-controller-manager-78d48bff9d-pxs5r\" (UID: \"01686458-debd-419a-90ec-cd27cc6953ec\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-pxs5r" Dec 07 16:16:55 crc kubenswrapper[4716]: I1207 16:16:55.545195 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/01686458-debd-419a-90ec-cd27cc6953ec-cert\") pod \"infra-operator-controller-manager-78d48bff9d-pxs5r\" (UID: \"01686458-debd-419a-90ec-cd27cc6953ec\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-pxs5r" Dec 07 16:16:55 crc kubenswrapper[4716]: I1207 16:16:55.577741 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-pxs5r" Dec 07 16:16:55 crc kubenswrapper[4716]: I1207 16:16:55.738729 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/34e0d69c-2f36-4ba6-a4e3-114cabe016c2-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879fdkpkx\" (UID: \"34e0d69c-2f36-4ba6-a4e3-114cabe016c2\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fdkpkx" Dec 07 16:16:55 crc kubenswrapper[4716]: I1207 16:16:55.747826 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/34e0d69c-2f36-4ba6-a4e3-114cabe016c2-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879fdkpkx\" (UID: \"34e0d69c-2f36-4ba6-a4e3-114cabe016c2\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fdkpkx" Dec 07 16:16:55 crc kubenswrapper[4716]: I1207 16:16:55.773601 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fdkpkx" Dec 07 16:16:55 crc kubenswrapper[4716]: I1207 16:16:55.987950 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-78d48bff9d-pxs5r"] Dec 07 16:16:55 crc kubenswrapper[4716]: W1207 16:16:55.988066 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod01686458_debd_419a_90ec_cd27cc6953ec.slice/crio-2dd9fab20bdf46f5efd3d3a361716e1c3f31f9ccb22c382120a817951e885031 WatchSource:0}: Error finding container 2dd9fab20bdf46f5efd3d3a361716e1c3f31f9ccb22c382120a817951e885031: Status 404 returned error can't find the container with id 2dd9fab20bdf46f5efd3d3a361716e1c3f31f9ccb22c382120a817951e885031 Dec 07 16:16:56 crc kubenswrapper[4716]: I1207 16:16:56.042256 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/9d840ba3-d5ef-48fc-87b2-6af4d164c9ba-webhook-certs\") pod \"openstack-operator-controller-manager-7db7d7886d-z6n25\" (UID: \"9d840ba3-d5ef-48fc-87b2-6af4d164c9ba\") " pod="openstack-operators/openstack-operator-controller-manager-7db7d7886d-z6n25" Dec 07 16:16:56 crc kubenswrapper[4716]: I1207 16:16:56.042335 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9d840ba3-d5ef-48fc-87b2-6af4d164c9ba-metrics-certs\") pod \"openstack-operator-controller-manager-7db7d7886d-z6n25\" (UID: \"9d840ba3-d5ef-48fc-87b2-6af4d164c9ba\") " pod="openstack-operators/openstack-operator-controller-manager-7db7d7886d-z6n25" Dec 07 16:16:56 crc kubenswrapper[4716]: I1207 16:16:56.045959 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9d840ba3-d5ef-48fc-87b2-6af4d164c9ba-metrics-certs\") pod \"openstack-operator-controller-manager-7db7d7886d-z6n25\" (UID: \"9d840ba3-d5ef-48fc-87b2-6af4d164c9ba\") " pod="openstack-operators/openstack-operator-controller-manager-7db7d7886d-z6n25" Dec 07 16:16:56 crc kubenswrapper[4716]: I1207 16:16:56.046363 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/9d840ba3-d5ef-48fc-87b2-6af4d164c9ba-webhook-certs\") pod \"openstack-operator-controller-manager-7db7d7886d-z6n25\" (UID: \"9d840ba3-d5ef-48fc-87b2-6af4d164c9ba\") " pod="openstack-operators/openstack-operator-controller-manager-7db7d7886d-z6n25" Dec 07 16:16:56 crc kubenswrapper[4716]: I1207 16:16:56.102746 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-7db7d7886d-z6n25" Dec 07 16:16:56 crc kubenswrapper[4716]: I1207 16:16:56.151848 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-pxs5r" event={"ID":"01686458-debd-419a-90ec-cd27cc6953ec","Type":"ContainerStarted","Data":"2dd9fab20bdf46f5efd3d3a361716e1c3f31f9ccb22c382120a817951e885031"} Dec 07 16:16:56 crc kubenswrapper[4716]: I1207 16:16:56.264746 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fdkpkx"] Dec 07 16:16:56 crc kubenswrapper[4716]: W1207 16:16:56.282038 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod34e0d69c_2f36_4ba6_a4e3_114cabe016c2.slice/crio-2899e63732a1f68c45b7dbca04959b32a454568253be7df9068007db852a45d7 WatchSource:0}: Error finding container 2899e63732a1f68c45b7dbca04959b32a454568253be7df9068007db852a45d7: Status 404 returned error can't find the container with id 2899e63732a1f68c45b7dbca04959b32a454568253be7df9068007db852a45d7 Dec 07 16:16:56 crc kubenswrapper[4716]: I1207 16:16:56.377610 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7db7d7886d-z6n25"] Dec 07 16:16:57 crc kubenswrapper[4716]: I1207 16:16:57.159197 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fdkpkx" event={"ID":"34e0d69c-2f36-4ba6-a4e3-114cabe016c2","Type":"ContainerStarted","Data":"2899e63732a1f68c45b7dbca04959b32a454568253be7df9068007db852a45d7"} Dec 07 16:16:57 crc kubenswrapper[4716]: I1207 16:16:57.161462 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-7db7d7886d-z6n25" event={"ID":"9d840ba3-d5ef-48fc-87b2-6af4d164c9ba","Type":"ContainerStarted","Data":"6213293fce32567682da11a4f6669d4d555afb58266321179225ef09aa6c4708"} Dec 07 16:16:57 crc kubenswrapper[4716]: I1207 16:16:57.161483 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-7db7d7886d-z6n25" event={"ID":"9d840ba3-d5ef-48fc-87b2-6af4d164c9ba","Type":"ContainerStarted","Data":"e2f4616556a37ad8916912ca42777d69c13c314b7932f89846653977f52b6881"} Dec 07 16:16:57 crc kubenswrapper[4716]: I1207 16:16:57.161629 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-7db7d7886d-z6n25" Dec 07 16:16:57 crc kubenswrapper[4716]: I1207 16:16:57.685420 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-7db7d7886d-z6n25" podStartSLOduration=33.685398805 podStartE2EDuration="33.685398805s" podCreationTimestamp="2025-12-07 16:16:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:16:57.186046163 +0000 UTC m=+879.876331075" watchObservedRunningTime="2025-12-07 16:16:57.685398805 +0000 UTC m=+880.375683707" Dec 07 16:16:59 crc kubenswrapper[4716]: I1207 16:16:59.179863 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-pxs5r" event={"ID":"01686458-debd-419a-90ec-cd27cc6953ec","Type":"ContainerStarted","Data":"93b65a70dfdd38ccfa11763835fd0cee213a5b044a195e9b63ef1992e21e1a47"} Dec 07 16:16:59 crc kubenswrapper[4716]: I1207 16:16:59.180469 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-pxs5r" event={"ID":"01686458-debd-419a-90ec-cd27cc6953ec","Type":"ContainerStarted","Data":"ea5ee5f0107931afca9d73694001d14a77e1edbcbbf4148bb4639bd1595e53b4"} Dec 07 16:16:59 crc kubenswrapper[4716]: I1207 16:16:59.180497 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-pxs5r" Dec 07 16:16:59 crc kubenswrapper[4716]: I1207 16:16:59.182025 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fdkpkx" event={"ID":"34e0d69c-2f36-4ba6-a4e3-114cabe016c2","Type":"ContainerStarted","Data":"e1c3a4f2591cc169efe737be4a2740077b2aa9838989cc3a3f0552727d34f2af"} Dec 07 16:16:59 crc kubenswrapper[4716]: I1207 16:16:59.182058 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fdkpkx" event={"ID":"34e0d69c-2f36-4ba6-a4e3-114cabe016c2","Type":"ContainerStarted","Data":"d8f55c1ead919ff38b2a707810c81d3e7e567007d0357b900914f80da16d7c18"} Dec 07 16:16:59 crc kubenswrapper[4716]: I1207 16:16:59.182895 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fdkpkx" Dec 07 16:16:59 crc kubenswrapper[4716]: I1207 16:16:59.209795 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-pxs5r" podStartSLOduration=33.466166246 podStartE2EDuration="36.209774575s" podCreationTimestamp="2025-12-07 16:16:23 +0000 UTC" firstStartedPulling="2025-12-07 16:16:55.98999781 +0000 UTC m=+878.680282722" lastFinishedPulling="2025-12-07 16:16:58.733606139 +0000 UTC m=+881.423891051" observedRunningTime="2025-12-07 16:16:59.20310712 +0000 UTC m=+881.893392042" watchObservedRunningTime="2025-12-07 16:16:59.209774575 +0000 UTC m=+881.900059497" Dec 07 16:16:59 crc kubenswrapper[4716]: I1207 16:16:59.237322 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fdkpkx" podStartSLOduration=33.772897751 podStartE2EDuration="36.237301412s" podCreationTimestamp="2025-12-07 16:16:23 +0000 UTC" firstStartedPulling="2025-12-07 16:16:56.284788333 +0000 UTC m=+878.975073245" lastFinishedPulling="2025-12-07 16:16:58.749191994 +0000 UTC m=+881.439476906" observedRunningTime="2025-12-07 16:16:59.231762048 +0000 UTC m=+881.922046990" watchObservedRunningTime="2025-12-07 16:16:59.237301412 +0000 UTC m=+881.927586324" Dec 07 16:16:59 crc kubenswrapper[4716]: I1207 16:16:59.424165 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-j9fcv"] Dec 07 16:16:59 crc kubenswrapper[4716]: I1207 16:16:59.425921 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-j9fcv" Dec 07 16:16:59 crc kubenswrapper[4716]: I1207 16:16:59.437597 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-j9fcv"] Dec 07 16:16:59 crc kubenswrapper[4716]: I1207 16:16:59.492355 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtgjc\" (UniqueName: \"kubernetes.io/projected/1a4797ee-7747-4d9d-abc6-113b15d0e888-kube-api-access-xtgjc\") pod \"redhat-marketplace-j9fcv\" (UID: \"1a4797ee-7747-4d9d-abc6-113b15d0e888\") " pod="openshift-marketplace/redhat-marketplace-j9fcv" Dec 07 16:16:59 crc kubenswrapper[4716]: I1207 16:16:59.492402 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a4797ee-7747-4d9d-abc6-113b15d0e888-catalog-content\") pod \"redhat-marketplace-j9fcv\" (UID: \"1a4797ee-7747-4d9d-abc6-113b15d0e888\") " pod="openshift-marketplace/redhat-marketplace-j9fcv" Dec 07 16:16:59 crc kubenswrapper[4716]: I1207 16:16:59.492426 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a4797ee-7747-4d9d-abc6-113b15d0e888-utilities\") pod \"redhat-marketplace-j9fcv\" (UID: \"1a4797ee-7747-4d9d-abc6-113b15d0e888\") " pod="openshift-marketplace/redhat-marketplace-j9fcv" Dec 07 16:16:59 crc kubenswrapper[4716]: I1207 16:16:59.593435 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xtgjc\" (UniqueName: \"kubernetes.io/projected/1a4797ee-7747-4d9d-abc6-113b15d0e888-kube-api-access-xtgjc\") pod \"redhat-marketplace-j9fcv\" (UID: \"1a4797ee-7747-4d9d-abc6-113b15d0e888\") " pod="openshift-marketplace/redhat-marketplace-j9fcv" Dec 07 16:16:59 crc kubenswrapper[4716]: I1207 16:16:59.593490 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a4797ee-7747-4d9d-abc6-113b15d0e888-catalog-content\") pod \"redhat-marketplace-j9fcv\" (UID: \"1a4797ee-7747-4d9d-abc6-113b15d0e888\") " pod="openshift-marketplace/redhat-marketplace-j9fcv" Dec 07 16:16:59 crc kubenswrapper[4716]: I1207 16:16:59.593516 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a4797ee-7747-4d9d-abc6-113b15d0e888-utilities\") pod \"redhat-marketplace-j9fcv\" (UID: \"1a4797ee-7747-4d9d-abc6-113b15d0e888\") " pod="openshift-marketplace/redhat-marketplace-j9fcv" Dec 07 16:16:59 crc kubenswrapper[4716]: I1207 16:16:59.593894 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a4797ee-7747-4d9d-abc6-113b15d0e888-catalog-content\") pod \"redhat-marketplace-j9fcv\" (UID: \"1a4797ee-7747-4d9d-abc6-113b15d0e888\") " pod="openshift-marketplace/redhat-marketplace-j9fcv" Dec 07 16:16:59 crc kubenswrapper[4716]: I1207 16:16:59.593969 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a4797ee-7747-4d9d-abc6-113b15d0e888-utilities\") pod \"redhat-marketplace-j9fcv\" (UID: \"1a4797ee-7747-4d9d-abc6-113b15d0e888\") " pod="openshift-marketplace/redhat-marketplace-j9fcv" Dec 07 16:16:59 crc kubenswrapper[4716]: I1207 16:16:59.620013 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtgjc\" (UniqueName: \"kubernetes.io/projected/1a4797ee-7747-4d9d-abc6-113b15d0e888-kube-api-access-xtgjc\") pod \"redhat-marketplace-j9fcv\" (UID: \"1a4797ee-7747-4d9d-abc6-113b15d0e888\") " pod="openshift-marketplace/redhat-marketplace-j9fcv" Dec 07 16:16:59 crc kubenswrapper[4716]: I1207 16:16:59.758307 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-j9fcv" Dec 07 16:17:00 crc kubenswrapper[4716]: I1207 16:17:00.227429 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-j9fcv"] Dec 07 16:17:01 crc kubenswrapper[4716]: I1207 16:17:01.199719 4716 generic.go:334] "Generic (PLEG): container finished" podID="1a4797ee-7747-4d9d-abc6-113b15d0e888" containerID="7ed1b4ee1ac483926ce033dc3e1fd5e3ae953f469d440cdb0b2c38d0791e2c5a" exitCode=0 Dec 07 16:17:01 crc kubenswrapper[4716]: I1207 16:17:01.199830 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j9fcv" event={"ID":"1a4797ee-7747-4d9d-abc6-113b15d0e888","Type":"ContainerDied","Data":"7ed1b4ee1ac483926ce033dc3e1fd5e3ae953f469d440cdb0b2c38d0791e2c5a"} Dec 07 16:17:01 crc kubenswrapper[4716]: I1207 16:17:01.200100 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j9fcv" event={"ID":"1a4797ee-7747-4d9d-abc6-113b15d0e888","Type":"ContainerStarted","Data":"b5b215031c6c7fbf58b6956a0ed7f4c3cef7e0a968b693138786ddec3f647bcc"} Dec 07 16:17:02 crc kubenswrapper[4716]: I1207 16:17:02.213697 4716 generic.go:334] "Generic (PLEG): container finished" podID="1a4797ee-7747-4d9d-abc6-113b15d0e888" containerID="13a9eb0edbdd0ccf8593e2f7ef2745501398a457200b3738778283f36a1037a7" exitCode=0 Dec 07 16:17:02 crc kubenswrapper[4716]: I1207 16:17:02.213773 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j9fcv" event={"ID":"1a4797ee-7747-4d9d-abc6-113b15d0e888","Type":"ContainerDied","Data":"13a9eb0edbdd0ccf8593e2f7ef2745501398a457200b3738778283f36a1037a7"} Dec 07 16:17:03 crc kubenswrapper[4716]: I1207 16:17:03.228300 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j9fcv" event={"ID":"1a4797ee-7747-4d9d-abc6-113b15d0e888","Type":"ContainerStarted","Data":"b31e92f451922389d848048a309c5efc59a1bd580aa2b19d89dccd5074400baf"} Dec 07 16:17:03 crc kubenswrapper[4716]: I1207 16:17:03.264835 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-j9fcv" podStartSLOduration=2.830565044 podStartE2EDuration="4.264817173s" podCreationTimestamp="2025-12-07 16:16:59 +0000 UTC" firstStartedPulling="2025-12-07 16:17:01.203140163 +0000 UTC m=+883.893425075" lastFinishedPulling="2025-12-07 16:17:02.637392262 +0000 UTC m=+885.327677204" observedRunningTime="2025-12-07 16:17:03.25288118 +0000 UTC m=+885.943166112" watchObservedRunningTime="2025-12-07 16:17:03.264817173 +0000 UTC m=+885.955102095" Dec 07 16:17:03 crc kubenswrapper[4716]: I1207 16:17:03.702154 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-zts6p" Dec 07 16:17:03 crc kubenswrapper[4716]: I1207 16:17:03.923925 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-wctdp" Dec 07 16:17:03 crc kubenswrapper[4716]: I1207 16:17:03.939377 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-kbzpp" Dec 07 16:17:04 crc kubenswrapper[4716]: I1207 16:17:04.126731 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-p2hbf" Dec 07 16:17:04 crc kubenswrapper[4716]: I1207 16:17:04.259268 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5854674fcc-4dnl8" Dec 07 16:17:04 crc kubenswrapper[4716]: I1207 16:17:04.279405 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-8zwtk" Dec 07 16:17:04 crc kubenswrapper[4716]: I1207 16:17:04.396589 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-78f8948974-8xrch" Dec 07 16:17:04 crc kubenswrapper[4716]: I1207 16:17:04.408736 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-z7lcv" Dec 07 16:17:04 crc kubenswrapper[4716]: I1207 16:17:04.542897 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-ggcr7" Dec 07 16:17:05 crc kubenswrapper[4716]: I1207 16:17:05.587535 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-pxs5r" Dec 07 16:17:05 crc kubenswrapper[4716]: I1207 16:17:05.781988 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fdkpkx" Dec 07 16:17:06 crc kubenswrapper[4716]: I1207 16:17:06.112816 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-7db7d7886d-z6n25" Dec 07 16:17:09 crc kubenswrapper[4716]: I1207 16:17:09.758469 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-j9fcv" Dec 07 16:17:09 crc kubenswrapper[4716]: I1207 16:17:09.759157 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-j9fcv" Dec 07 16:17:09 crc kubenswrapper[4716]: I1207 16:17:09.813169 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-j9fcv" Dec 07 16:17:10 crc kubenswrapper[4716]: I1207 16:17:10.352105 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-j9fcv" Dec 07 16:17:10 crc kubenswrapper[4716]: I1207 16:17:10.413360 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-j9fcv"] Dec 07 16:17:12 crc kubenswrapper[4716]: I1207 16:17:12.307701 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-j9fcv" podUID="1a4797ee-7747-4d9d-abc6-113b15d0e888" containerName="registry-server" containerID="cri-o://b31e92f451922389d848048a309c5efc59a1bd580aa2b19d89dccd5074400baf" gracePeriod=2 Dec 07 16:17:13 crc kubenswrapper[4716]: I1207 16:17:13.319301 4716 generic.go:334] "Generic (PLEG): container finished" podID="1a4797ee-7747-4d9d-abc6-113b15d0e888" containerID="b31e92f451922389d848048a309c5efc59a1bd580aa2b19d89dccd5074400baf" exitCode=0 Dec 07 16:17:13 crc kubenswrapper[4716]: I1207 16:17:13.319355 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j9fcv" event={"ID":"1a4797ee-7747-4d9d-abc6-113b15d0e888","Type":"ContainerDied","Data":"b31e92f451922389d848048a309c5efc59a1bd580aa2b19d89dccd5074400baf"} Dec 07 16:17:17 crc kubenswrapper[4716]: I1207 16:17:17.775893 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-2q7l4"] Dec 07 16:17:17 crc kubenswrapper[4716]: I1207 16:17:17.778823 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2q7l4" Dec 07 16:17:17 crc kubenswrapper[4716]: I1207 16:17:17.845182 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2q7l4"] Dec 07 16:17:17 crc kubenswrapper[4716]: I1207 16:17:17.859693 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f98dfb6b-5a13-42e4-bca1-030f964b78ce-utilities\") pod \"redhat-operators-2q7l4\" (UID: \"f98dfb6b-5a13-42e4-bca1-030f964b78ce\") " pod="openshift-marketplace/redhat-operators-2q7l4" Dec 07 16:17:17 crc kubenswrapper[4716]: I1207 16:17:17.859764 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twkvk\" (UniqueName: \"kubernetes.io/projected/f98dfb6b-5a13-42e4-bca1-030f964b78ce-kube-api-access-twkvk\") pod \"redhat-operators-2q7l4\" (UID: \"f98dfb6b-5a13-42e4-bca1-030f964b78ce\") " pod="openshift-marketplace/redhat-operators-2q7l4" Dec 07 16:17:17 crc kubenswrapper[4716]: I1207 16:17:17.859969 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f98dfb6b-5a13-42e4-bca1-030f964b78ce-catalog-content\") pod \"redhat-operators-2q7l4\" (UID: \"f98dfb6b-5a13-42e4-bca1-030f964b78ce\") " pod="openshift-marketplace/redhat-operators-2q7l4" Dec 07 16:17:17 crc kubenswrapper[4716]: I1207 16:17:17.960985 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f98dfb6b-5a13-42e4-bca1-030f964b78ce-utilities\") pod \"redhat-operators-2q7l4\" (UID: \"f98dfb6b-5a13-42e4-bca1-030f964b78ce\") " pod="openshift-marketplace/redhat-operators-2q7l4" Dec 07 16:17:17 crc kubenswrapper[4716]: I1207 16:17:17.961028 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twkvk\" (UniqueName: \"kubernetes.io/projected/f98dfb6b-5a13-42e4-bca1-030f964b78ce-kube-api-access-twkvk\") pod \"redhat-operators-2q7l4\" (UID: \"f98dfb6b-5a13-42e4-bca1-030f964b78ce\") " pod="openshift-marketplace/redhat-operators-2q7l4" Dec 07 16:17:17 crc kubenswrapper[4716]: I1207 16:17:17.961112 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f98dfb6b-5a13-42e4-bca1-030f964b78ce-catalog-content\") pod \"redhat-operators-2q7l4\" (UID: \"f98dfb6b-5a13-42e4-bca1-030f964b78ce\") " pod="openshift-marketplace/redhat-operators-2q7l4" Dec 07 16:17:17 crc kubenswrapper[4716]: I1207 16:17:17.961552 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f98dfb6b-5a13-42e4-bca1-030f964b78ce-catalog-content\") pod \"redhat-operators-2q7l4\" (UID: \"f98dfb6b-5a13-42e4-bca1-030f964b78ce\") " pod="openshift-marketplace/redhat-operators-2q7l4" Dec 07 16:17:17 crc kubenswrapper[4716]: I1207 16:17:17.961620 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f98dfb6b-5a13-42e4-bca1-030f964b78ce-utilities\") pod \"redhat-operators-2q7l4\" (UID: \"f98dfb6b-5a13-42e4-bca1-030f964b78ce\") " pod="openshift-marketplace/redhat-operators-2q7l4" Dec 07 16:17:17 crc kubenswrapper[4716]: I1207 16:17:17.983887 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twkvk\" (UniqueName: \"kubernetes.io/projected/f98dfb6b-5a13-42e4-bca1-030f964b78ce-kube-api-access-twkvk\") pod \"redhat-operators-2q7l4\" (UID: \"f98dfb6b-5a13-42e4-bca1-030f964b78ce\") " pod="openshift-marketplace/redhat-operators-2q7l4" Dec 07 16:17:18 crc kubenswrapper[4716]: I1207 16:17:18.100395 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2q7l4" Dec 07 16:17:18 crc kubenswrapper[4716]: I1207 16:17:18.174960 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-j9fcv" Dec 07 16:17:18 crc kubenswrapper[4716]: I1207 16:17:18.358176 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j9fcv" event={"ID":"1a4797ee-7747-4d9d-abc6-113b15d0e888","Type":"ContainerDied","Data":"b5b215031c6c7fbf58b6956a0ed7f4c3cef7e0a968b693138786ddec3f647bcc"} Dec 07 16:17:18 crc kubenswrapper[4716]: I1207 16:17:18.358225 4716 scope.go:117] "RemoveContainer" containerID="b31e92f451922389d848048a309c5efc59a1bd580aa2b19d89dccd5074400baf" Dec 07 16:17:18 crc kubenswrapper[4716]: I1207 16:17:18.358269 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-j9fcv" Dec 07 16:17:18 crc kubenswrapper[4716]: I1207 16:17:18.366302 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xtgjc\" (UniqueName: \"kubernetes.io/projected/1a4797ee-7747-4d9d-abc6-113b15d0e888-kube-api-access-xtgjc\") pod \"1a4797ee-7747-4d9d-abc6-113b15d0e888\" (UID: \"1a4797ee-7747-4d9d-abc6-113b15d0e888\") " Dec 07 16:17:18 crc kubenswrapper[4716]: I1207 16:17:18.366350 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a4797ee-7747-4d9d-abc6-113b15d0e888-catalog-content\") pod \"1a4797ee-7747-4d9d-abc6-113b15d0e888\" (UID: \"1a4797ee-7747-4d9d-abc6-113b15d0e888\") " Dec 07 16:17:18 crc kubenswrapper[4716]: I1207 16:17:18.366412 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a4797ee-7747-4d9d-abc6-113b15d0e888-utilities\") pod \"1a4797ee-7747-4d9d-abc6-113b15d0e888\" (UID: \"1a4797ee-7747-4d9d-abc6-113b15d0e888\") " Dec 07 16:17:18 crc kubenswrapper[4716]: I1207 16:17:18.367421 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a4797ee-7747-4d9d-abc6-113b15d0e888-utilities" (OuterVolumeSpecName: "utilities") pod "1a4797ee-7747-4d9d-abc6-113b15d0e888" (UID: "1a4797ee-7747-4d9d-abc6-113b15d0e888"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:17:18 crc kubenswrapper[4716]: I1207 16:17:18.374411 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a4797ee-7747-4d9d-abc6-113b15d0e888-kube-api-access-xtgjc" (OuterVolumeSpecName: "kube-api-access-xtgjc") pod "1a4797ee-7747-4d9d-abc6-113b15d0e888" (UID: "1a4797ee-7747-4d9d-abc6-113b15d0e888"). InnerVolumeSpecName "kube-api-access-xtgjc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:17:18 crc kubenswrapper[4716]: I1207 16:17:18.383631 4716 scope.go:117] "RemoveContainer" containerID="13a9eb0edbdd0ccf8593e2f7ef2745501398a457200b3738778283f36a1037a7" Dec 07 16:17:18 crc kubenswrapper[4716]: I1207 16:17:18.388538 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a4797ee-7747-4d9d-abc6-113b15d0e888-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1a4797ee-7747-4d9d-abc6-113b15d0e888" (UID: "1a4797ee-7747-4d9d-abc6-113b15d0e888"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:17:18 crc kubenswrapper[4716]: I1207 16:17:18.408530 4716 scope.go:117] "RemoveContainer" containerID="7ed1b4ee1ac483926ce033dc3e1fd5e3ae953f469d440cdb0b2c38d0791e2c5a" Dec 07 16:17:18 crc kubenswrapper[4716]: I1207 16:17:18.468465 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xtgjc\" (UniqueName: \"kubernetes.io/projected/1a4797ee-7747-4d9d-abc6-113b15d0e888-kube-api-access-xtgjc\") on node \"crc\" DevicePath \"\"" Dec 07 16:17:18 crc kubenswrapper[4716]: I1207 16:17:18.468509 4716 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a4797ee-7747-4d9d-abc6-113b15d0e888-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 07 16:17:18 crc kubenswrapper[4716]: I1207 16:17:18.468522 4716 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a4797ee-7747-4d9d-abc6-113b15d0e888-utilities\") on node \"crc\" DevicePath \"\"" Dec 07 16:17:18 crc kubenswrapper[4716]: I1207 16:17:18.581069 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2q7l4"] Dec 07 16:17:18 crc kubenswrapper[4716]: I1207 16:17:18.691966 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-j9fcv"] Dec 07 16:17:18 crc kubenswrapper[4716]: I1207 16:17:18.699142 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-j9fcv"] Dec 07 16:17:19 crc kubenswrapper[4716]: I1207 16:17:19.385596 4716 generic.go:334] "Generic (PLEG): container finished" podID="f98dfb6b-5a13-42e4-bca1-030f964b78ce" containerID="4f400069cb58c65bdc2c1a9f100546822c743a27bdf17cfdbba425cea44b7d4a" exitCode=0 Dec 07 16:17:19 crc kubenswrapper[4716]: I1207 16:17:19.386025 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2q7l4" event={"ID":"f98dfb6b-5a13-42e4-bca1-030f964b78ce","Type":"ContainerDied","Data":"4f400069cb58c65bdc2c1a9f100546822c743a27bdf17cfdbba425cea44b7d4a"} Dec 07 16:17:19 crc kubenswrapper[4716]: I1207 16:17:19.386055 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2q7l4" event={"ID":"f98dfb6b-5a13-42e4-bca1-030f964b78ce","Type":"ContainerStarted","Data":"8c77248ccf81982da297e70d1f01df70fddcecc3bbe829e1707c28ff865951b7"} Dec 07 16:17:19 crc kubenswrapper[4716]: I1207 16:17:19.388344 4716 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 07 16:17:19 crc kubenswrapper[4716]: I1207 16:17:19.667167 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a4797ee-7747-4d9d-abc6-113b15d0e888" path="/var/lib/kubelet/pods/1a4797ee-7747-4d9d-abc6-113b15d0e888/volumes" Dec 07 16:17:20 crc kubenswrapper[4716]: I1207 16:17:20.397844 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2q7l4" event={"ID":"f98dfb6b-5a13-42e4-bca1-030f964b78ce","Type":"ContainerStarted","Data":"cf57458fb0cb2dccc8bdf4918c05ed2518feea9353396a4b3c291293f6f3ceae"} Dec 07 16:17:21 crc kubenswrapper[4716]: I1207 16:17:21.407473 4716 generic.go:334] "Generic (PLEG): container finished" podID="f98dfb6b-5a13-42e4-bca1-030f964b78ce" containerID="cf57458fb0cb2dccc8bdf4918c05ed2518feea9353396a4b3c291293f6f3ceae" exitCode=0 Dec 07 16:17:21 crc kubenswrapper[4716]: I1207 16:17:21.407528 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2q7l4" event={"ID":"f98dfb6b-5a13-42e4-bca1-030f964b78ce","Type":"ContainerDied","Data":"cf57458fb0cb2dccc8bdf4918c05ed2518feea9353396a4b3c291293f6f3ceae"} Dec 07 16:17:22 crc kubenswrapper[4716]: I1207 16:17:22.416605 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2q7l4" event={"ID":"f98dfb6b-5a13-42e4-bca1-030f964b78ce","Type":"ContainerStarted","Data":"db98b29ab6b961e8d20058c659be7ccc34873b22ee13c03e7161937da0248b2b"} Dec 07 16:17:22 crc kubenswrapper[4716]: I1207 16:17:22.453589 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-2q7l4" podStartSLOduration=2.965335138 podStartE2EDuration="5.453555873s" podCreationTimestamp="2025-12-07 16:17:17 +0000 UTC" firstStartedPulling="2025-12-07 16:17:19.387978153 +0000 UTC m=+902.078263075" lastFinishedPulling="2025-12-07 16:17:21.876198888 +0000 UTC m=+904.566483810" observedRunningTime="2025-12-07 16:17:22.441223869 +0000 UTC m=+905.131508791" watchObservedRunningTime="2025-12-07 16:17:22.453555873 +0000 UTC m=+905.143840795" Dec 07 16:17:23 crc kubenswrapper[4716]: I1207 16:17:23.010329 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-n7br7"] Dec 07 16:17:23 crc kubenswrapper[4716]: E1207 16:17:23.015839 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a4797ee-7747-4d9d-abc6-113b15d0e888" containerName="extract-content" Dec 07 16:17:23 crc kubenswrapper[4716]: I1207 16:17:23.015941 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a4797ee-7747-4d9d-abc6-113b15d0e888" containerName="extract-content" Dec 07 16:17:23 crc kubenswrapper[4716]: E1207 16:17:23.016055 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a4797ee-7747-4d9d-abc6-113b15d0e888" containerName="registry-server" Dec 07 16:17:23 crc kubenswrapper[4716]: I1207 16:17:23.016183 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a4797ee-7747-4d9d-abc6-113b15d0e888" containerName="registry-server" Dec 07 16:17:23 crc kubenswrapper[4716]: E1207 16:17:23.019193 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a4797ee-7747-4d9d-abc6-113b15d0e888" containerName="extract-utilities" Dec 07 16:17:23 crc kubenswrapper[4716]: I1207 16:17:23.019227 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a4797ee-7747-4d9d-abc6-113b15d0e888" containerName="extract-utilities" Dec 07 16:17:23 crc kubenswrapper[4716]: I1207 16:17:23.019537 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a4797ee-7747-4d9d-abc6-113b15d0e888" containerName="registry-server" Dec 07 16:17:23 crc kubenswrapper[4716]: I1207 16:17:23.020323 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-n7br7" Dec 07 16:17:23 crc kubenswrapper[4716]: I1207 16:17:23.021880 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-khkjx" Dec 07 16:17:23 crc kubenswrapper[4716]: I1207 16:17:23.022594 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Dec 07 16:17:23 crc kubenswrapper[4716]: I1207 16:17:23.023307 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Dec 07 16:17:23 crc kubenswrapper[4716]: I1207 16:17:23.023336 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Dec 07 16:17:23 crc kubenswrapper[4716]: I1207 16:17:23.026202 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-n7br7"] Dec 07 16:17:23 crc kubenswrapper[4716]: I1207 16:17:23.039765 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81351bb9-bf61-4c75-8a31-41769f3c09bf-config\") pod \"dnsmasq-dns-675f4bcbfc-n7br7\" (UID: \"81351bb9-bf61-4c75-8a31-41769f3c09bf\") " pod="openstack/dnsmasq-dns-675f4bcbfc-n7br7" Dec 07 16:17:23 crc kubenswrapper[4716]: I1207 16:17:23.039986 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9258x\" (UniqueName: \"kubernetes.io/projected/81351bb9-bf61-4c75-8a31-41769f3c09bf-kube-api-access-9258x\") pod \"dnsmasq-dns-675f4bcbfc-n7br7\" (UID: \"81351bb9-bf61-4c75-8a31-41769f3c09bf\") " pod="openstack/dnsmasq-dns-675f4bcbfc-n7br7" Dec 07 16:17:23 crc kubenswrapper[4716]: I1207 16:17:23.051684 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-rjq6f"] Dec 07 16:17:23 crc kubenswrapper[4716]: I1207 16:17:23.052757 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-rjq6f" Dec 07 16:17:23 crc kubenswrapper[4716]: I1207 16:17:23.058317 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Dec 07 16:17:23 crc kubenswrapper[4716]: I1207 16:17:23.064434 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-rjq6f"] Dec 07 16:17:23 crc kubenswrapper[4716]: I1207 16:17:23.144927 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7e9041a0-0258-4e21-b0a0-07edb6549139-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-rjq6f\" (UID: \"7e9041a0-0258-4e21-b0a0-07edb6549139\") " pod="openstack/dnsmasq-dns-78dd6ddcc-rjq6f" Dec 07 16:17:23 crc kubenswrapper[4716]: I1207 16:17:23.145026 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e9041a0-0258-4e21-b0a0-07edb6549139-config\") pod \"dnsmasq-dns-78dd6ddcc-rjq6f\" (UID: \"7e9041a0-0258-4e21-b0a0-07edb6549139\") " pod="openstack/dnsmasq-dns-78dd6ddcc-rjq6f" Dec 07 16:17:23 crc kubenswrapper[4716]: I1207 16:17:23.145057 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81351bb9-bf61-4c75-8a31-41769f3c09bf-config\") pod \"dnsmasq-dns-675f4bcbfc-n7br7\" (UID: \"81351bb9-bf61-4c75-8a31-41769f3c09bf\") " pod="openstack/dnsmasq-dns-675f4bcbfc-n7br7" Dec 07 16:17:23 crc kubenswrapper[4716]: I1207 16:17:23.145093 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9258x\" (UniqueName: \"kubernetes.io/projected/81351bb9-bf61-4c75-8a31-41769f3c09bf-kube-api-access-9258x\") pod \"dnsmasq-dns-675f4bcbfc-n7br7\" (UID: \"81351bb9-bf61-4c75-8a31-41769f3c09bf\") " pod="openstack/dnsmasq-dns-675f4bcbfc-n7br7" Dec 07 16:17:23 crc kubenswrapper[4716]: I1207 16:17:23.145113 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h79wx\" (UniqueName: \"kubernetes.io/projected/7e9041a0-0258-4e21-b0a0-07edb6549139-kube-api-access-h79wx\") pod \"dnsmasq-dns-78dd6ddcc-rjq6f\" (UID: \"7e9041a0-0258-4e21-b0a0-07edb6549139\") " pod="openstack/dnsmasq-dns-78dd6ddcc-rjq6f" Dec 07 16:17:23 crc kubenswrapper[4716]: I1207 16:17:23.146032 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81351bb9-bf61-4c75-8a31-41769f3c09bf-config\") pod \"dnsmasq-dns-675f4bcbfc-n7br7\" (UID: \"81351bb9-bf61-4c75-8a31-41769f3c09bf\") " pod="openstack/dnsmasq-dns-675f4bcbfc-n7br7" Dec 07 16:17:23 crc kubenswrapper[4716]: I1207 16:17:23.173198 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9258x\" (UniqueName: \"kubernetes.io/projected/81351bb9-bf61-4c75-8a31-41769f3c09bf-kube-api-access-9258x\") pod \"dnsmasq-dns-675f4bcbfc-n7br7\" (UID: \"81351bb9-bf61-4c75-8a31-41769f3c09bf\") " pod="openstack/dnsmasq-dns-675f4bcbfc-n7br7" Dec 07 16:17:23 crc kubenswrapper[4716]: I1207 16:17:23.245842 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e9041a0-0258-4e21-b0a0-07edb6549139-config\") pod \"dnsmasq-dns-78dd6ddcc-rjq6f\" (UID: \"7e9041a0-0258-4e21-b0a0-07edb6549139\") " pod="openstack/dnsmasq-dns-78dd6ddcc-rjq6f" Dec 07 16:17:23 crc kubenswrapper[4716]: I1207 16:17:23.245900 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h79wx\" (UniqueName: \"kubernetes.io/projected/7e9041a0-0258-4e21-b0a0-07edb6549139-kube-api-access-h79wx\") pod \"dnsmasq-dns-78dd6ddcc-rjq6f\" (UID: \"7e9041a0-0258-4e21-b0a0-07edb6549139\") " pod="openstack/dnsmasq-dns-78dd6ddcc-rjq6f" Dec 07 16:17:23 crc kubenswrapper[4716]: I1207 16:17:23.245941 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7e9041a0-0258-4e21-b0a0-07edb6549139-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-rjq6f\" (UID: \"7e9041a0-0258-4e21-b0a0-07edb6549139\") " pod="openstack/dnsmasq-dns-78dd6ddcc-rjq6f" Dec 07 16:17:23 crc kubenswrapper[4716]: I1207 16:17:23.246724 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7e9041a0-0258-4e21-b0a0-07edb6549139-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-rjq6f\" (UID: \"7e9041a0-0258-4e21-b0a0-07edb6549139\") " pod="openstack/dnsmasq-dns-78dd6ddcc-rjq6f" Dec 07 16:17:23 crc kubenswrapper[4716]: I1207 16:17:23.246724 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e9041a0-0258-4e21-b0a0-07edb6549139-config\") pod \"dnsmasq-dns-78dd6ddcc-rjq6f\" (UID: \"7e9041a0-0258-4e21-b0a0-07edb6549139\") " pod="openstack/dnsmasq-dns-78dd6ddcc-rjq6f" Dec 07 16:17:23 crc kubenswrapper[4716]: I1207 16:17:23.265828 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h79wx\" (UniqueName: \"kubernetes.io/projected/7e9041a0-0258-4e21-b0a0-07edb6549139-kube-api-access-h79wx\") pod \"dnsmasq-dns-78dd6ddcc-rjq6f\" (UID: \"7e9041a0-0258-4e21-b0a0-07edb6549139\") " pod="openstack/dnsmasq-dns-78dd6ddcc-rjq6f" Dec 07 16:17:23 crc kubenswrapper[4716]: I1207 16:17:23.338266 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-n7br7" Dec 07 16:17:23 crc kubenswrapper[4716]: I1207 16:17:23.369355 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-rjq6f" Dec 07 16:17:23 crc kubenswrapper[4716]: I1207 16:17:23.816746 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-n7br7"] Dec 07 16:17:23 crc kubenswrapper[4716]: W1207 16:17:23.823738 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod81351bb9_bf61_4c75_8a31_41769f3c09bf.slice/crio-8e65c04758f9dcfd7a9d71a96b034416c28e4aea2a6d20b2c01413c17912c587 WatchSource:0}: Error finding container 8e65c04758f9dcfd7a9d71a96b034416c28e4aea2a6d20b2c01413c17912c587: Status 404 returned error can't find the container with id 8e65c04758f9dcfd7a9d71a96b034416c28e4aea2a6d20b2c01413c17912c587 Dec 07 16:17:23 crc kubenswrapper[4716]: I1207 16:17:23.875510 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-rjq6f"] Dec 07 16:17:23 crc kubenswrapper[4716]: W1207 16:17:23.881164 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7e9041a0_0258_4e21_b0a0_07edb6549139.slice/crio-6cf210dcb88d6dfbf796d41105cc4c88595ad110329ba0146984755d1ee60b31 WatchSource:0}: Error finding container 6cf210dcb88d6dfbf796d41105cc4c88595ad110329ba0146984755d1ee60b31: Status 404 returned error can't find the container with id 6cf210dcb88d6dfbf796d41105cc4c88595ad110329ba0146984755d1ee60b31 Dec 07 16:17:24 crc kubenswrapper[4716]: I1207 16:17:24.441205 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-rjq6f" event={"ID":"7e9041a0-0258-4e21-b0a0-07edb6549139","Type":"ContainerStarted","Data":"6cf210dcb88d6dfbf796d41105cc4c88595ad110329ba0146984755d1ee60b31"} Dec 07 16:17:24 crc kubenswrapper[4716]: I1207 16:17:24.442684 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-n7br7" event={"ID":"81351bb9-bf61-4c75-8a31-41769f3c09bf","Type":"ContainerStarted","Data":"8e65c04758f9dcfd7a9d71a96b034416c28e4aea2a6d20b2c01413c17912c587"} Dec 07 16:17:25 crc kubenswrapper[4716]: I1207 16:17:25.896793 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-n7br7"] Dec 07 16:17:25 crc kubenswrapper[4716]: I1207 16:17:25.916110 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-dwp2b"] Dec 07 16:17:25 crc kubenswrapper[4716]: I1207 16:17:25.918652 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-dwp2b" Dec 07 16:17:25 crc kubenswrapper[4716]: I1207 16:17:25.922466 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-dwp2b"] Dec 07 16:17:25 crc kubenswrapper[4716]: I1207 16:17:25.993936 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b4f34fda-3133-46e3-99d6-d49f0ae032d5-dns-svc\") pod \"dnsmasq-dns-666b6646f7-dwp2b\" (UID: \"b4f34fda-3133-46e3-99d6-d49f0ae032d5\") " pod="openstack/dnsmasq-dns-666b6646f7-dwp2b" Dec 07 16:17:25 crc kubenswrapper[4716]: I1207 16:17:25.994023 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4f34fda-3133-46e3-99d6-d49f0ae032d5-config\") pod \"dnsmasq-dns-666b6646f7-dwp2b\" (UID: \"b4f34fda-3133-46e3-99d6-d49f0ae032d5\") " pod="openstack/dnsmasq-dns-666b6646f7-dwp2b" Dec 07 16:17:25 crc kubenswrapper[4716]: I1207 16:17:25.994053 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-plj7t\" (UniqueName: \"kubernetes.io/projected/b4f34fda-3133-46e3-99d6-d49f0ae032d5-kube-api-access-plj7t\") pod \"dnsmasq-dns-666b6646f7-dwp2b\" (UID: \"b4f34fda-3133-46e3-99d6-d49f0ae032d5\") " pod="openstack/dnsmasq-dns-666b6646f7-dwp2b" Dec 07 16:17:26 crc kubenswrapper[4716]: I1207 16:17:26.094907 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b4f34fda-3133-46e3-99d6-d49f0ae032d5-dns-svc\") pod \"dnsmasq-dns-666b6646f7-dwp2b\" (UID: \"b4f34fda-3133-46e3-99d6-d49f0ae032d5\") " pod="openstack/dnsmasq-dns-666b6646f7-dwp2b" Dec 07 16:17:26 crc kubenswrapper[4716]: I1207 16:17:26.094966 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4f34fda-3133-46e3-99d6-d49f0ae032d5-config\") pod \"dnsmasq-dns-666b6646f7-dwp2b\" (UID: \"b4f34fda-3133-46e3-99d6-d49f0ae032d5\") " pod="openstack/dnsmasq-dns-666b6646f7-dwp2b" Dec 07 16:17:26 crc kubenswrapper[4716]: I1207 16:17:26.094987 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-plj7t\" (UniqueName: \"kubernetes.io/projected/b4f34fda-3133-46e3-99d6-d49f0ae032d5-kube-api-access-plj7t\") pod \"dnsmasq-dns-666b6646f7-dwp2b\" (UID: \"b4f34fda-3133-46e3-99d6-d49f0ae032d5\") " pod="openstack/dnsmasq-dns-666b6646f7-dwp2b" Dec 07 16:17:26 crc kubenswrapper[4716]: I1207 16:17:26.095795 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4f34fda-3133-46e3-99d6-d49f0ae032d5-config\") pod \"dnsmasq-dns-666b6646f7-dwp2b\" (UID: \"b4f34fda-3133-46e3-99d6-d49f0ae032d5\") " pod="openstack/dnsmasq-dns-666b6646f7-dwp2b" Dec 07 16:17:26 crc kubenswrapper[4716]: I1207 16:17:26.095795 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b4f34fda-3133-46e3-99d6-d49f0ae032d5-dns-svc\") pod \"dnsmasq-dns-666b6646f7-dwp2b\" (UID: \"b4f34fda-3133-46e3-99d6-d49f0ae032d5\") " pod="openstack/dnsmasq-dns-666b6646f7-dwp2b" Dec 07 16:17:26 crc kubenswrapper[4716]: I1207 16:17:26.136360 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-plj7t\" (UniqueName: \"kubernetes.io/projected/b4f34fda-3133-46e3-99d6-d49f0ae032d5-kube-api-access-plj7t\") pod \"dnsmasq-dns-666b6646f7-dwp2b\" (UID: \"b4f34fda-3133-46e3-99d6-d49f0ae032d5\") " pod="openstack/dnsmasq-dns-666b6646f7-dwp2b" Dec 07 16:17:26 crc kubenswrapper[4716]: I1207 16:17:26.179686 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-rjq6f"] Dec 07 16:17:26 crc kubenswrapper[4716]: I1207 16:17:26.207714 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-ktd8n"] Dec 07 16:17:26 crc kubenswrapper[4716]: I1207 16:17:26.208856 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-ktd8n" Dec 07 16:17:26 crc kubenswrapper[4716]: I1207 16:17:26.219479 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-ktd8n"] Dec 07 16:17:26 crc kubenswrapper[4716]: I1207 16:17:26.245957 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-dwp2b" Dec 07 16:17:26 crc kubenswrapper[4716]: I1207 16:17:26.423747 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cx6ns\" (UniqueName: \"kubernetes.io/projected/240ebe8a-eb59-4726-a66f-4ce1e2900e4c-kube-api-access-cx6ns\") pod \"dnsmasq-dns-57d769cc4f-ktd8n\" (UID: \"240ebe8a-eb59-4726-a66f-4ce1e2900e4c\") " pod="openstack/dnsmasq-dns-57d769cc4f-ktd8n" Dec 07 16:17:26 crc kubenswrapper[4716]: I1207 16:17:26.426484 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/240ebe8a-eb59-4726-a66f-4ce1e2900e4c-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-ktd8n\" (UID: \"240ebe8a-eb59-4726-a66f-4ce1e2900e4c\") " pod="openstack/dnsmasq-dns-57d769cc4f-ktd8n" Dec 07 16:17:26 crc kubenswrapper[4716]: I1207 16:17:26.426723 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/240ebe8a-eb59-4726-a66f-4ce1e2900e4c-config\") pod \"dnsmasq-dns-57d769cc4f-ktd8n\" (UID: \"240ebe8a-eb59-4726-a66f-4ce1e2900e4c\") " pod="openstack/dnsmasq-dns-57d769cc4f-ktd8n" Dec 07 16:17:26 crc kubenswrapper[4716]: I1207 16:17:26.528123 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/240ebe8a-eb59-4726-a66f-4ce1e2900e4c-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-ktd8n\" (UID: \"240ebe8a-eb59-4726-a66f-4ce1e2900e4c\") " pod="openstack/dnsmasq-dns-57d769cc4f-ktd8n" Dec 07 16:17:26 crc kubenswrapper[4716]: I1207 16:17:26.528209 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/240ebe8a-eb59-4726-a66f-4ce1e2900e4c-config\") pod \"dnsmasq-dns-57d769cc4f-ktd8n\" (UID: \"240ebe8a-eb59-4726-a66f-4ce1e2900e4c\") " pod="openstack/dnsmasq-dns-57d769cc4f-ktd8n" Dec 07 16:17:26 crc kubenswrapper[4716]: I1207 16:17:26.528260 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cx6ns\" (UniqueName: \"kubernetes.io/projected/240ebe8a-eb59-4726-a66f-4ce1e2900e4c-kube-api-access-cx6ns\") pod \"dnsmasq-dns-57d769cc4f-ktd8n\" (UID: \"240ebe8a-eb59-4726-a66f-4ce1e2900e4c\") " pod="openstack/dnsmasq-dns-57d769cc4f-ktd8n" Dec 07 16:17:26 crc kubenswrapper[4716]: I1207 16:17:26.529215 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/240ebe8a-eb59-4726-a66f-4ce1e2900e4c-config\") pod \"dnsmasq-dns-57d769cc4f-ktd8n\" (UID: \"240ebe8a-eb59-4726-a66f-4ce1e2900e4c\") " pod="openstack/dnsmasq-dns-57d769cc4f-ktd8n" Dec 07 16:17:26 crc kubenswrapper[4716]: I1207 16:17:26.529355 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/240ebe8a-eb59-4726-a66f-4ce1e2900e4c-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-ktd8n\" (UID: \"240ebe8a-eb59-4726-a66f-4ce1e2900e4c\") " pod="openstack/dnsmasq-dns-57d769cc4f-ktd8n" Dec 07 16:17:26 crc kubenswrapper[4716]: I1207 16:17:26.548364 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cx6ns\" (UniqueName: \"kubernetes.io/projected/240ebe8a-eb59-4726-a66f-4ce1e2900e4c-kube-api-access-cx6ns\") pod \"dnsmasq-dns-57d769cc4f-ktd8n\" (UID: \"240ebe8a-eb59-4726-a66f-4ce1e2900e4c\") " pod="openstack/dnsmasq-dns-57d769cc4f-ktd8n" Dec 07 16:17:26 crc kubenswrapper[4716]: I1207 16:17:26.809932 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-dwp2b"] Dec 07 16:17:26 crc kubenswrapper[4716]: W1207 16:17:26.829228 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb4f34fda_3133_46e3_99d6_d49f0ae032d5.slice/crio-13f05e456fc8da842a3bc47f553afcaa206a2a51512c74c494302cee89f3f6b4 WatchSource:0}: Error finding container 13f05e456fc8da842a3bc47f553afcaa206a2a51512c74c494302cee89f3f6b4: Status 404 returned error can't find the container with id 13f05e456fc8da842a3bc47f553afcaa206a2a51512c74c494302cee89f3f6b4 Dec 07 16:17:26 crc kubenswrapper[4716]: I1207 16:17:26.832674 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-ktd8n" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.057470 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.059009 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.060024 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.061590 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.061975 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.062118 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.062168 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.062392 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-5tlk9" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.062496 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.063703 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.100323 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-ktd8n"] Dec 07 16:17:27 crc kubenswrapper[4716]: W1207 16:17:27.115221 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod240ebe8a_eb59_4726_a66f_4ce1e2900e4c.slice/crio-e9b4fd94e94a98f504eeef3860b2d6161de289fdcc23e3137c8f71f653b02250 WatchSource:0}: Error finding container e9b4fd94e94a98f504eeef3860b2d6161de289fdcc23e3137c8f71f653b02250: Status 404 returned error can't find the container with id e9b4fd94e94a98f504eeef3860b2d6161de289fdcc23e3137c8f71f653b02250 Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.137388 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f653df87-c234-4cb6-8ee8-6b16b4fe0044-server-conf\") pod \"rabbitmq-server-0\" (UID: \"f653df87-c234-4cb6-8ee8-6b16b4fe0044\") " pod="openstack/rabbitmq-server-0" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.137442 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f653df87-c234-4cb6-8ee8-6b16b4fe0044-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"f653df87-c234-4cb6-8ee8-6b16b4fe0044\") " pod="openstack/rabbitmq-server-0" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.137463 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"f653df87-c234-4cb6-8ee8-6b16b4fe0044\") " pod="openstack/rabbitmq-server-0" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.137487 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f653df87-c234-4cb6-8ee8-6b16b4fe0044-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"f653df87-c234-4cb6-8ee8-6b16b4fe0044\") " pod="openstack/rabbitmq-server-0" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.137536 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f653df87-c234-4cb6-8ee8-6b16b4fe0044-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"f653df87-c234-4cb6-8ee8-6b16b4fe0044\") " pod="openstack/rabbitmq-server-0" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.137561 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v8bpb\" (UniqueName: \"kubernetes.io/projected/f653df87-c234-4cb6-8ee8-6b16b4fe0044-kube-api-access-v8bpb\") pod \"rabbitmq-server-0\" (UID: \"f653df87-c234-4cb6-8ee8-6b16b4fe0044\") " pod="openstack/rabbitmq-server-0" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.137580 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f653df87-c234-4cb6-8ee8-6b16b4fe0044-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"f653df87-c234-4cb6-8ee8-6b16b4fe0044\") " pod="openstack/rabbitmq-server-0" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.137608 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f653df87-c234-4cb6-8ee8-6b16b4fe0044-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"f653df87-c234-4cb6-8ee8-6b16b4fe0044\") " pod="openstack/rabbitmq-server-0" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.137622 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f653df87-c234-4cb6-8ee8-6b16b4fe0044-pod-info\") pod \"rabbitmq-server-0\" (UID: \"f653df87-c234-4cb6-8ee8-6b16b4fe0044\") " pod="openstack/rabbitmq-server-0" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.137637 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f653df87-c234-4cb6-8ee8-6b16b4fe0044-config-data\") pod \"rabbitmq-server-0\" (UID: \"f653df87-c234-4cb6-8ee8-6b16b4fe0044\") " pod="openstack/rabbitmq-server-0" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.137693 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f653df87-c234-4cb6-8ee8-6b16b4fe0044-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"f653df87-c234-4cb6-8ee8-6b16b4fe0044\") " pod="openstack/rabbitmq-server-0" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.239243 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f653df87-c234-4cb6-8ee8-6b16b4fe0044-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"f653df87-c234-4cb6-8ee8-6b16b4fe0044\") " pod="openstack/rabbitmq-server-0" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.239305 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f653df87-c234-4cb6-8ee8-6b16b4fe0044-server-conf\") pod \"rabbitmq-server-0\" (UID: \"f653df87-c234-4cb6-8ee8-6b16b4fe0044\") " pod="openstack/rabbitmq-server-0" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.239338 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f653df87-c234-4cb6-8ee8-6b16b4fe0044-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"f653df87-c234-4cb6-8ee8-6b16b4fe0044\") " pod="openstack/rabbitmq-server-0" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.239360 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"f653df87-c234-4cb6-8ee8-6b16b4fe0044\") " pod="openstack/rabbitmq-server-0" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.239380 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f653df87-c234-4cb6-8ee8-6b16b4fe0044-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"f653df87-c234-4cb6-8ee8-6b16b4fe0044\") " pod="openstack/rabbitmq-server-0" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.239433 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f653df87-c234-4cb6-8ee8-6b16b4fe0044-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"f653df87-c234-4cb6-8ee8-6b16b4fe0044\") " pod="openstack/rabbitmq-server-0" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.239463 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v8bpb\" (UniqueName: \"kubernetes.io/projected/f653df87-c234-4cb6-8ee8-6b16b4fe0044-kube-api-access-v8bpb\") pod \"rabbitmq-server-0\" (UID: \"f653df87-c234-4cb6-8ee8-6b16b4fe0044\") " pod="openstack/rabbitmq-server-0" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.239487 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f653df87-c234-4cb6-8ee8-6b16b4fe0044-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"f653df87-c234-4cb6-8ee8-6b16b4fe0044\") " pod="openstack/rabbitmq-server-0" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.239525 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f653df87-c234-4cb6-8ee8-6b16b4fe0044-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"f653df87-c234-4cb6-8ee8-6b16b4fe0044\") " pod="openstack/rabbitmq-server-0" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.239548 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f653df87-c234-4cb6-8ee8-6b16b4fe0044-pod-info\") pod \"rabbitmq-server-0\" (UID: \"f653df87-c234-4cb6-8ee8-6b16b4fe0044\") " pod="openstack/rabbitmq-server-0" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.239569 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f653df87-c234-4cb6-8ee8-6b16b4fe0044-config-data\") pod \"rabbitmq-server-0\" (UID: \"f653df87-c234-4cb6-8ee8-6b16b4fe0044\") " pod="openstack/rabbitmq-server-0" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.241663 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f653df87-c234-4cb6-8ee8-6b16b4fe0044-config-data\") pod \"rabbitmq-server-0\" (UID: \"f653df87-c234-4cb6-8ee8-6b16b4fe0044\") " pod="openstack/rabbitmq-server-0" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.242278 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f653df87-c234-4cb6-8ee8-6b16b4fe0044-server-conf\") pod \"rabbitmq-server-0\" (UID: \"f653df87-c234-4cb6-8ee8-6b16b4fe0044\") " pod="openstack/rabbitmq-server-0" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.242553 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f653df87-c234-4cb6-8ee8-6b16b4fe0044-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"f653df87-c234-4cb6-8ee8-6b16b4fe0044\") " pod="openstack/rabbitmq-server-0" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.242849 4716 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"f653df87-c234-4cb6-8ee8-6b16b4fe0044\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/rabbitmq-server-0" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.243025 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f653df87-c234-4cb6-8ee8-6b16b4fe0044-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"f653df87-c234-4cb6-8ee8-6b16b4fe0044\") " pod="openstack/rabbitmq-server-0" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.243719 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f653df87-c234-4cb6-8ee8-6b16b4fe0044-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"f653df87-c234-4cb6-8ee8-6b16b4fe0044\") " pod="openstack/rabbitmq-server-0" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.246295 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f653df87-c234-4cb6-8ee8-6b16b4fe0044-pod-info\") pod \"rabbitmq-server-0\" (UID: \"f653df87-c234-4cb6-8ee8-6b16b4fe0044\") " pod="openstack/rabbitmq-server-0" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.246717 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f653df87-c234-4cb6-8ee8-6b16b4fe0044-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"f653df87-c234-4cb6-8ee8-6b16b4fe0044\") " pod="openstack/rabbitmq-server-0" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.247002 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f653df87-c234-4cb6-8ee8-6b16b4fe0044-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"f653df87-c234-4cb6-8ee8-6b16b4fe0044\") " pod="openstack/rabbitmq-server-0" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.250132 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f653df87-c234-4cb6-8ee8-6b16b4fe0044-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"f653df87-c234-4cb6-8ee8-6b16b4fe0044\") " pod="openstack/rabbitmq-server-0" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.258239 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v8bpb\" (UniqueName: \"kubernetes.io/projected/f653df87-c234-4cb6-8ee8-6b16b4fe0044-kube-api-access-v8bpb\") pod \"rabbitmq-server-0\" (UID: \"f653df87-c234-4cb6-8ee8-6b16b4fe0044\") " pod="openstack/rabbitmq-server-0" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.271572 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"f653df87-c234-4cb6-8ee8-6b16b4fe0044\") " pod="openstack/rabbitmq-server-0" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.349733 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.352419 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.354434 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.355527 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-b6fbd" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.356257 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.356457 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.356546 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.356585 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.356698 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.360572 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.386393 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.443460 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.443508 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.443543 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.443583 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.443615 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.443639 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.443658 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.443734 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.444923 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.444988 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.445017 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6zsm9\" (UniqueName: \"kubernetes.io/projected/9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe-kube-api-access-6zsm9\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.469706 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-dwp2b" event={"ID":"b4f34fda-3133-46e3-99d6-d49f0ae032d5","Type":"ContainerStarted","Data":"13f05e456fc8da842a3bc47f553afcaa206a2a51512c74c494302cee89f3f6b4"} Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.471853 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-ktd8n" event={"ID":"240ebe8a-eb59-4726-a66f-4ce1e2900e4c","Type":"ContainerStarted","Data":"e9b4fd94e94a98f504eeef3860b2d6161de289fdcc23e3137c8f71f653b02250"} Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.545965 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.546015 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.546051 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.546110 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.546147 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.546176 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.546194 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.546234 4716 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.546256 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.546289 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.546317 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6zsm9\" (UniqueName: \"kubernetes.io/projected/9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe-kube-api-access-6zsm9\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.546340 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.547255 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.549036 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.549506 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.550548 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.552728 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.552867 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.556126 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.560375 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.565791 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.569552 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6zsm9\" (UniqueName: \"kubernetes.io/projected/9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe-kube-api-access-6zsm9\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.575403 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.676126 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:17:27 crc kubenswrapper[4716]: I1207 16:17:27.872450 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 07 16:17:28 crc kubenswrapper[4716]: I1207 16:17:28.101402 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-2q7l4" Dec 07 16:17:28 crc kubenswrapper[4716]: I1207 16:17:28.101521 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-2q7l4" Dec 07 16:17:28 crc kubenswrapper[4716]: I1207 16:17:28.179794 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-2q7l4" Dec 07 16:17:28 crc kubenswrapper[4716]: I1207 16:17:28.211329 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 07 16:17:28 crc kubenswrapper[4716]: I1207 16:17:28.479303 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe","Type":"ContainerStarted","Data":"0c115d03172b1cc97cbbfbf14c42f1452617bd97a6f36bca2ea6f8acc3f24cf5"} Dec 07 16:17:28 crc kubenswrapper[4716]: I1207 16:17:28.481554 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f653df87-c234-4cb6-8ee8-6b16b4fe0044","Type":"ContainerStarted","Data":"f7fef3c71a02a05d212998dcea6ed0986974ed48e929dc2cca1cd5fd191035db"} Dec 07 16:17:28 crc kubenswrapper[4716]: I1207 16:17:28.526860 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-2q7l4" Dec 07 16:17:28 crc kubenswrapper[4716]: I1207 16:17:28.566945 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2q7l4"] Dec 07 16:17:28 crc kubenswrapper[4716]: I1207 16:17:28.785028 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Dec 07 16:17:28 crc kubenswrapper[4716]: I1207 16:17:28.786299 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 07 16:17:28 crc kubenswrapper[4716]: I1207 16:17:28.789254 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Dec 07 16:17:28 crc kubenswrapper[4716]: I1207 16:17:28.789383 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-fxl7n" Dec 07 16:17:28 crc kubenswrapper[4716]: I1207 16:17:28.790178 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Dec 07 16:17:28 crc kubenswrapper[4716]: I1207 16:17:28.790335 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Dec 07 16:17:28 crc kubenswrapper[4716]: I1207 16:17:28.790475 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 07 16:17:28 crc kubenswrapper[4716]: I1207 16:17:28.805809 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Dec 07 16:17:28 crc kubenswrapper[4716]: I1207 16:17:28.878788 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/c450ef40-3a23-4bab-9d64-4475f12d0490-config-data-default\") pod \"openstack-galera-0\" (UID: \"c450ef40-3a23-4bab-9d64-4475f12d0490\") " pod="openstack/openstack-galera-0" Dec 07 16:17:28 crc kubenswrapper[4716]: I1207 16:17:28.879033 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/c450ef40-3a23-4bab-9d64-4475f12d0490-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"c450ef40-3a23-4bab-9d64-4475f12d0490\") " pod="openstack/openstack-galera-0" Dec 07 16:17:28 crc kubenswrapper[4716]: I1207 16:17:28.879159 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c450ef40-3a23-4bab-9d64-4475f12d0490-kolla-config\") pod \"openstack-galera-0\" (UID: \"c450ef40-3a23-4bab-9d64-4475f12d0490\") " pod="openstack/openstack-galera-0" Dec 07 16:17:28 crc kubenswrapper[4716]: I1207 16:17:28.879274 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/c450ef40-3a23-4bab-9d64-4475f12d0490-config-data-generated\") pod \"openstack-galera-0\" (UID: \"c450ef40-3a23-4bab-9d64-4475f12d0490\") " pod="openstack/openstack-galera-0" Dec 07 16:17:28 crc kubenswrapper[4716]: I1207 16:17:28.879354 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c450ef40-3a23-4bab-9d64-4475f12d0490-operator-scripts\") pod \"openstack-galera-0\" (UID: \"c450ef40-3a23-4bab-9d64-4475f12d0490\") " pod="openstack/openstack-galera-0" Dec 07 16:17:28 crc kubenswrapper[4716]: I1207 16:17:28.879442 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c450ef40-3a23-4bab-9d64-4475f12d0490-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"c450ef40-3a23-4bab-9d64-4475f12d0490\") " pod="openstack/openstack-galera-0" Dec 07 16:17:28 crc kubenswrapper[4716]: I1207 16:17:28.879533 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-galera-0\" (UID: \"c450ef40-3a23-4bab-9d64-4475f12d0490\") " pod="openstack/openstack-galera-0" Dec 07 16:17:28 crc kubenswrapper[4716]: I1207 16:17:28.879620 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mbwx4\" (UniqueName: \"kubernetes.io/projected/c450ef40-3a23-4bab-9d64-4475f12d0490-kube-api-access-mbwx4\") pod \"openstack-galera-0\" (UID: \"c450ef40-3a23-4bab-9d64-4475f12d0490\") " pod="openstack/openstack-galera-0" Dec 07 16:17:28 crc kubenswrapper[4716]: I1207 16:17:28.981061 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/c450ef40-3a23-4bab-9d64-4475f12d0490-config-data-generated\") pod \"openstack-galera-0\" (UID: \"c450ef40-3a23-4bab-9d64-4475f12d0490\") " pod="openstack/openstack-galera-0" Dec 07 16:17:28 crc kubenswrapper[4716]: I1207 16:17:28.981115 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c450ef40-3a23-4bab-9d64-4475f12d0490-operator-scripts\") pod \"openstack-galera-0\" (UID: \"c450ef40-3a23-4bab-9d64-4475f12d0490\") " pod="openstack/openstack-galera-0" Dec 07 16:17:28 crc kubenswrapper[4716]: I1207 16:17:28.981133 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c450ef40-3a23-4bab-9d64-4475f12d0490-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"c450ef40-3a23-4bab-9d64-4475f12d0490\") " pod="openstack/openstack-galera-0" Dec 07 16:17:28 crc kubenswrapper[4716]: I1207 16:17:28.981176 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-galera-0\" (UID: \"c450ef40-3a23-4bab-9d64-4475f12d0490\") " pod="openstack/openstack-galera-0" Dec 07 16:17:28 crc kubenswrapper[4716]: I1207 16:17:28.981209 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mbwx4\" (UniqueName: \"kubernetes.io/projected/c450ef40-3a23-4bab-9d64-4475f12d0490-kube-api-access-mbwx4\") pod \"openstack-galera-0\" (UID: \"c450ef40-3a23-4bab-9d64-4475f12d0490\") " pod="openstack/openstack-galera-0" Dec 07 16:17:28 crc kubenswrapper[4716]: I1207 16:17:28.981255 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/c450ef40-3a23-4bab-9d64-4475f12d0490-config-data-default\") pod \"openstack-galera-0\" (UID: \"c450ef40-3a23-4bab-9d64-4475f12d0490\") " pod="openstack/openstack-galera-0" Dec 07 16:17:28 crc kubenswrapper[4716]: I1207 16:17:28.981284 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/c450ef40-3a23-4bab-9d64-4475f12d0490-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"c450ef40-3a23-4bab-9d64-4475f12d0490\") " pod="openstack/openstack-galera-0" Dec 07 16:17:28 crc kubenswrapper[4716]: I1207 16:17:28.981309 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c450ef40-3a23-4bab-9d64-4475f12d0490-kolla-config\") pod \"openstack-galera-0\" (UID: \"c450ef40-3a23-4bab-9d64-4475f12d0490\") " pod="openstack/openstack-galera-0" Dec 07 16:17:28 crc kubenswrapper[4716]: I1207 16:17:28.981932 4716 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-galera-0\" (UID: \"c450ef40-3a23-4bab-9d64-4475f12d0490\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/openstack-galera-0" Dec 07 16:17:28 crc kubenswrapper[4716]: I1207 16:17:28.982092 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c450ef40-3a23-4bab-9d64-4475f12d0490-kolla-config\") pod \"openstack-galera-0\" (UID: \"c450ef40-3a23-4bab-9d64-4475f12d0490\") " pod="openstack/openstack-galera-0" Dec 07 16:17:28 crc kubenswrapper[4716]: I1207 16:17:28.982427 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/c450ef40-3a23-4bab-9d64-4475f12d0490-config-data-default\") pod \"openstack-galera-0\" (UID: \"c450ef40-3a23-4bab-9d64-4475f12d0490\") " pod="openstack/openstack-galera-0" Dec 07 16:17:28 crc kubenswrapper[4716]: I1207 16:17:28.982841 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/c450ef40-3a23-4bab-9d64-4475f12d0490-config-data-generated\") pod \"openstack-galera-0\" (UID: \"c450ef40-3a23-4bab-9d64-4475f12d0490\") " pod="openstack/openstack-galera-0" Dec 07 16:17:28 crc kubenswrapper[4716]: I1207 16:17:28.987610 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c450ef40-3a23-4bab-9d64-4475f12d0490-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"c450ef40-3a23-4bab-9d64-4475f12d0490\") " pod="openstack/openstack-galera-0" Dec 07 16:17:28 crc kubenswrapper[4716]: I1207 16:17:28.994577 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/c450ef40-3a23-4bab-9d64-4475f12d0490-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"c450ef40-3a23-4bab-9d64-4475f12d0490\") " pod="openstack/openstack-galera-0" Dec 07 16:17:29 crc kubenswrapper[4716]: I1207 16:17:29.005626 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c450ef40-3a23-4bab-9d64-4475f12d0490-operator-scripts\") pod \"openstack-galera-0\" (UID: \"c450ef40-3a23-4bab-9d64-4475f12d0490\") " pod="openstack/openstack-galera-0" Dec 07 16:17:29 crc kubenswrapper[4716]: I1207 16:17:29.008517 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mbwx4\" (UniqueName: \"kubernetes.io/projected/c450ef40-3a23-4bab-9d64-4475f12d0490-kube-api-access-mbwx4\") pod \"openstack-galera-0\" (UID: \"c450ef40-3a23-4bab-9d64-4475f12d0490\") " pod="openstack/openstack-galera-0" Dec 07 16:17:29 crc kubenswrapper[4716]: I1207 16:17:29.013759 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-galera-0\" (UID: \"c450ef40-3a23-4bab-9d64-4475f12d0490\") " pod="openstack/openstack-galera-0" Dec 07 16:17:29 crc kubenswrapper[4716]: I1207 16:17:29.117767 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 07 16:17:30 crc kubenswrapper[4716]: I1207 16:17:30.073869 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 07 16:17:30 crc kubenswrapper[4716]: I1207 16:17:30.077570 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 07 16:17:30 crc kubenswrapper[4716]: I1207 16:17:30.086345 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Dec 07 16:17:30 crc kubenswrapper[4716]: I1207 16:17:30.086520 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-vnmww" Dec 07 16:17:30 crc kubenswrapper[4716]: I1207 16:17:30.086620 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Dec 07 16:17:30 crc kubenswrapper[4716]: I1207 16:17:30.086745 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Dec 07 16:17:30 crc kubenswrapper[4716]: I1207 16:17:30.108620 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 07 16:17:30 crc kubenswrapper[4716]: I1207 16:17:30.120065 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 07 16:17:30 crc kubenswrapper[4716]: I1207 16:17:30.207850 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f5bcaa65-381c-40e7-97bd-4fc0eafc572a-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"f5bcaa65-381c-40e7-97bd-4fc0eafc572a\") " pod="openstack/openstack-cell1-galera-0" Dec 07 16:17:30 crc kubenswrapper[4716]: I1207 16:17:30.207917 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5bcaa65-381c-40e7-97bd-4fc0eafc572a-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"f5bcaa65-381c-40e7-97bd-4fc0eafc572a\") " pod="openstack/openstack-cell1-galera-0" Dec 07 16:17:30 crc kubenswrapper[4716]: I1207 16:17:30.207941 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6cj9d\" (UniqueName: \"kubernetes.io/projected/f5bcaa65-381c-40e7-97bd-4fc0eafc572a-kube-api-access-6cj9d\") pod \"openstack-cell1-galera-0\" (UID: \"f5bcaa65-381c-40e7-97bd-4fc0eafc572a\") " pod="openstack/openstack-cell1-galera-0" Dec 07 16:17:30 crc kubenswrapper[4716]: I1207 16:17:30.208028 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/f5bcaa65-381c-40e7-97bd-4fc0eafc572a-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"f5bcaa65-381c-40e7-97bd-4fc0eafc572a\") " pod="openstack/openstack-cell1-galera-0" Dec 07 16:17:30 crc kubenswrapper[4716]: I1207 16:17:30.208089 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f5bcaa65-381c-40e7-97bd-4fc0eafc572a-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"f5bcaa65-381c-40e7-97bd-4fc0eafc572a\") " pod="openstack/openstack-cell1-galera-0" Dec 07 16:17:30 crc kubenswrapper[4716]: I1207 16:17:30.208151 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-cell1-galera-0\" (UID: \"f5bcaa65-381c-40e7-97bd-4fc0eafc572a\") " pod="openstack/openstack-cell1-galera-0" Dec 07 16:17:30 crc kubenswrapper[4716]: I1207 16:17:30.208196 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f5bcaa65-381c-40e7-97bd-4fc0eafc572a-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"f5bcaa65-381c-40e7-97bd-4fc0eafc572a\") " pod="openstack/openstack-cell1-galera-0" Dec 07 16:17:30 crc kubenswrapper[4716]: I1207 16:17:30.208244 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f5bcaa65-381c-40e7-97bd-4fc0eafc572a-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"f5bcaa65-381c-40e7-97bd-4fc0eafc572a\") " pod="openstack/openstack-cell1-galera-0" Dec 07 16:17:30 crc kubenswrapper[4716]: I1207 16:17:30.309793 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-cell1-galera-0\" (UID: \"f5bcaa65-381c-40e7-97bd-4fc0eafc572a\") " pod="openstack/openstack-cell1-galera-0" Dec 07 16:17:30 crc kubenswrapper[4716]: I1207 16:17:30.309850 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f5bcaa65-381c-40e7-97bd-4fc0eafc572a-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"f5bcaa65-381c-40e7-97bd-4fc0eafc572a\") " pod="openstack/openstack-cell1-galera-0" Dec 07 16:17:30 crc kubenswrapper[4716]: I1207 16:17:30.309897 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f5bcaa65-381c-40e7-97bd-4fc0eafc572a-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"f5bcaa65-381c-40e7-97bd-4fc0eafc572a\") " pod="openstack/openstack-cell1-galera-0" Dec 07 16:17:30 crc kubenswrapper[4716]: I1207 16:17:30.309918 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f5bcaa65-381c-40e7-97bd-4fc0eafc572a-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"f5bcaa65-381c-40e7-97bd-4fc0eafc572a\") " pod="openstack/openstack-cell1-galera-0" Dec 07 16:17:30 crc kubenswrapper[4716]: I1207 16:17:30.309934 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5bcaa65-381c-40e7-97bd-4fc0eafc572a-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"f5bcaa65-381c-40e7-97bd-4fc0eafc572a\") " pod="openstack/openstack-cell1-galera-0" Dec 07 16:17:30 crc kubenswrapper[4716]: I1207 16:17:30.309949 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6cj9d\" (UniqueName: \"kubernetes.io/projected/f5bcaa65-381c-40e7-97bd-4fc0eafc572a-kube-api-access-6cj9d\") pod \"openstack-cell1-galera-0\" (UID: \"f5bcaa65-381c-40e7-97bd-4fc0eafc572a\") " pod="openstack/openstack-cell1-galera-0" Dec 07 16:17:30 crc kubenswrapper[4716]: I1207 16:17:30.310006 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/f5bcaa65-381c-40e7-97bd-4fc0eafc572a-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"f5bcaa65-381c-40e7-97bd-4fc0eafc572a\") " pod="openstack/openstack-cell1-galera-0" Dec 07 16:17:30 crc kubenswrapper[4716]: I1207 16:17:30.310034 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f5bcaa65-381c-40e7-97bd-4fc0eafc572a-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"f5bcaa65-381c-40e7-97bd-4fc0eafc572a\") " pod="openstack/openstack-cell1-galera-0" Dec 07 16:17:30 crc kubenswrapper[4716]: I1207 16:17:30.310181 4716 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-cell1-galera-0\" (UID: \"f5bcaa65-381c-40e7-97bd-4fc0eafc572a\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/openstack-cell1-galera-0" Dec 07 16:17:30 crc kubenswrapper[4716]: I1207 16:17:30.310430 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f5bcaa65-381c-40e7-97bd-4fc0eafc572a-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"f5bcaa65-381c-40e7-97bd-4fc0eafc572a\") " pod="openstack/openstack-cell1-galera-0" Dec 07 16:17:30 crc kubenswrapper[4716]: I1207 16:17:30.311717 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f5bcaa65-381c-40e7-97bd-4fc0eafc572a-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"f5bcaa65-381c-40e7-97bd-4fc0eafc572a\") " pod="openstack/openstack-cell1-galera-0" Dec 07 16:17:30 crc kubenswrapper[4716]: I1207 16:17:30.312658 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f5bcaa65-381c-40e7-97bd-4fc0eafc572a-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"f5bcaa65-381c-40e7-97bd-4fc0eafc572a\") " pod="openstack/openstack-cell1-galera-0" Dec 07 16:17:30 crc kubenswrapper[4716]: I1207 16:17:30.313181 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f5bcaa65-381c-40e7-97bd-4fc0eafc572a-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"f5bcaa65-381c-40e7-97bd-4fc0eafc572a\") " pod="openstack/openstack-cell1-galera-0" Dec 07 16:17:30 crc kubenswrapper[4716]: I1207 16:17:30.319638 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5bcaa65-381c-40e7-97bd-4fc0eafc572a-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"f5bcaa65-381c-40e7-97bd-4fc0eafc572a\") " pod="openstack/openstack-cell1-galera-0" Dec 07 16:17:30 crc kubenswrapper[4716]: I1207 16:17:30.320277 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/f5bcaa65-381c-40e7-97bd-4fc0eafc572a-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"f5bcaa65-381c-40e7-97bd-4fc0eafc572a\") " pod="openstack/openstack-cell1-galera-0" Dec 07 16:17:30 crc kubenswrapper[4716]: I1207 16:17:30.327923 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-cell1-galera-0\" (UID: \"f5bcaa65-381c-40e7-97bd-4fc0eafc572a\") " pod="openstack/openstack-cell1-galera-0" Dec 07 16:17:30 crc kubenswrapper[4716]: I1207 16:17:30.337688 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6cj9d\" (UniqueName: \"kubernetes.io/projected/f5bcaa65-381c-40e7-97bd-4fc0eafc572a-kube-api-access-6cj9d\") pod \"openstack-cell1-galera-0\" (UID: \"f5bcaa65-381c-40e7-97bd-4fc0eafc572a\") " pod="openstack/openstack-cell1-galera-0" Dec 07 16:17:30 crc kubenswrapper[4716]: I1207 16:17:30.399637 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 07 16:17:30 crc kubenswrapper[4716]: I1207 16:17:30.407762 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Dec 07 16:17:30 crc kubenswrapper[4716]: I1207 16:17:30.409021 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 07 16:17:30 crc kubenswrapper[4716]: I1207 16:17:30.411006 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Dec 07 16:17:30 crc kubenswrapper[4716]: I1207 16:17:30.411448 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-8wfxp" Dec 07 16:17:30 crc kubenswrapper[4716]: I1207 16:17:30.411629 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Dec 07 16:17:30 crc kubenswrapper[4716]: I1207 16:17:30.422106 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 07 16:17:30 crc kubenswrapper[4716]: I1207 16:17:30.506957 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-2q7l4" podUID="f98dfb6b-5a13-42e4-bca1-030f964b78ce" containerName="registry-server" containerID="cri-o://db98b29ab6b961e8d20058c659be7ccc34873b22ee13c03e7161937da0248b2b" gracePeriod=2 Dec 07 16:17:30 crc kubenswrapper[4716]: I1207 16:17:30.507054 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"c450ef40-3a23-4bab-9d64-4475f12d0490","Type":"ContainerStarted","Data":"dfa3ab1478d2b3c268f9904f9546d934415cad4c85f75d3d48eb041b83e52505"} Dec 07 16:17:30 crc kubenswrapper[4716]: I1207 16:17:30.514751 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/0f4b5de9-9760-4930-bc8e-d7e47a13ad2e-memcached-tls-certs\") pod \"memcached-0\" (UID: \"0f4b5de9-9760-4930-bc8e-d7e47a13ad2e\") " pod="openstack/memcached-0" Dec 07 16:17:30 crc kubenswrapper[4716]: I1207 16:17:30.514800 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f4b5de9-9760-4930-bc8e-d7e47a13ad2e-combined-ca-bundle\") pod \"memcached-0\" (UID: \"0f4b5de9-9760-4930-bc8e-d7e47a13ad2e\") " pod="openstack/memcached-0" Dec 07 16:17:30 crc kubenswrapper[4716]: I1207 16:17:30.514833 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0f4b5de9-9760-4930-bc8e-d7e47a13ad2e-kolla-config\") pod \"memcached-0\" (UID: \"0f4b5de9-9760-4930-bc8e-d7e47a13ad2e\") " pod="openstack/memcached-0" Dec 07 16:17:30 crc kubenswrapper[4716]: I1207 16:17:30.514867 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xckdg\" (UniqueName: \"kubernetes.io/projected/0f4b5de9-9760-4930-bc8e-d7e47a13ad2e-kube-api-access-xckdg\") pod \"memcached-0\" (UID: \"0f4b5de9-9760-4930-bc8e-d7e47a13ad2e\") " pod="openstack/memcached-0" Dec 07 16:17:30 crc kubenswrapper[4716]: I1207 16:17:30.514927 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0f4b5de9-9760-4930-bc8e-d7e47a13ad2e-config-data\") pod \"memcached-0\" (UID: \"0f4b5de9-9760-4930-bc8e-d7e47a13ad2e\") " pod="openstack/memcached-0" Dec 07 16:17:30 crc kubenswrapper[4716]: I1207 16:17:30.616869 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0f4b5de9-9760-4930-bc8e-d7e47a13ad2e-config-data\") pod \"memcached-0\" (UID: \"0f4b5de9-9760-4930-bc8e-d7e47a13ad2e\") " pod="openstack/memcached-0" Dec 07 16:17:30 crc kubenswrapper[4716]: I1207 16:17:30.616944 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/0f4b5de9-9760-4930-bc8e-d7e47a13ad2e-memcached-tls-certs\") pod \"memcached-0\" (UID: \"0f4b5de9-9760-4930-bc8e-d7e47a13ad2e\") " pod="openstack/memcached-0" Dec 07 16:17:30 crc kubenswrapper[4716]: I1207 16:17:30.616982 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f4b5de9-9760-4930-bc8e-d7e47a13ad2e-combined-ca-bundle\") pod \"memcached-0\" (UID: \"0f4b5de9-9760-4930-bc8e-d7e47a13ad2e\") " pod="openstack/memcached-0" Dec 07 16:17:30 crc kubenswrapper[4716]: I1207 16:17:30.617017 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0f4b5de9-9760-4930-bc8e-d7e47a13ad2e-kolla-config\") pod \"memcached-0\" (UID: \"0f4b5de9-9760-4930-bc8e-d7e47a13ad2e\") " pod="openstack/memcached-0" Dec 07 16:17:30 crc kubenswrapper[4716]: I1207 16:17:30.617115 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xckdg\" (UniqueName: \"kubernetes.io/projected/0f4b5de9-9760-4930-bc8e-d7e47a13ad2e-kube-api-access-xckdg\") pod \"memcached-0\" (UID: \"0f4b5de9-9760-4930-bc8e-d7e47a13ad2e\") " pod="openstack/memcached-0" Dec 07 16:17:30 crc kubenswrapper[4716]: I1207 16:17:30.618739 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0f4b5de9-9760-4930-bc8e-d7e47a13ad2e-config-data\") pod \"memcached-0\" (UID: \"0f4b5de9-9760-4930-bc8e-d7e47a13ad2e\") " pod="openstack/memcached-0" Dec 07 16:17:30 crc kubenswrapper[4716]: I1207 16:17:30.619159 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0f4b5de9-9760-4930-bc8e-d7e47a13ad2e-kolla-config\") pod \"memcached-0\" (UID: \"0f4b5de9-9760-4930-bc8e-d7e47a13ad2e\") " pod="openstack/memcached-0" Dec 07 16:17:30 crc kubenswrapper[4716]: I1207 16:17:30.627334 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f4b5de9-9760-4930-bc8e-d7e47a13ad2e-combined-ca-bundle\") pod \"memcached-0\" (UID: \"0f4b5de9-9760-4930-bc8e-d7e47a13ad2e\") " pod="openstack/memcached-0" Dec 07 16:17:30 crc kubenswrapper[4716]: I1207 16:17:30.627665 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/0f4b5de9-9760-4930-bc8e-d7e47a13ad2e-memcached-tls-certs\") pod \"memcached-0\" (UID: \"0f4b5de9-9760-4930-bc8e-d7e47a13ad2e\") " pod="openstack/memcached-0" Dec 07 16:17:30 crc kubenswrapper[4716]: I1207 16:17:30.638629 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xckdg\" (UniqueName: \"kubernetes.io/projected/0f4b5de9-9760-4930-bc8e-d7e47a13ad2e-kube-api-access-xckdg\") pod \"memcached-0\" (UID: \"0f4b5de9-9760-4930-bc8e-d7e47a13ad2e\") " pod="openstack/memcached-0" Dec 07 16:17:30 crc kubenswrapper[4716]: I1207 16:17:30.739788 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 07 16:17:30 crc kubenswrapper[4716]: W1207 16:17:30.749950 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf5bcaa65_381c_40e7_97bd_4fc0eafc572a.slice/crio-9af1cf096bdb82018a22a87b942f7087020bb803ba23ade016b89d7b4987768b WatchSource:0}: Error finding container 9af1cf096bdb82018a22a87b942f7087020bb803ba23ade016b89d7b4987768b: Status 404 returned error can't find the container with id 9af1cf096bdb82018a22a87b942f7087020bb803ba23ade016b89d7b4987768b Dec 07 16:17:30 crc kubenswrapper[4716]: I1207 16:17:30.807969 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 07 16:17:31 crc kubenswrapper[4716]: I1207 16:17:31.218767 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 07 16:17:31 crc kubenswrapper[4716]: I1207 16:17:31.516677 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"f5bcaa65-381c-40e7-97bd-4fc0eafc572a","Type":"ContainerStarted","Data":"9af1cf096bdb82018a22a87b942f7087020bb803ba23ade016b89d7b4987768b"} Dec 07 16:17:31 crc kubenswrapper[4716]: I1207 16:17:31.518242 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"0f4b5de9-9760-4930-bc8e-d7e47a13ad2e","Type":"ContainerStarted","Data":"0589ebc15630f3f402a214eee21b1d998f8d613d110816097c443db1b8b098b6"} Dec 07 16:17:31 crc kubenswrapper[4716]: I1207 16:17:31.520830 4716 generic.go:334] "Generic (PLEG): container finished" podID="f98dfb6b-5a13-42e4-bca1-030f964b78ce" containerID="db98b29ab6b961e8d20058c659be7ccc34873b22ee13c03e7161937da0248b2b" exitCode=0 Dec 07 16:17:31 crc kubenswrapper[4716]: I1207 16:17:31.520887 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2q7l4" event={"ID":"f98dfb6b-5a13-42e4-bca1-030f964b78ce","Type":"ContainerDied","Data":"db98b29ab6b961e8d20058c659be7ccc34873b22ee13c03e7161937da0248b2b"} Dec 07 16:17:32 crc kubenswrapper[4716]: I1207 16:17:32.297117 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 07 16:17:32 crc kubenswrapper[4716]: I1207 16:17:32.298258 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 07 16:17:32 crc kubenswrapper[4716]: I1207 16:17:32.300203 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-czc42" Dec 07 16:17:32 crc kubenswrapper[4716]: I1207 16:17:32.311742 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 07 16:17:32 crc kubenswrapper[4716]: I1207 16:17:32.349639 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-glmlq\" (UniqueName: \"kubernetes.io/projected/c088cd1b-83a8-4f05-b7db-11c8ffb59fab-kube-api-access-glmlq\") pod \"kube-state-metrics-0\" (UID: \"c088cd1b-83a8-4f05-b7db-11c8ffb59fab\") " pod="openstack/kube-state-metrics-0" Dec 07 16:17:32 crc kubenswrapper[4716]: I1207 16:17:32.450854 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-glmlq\" (UniqueName: \"kubernetes.io/projected/c088cd1b-83a8-4f05-b7db-11c8ffb59fab-kube-api-access-glmlq\") pod \"kube-state-metrics-0\" (UID: \"c088cd1b-83a8-4f05-b7db-11c8ffb59fab\") " pod="openstack/kube-state-metrics-0" Dec 07 16:17:32 crc kubenswrapper[4716]: I1207 16:17:32.477900 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-glmlq\" (UniqueName: \"kubernetes.io/projected/c088cd1b-83a8-4f05-b7db-11c8ffb59fab-kube-api-access-glmlq\") pod \"kube-state-metrics-0\" (UID: \"c088cd1b-83a8-4f05-b7db-11c8ffb59fab\") " pod="openstack/kube-state-metrics-0" Dec 07 16:17:32 crc kubenswrapper[4716]: I1207 16:17:32.620515 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 07 16:17:34 crc kubenswrapper[4716]: I1207 16:17:34.130158 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2q7l4" Dec 07 16:17:34 crc kubenswrapper[4716]: I1207 16:17:34.194954 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f98dfb6b-5a13-42e4-bca1-030f964b78ce-utilities\") pod \"f98dfb6b-5a13-42e4-bca1-030f964b78ce\" (UID: \"f98dfb6b-5a13-42e4-bca1-030f964b78ce\") " Dec 07 16:17:34 crc kubenswrapper[4716]: I1207 16:17:34.195062 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-twkvk\" (UniqueName: \"kubernetes.io/projected/f98dfb6b-5a13-42e4-bca1-030f964b78ce-kube-api-access-twkvk\") pod \"f98dfb6b-5a13-42e4-bca1-030f964b78ce\" (UID: \"f98dfb6b-5a13-42e4-bca1-030f964b78ce\") " Dec 07 16:17:34 crc kubenswrapper[4716]: I1207 16:17:34.195172 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f98dfb6b-5a13-42e4-bca1-030f964b78ce-catalog-content\") pod \"f98dfb6b-5a13-42e4-bca1-030f964b78ce\" (UID: \"f98dfb6b-5a13-42e4-bca1-030f964b78ce\") " Dec 07 16:17:34 crc kubenswrapper[4716]: I1207 16:17:34.196032 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f98dfb6b-5a13-42e4-bca1-030f964b78ce-utilities" (OuterVolumeSpecName: "utilities") pod "f98dfb6b-5a13-42e4-bca1-030f964b78ce" (UID: "f98dfb6b-5a13-42e4-bca1-030f964b78ce"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:17:34 crc kubenswrapper[4716]: I1207 16:17:34.200480 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f98dfb6b-5a13-42e4-bca1-030f964b78ce-kube-api-access-twkvk" (OuterVolumeSpecName: "kube-api-access-twkvk") pod "f98dfb6b-5a13-42e4-bca1-030f964b78ce" (UID: "f98dfb6b-5a13-42e4-bca1-030f964b78ce"). InnerVolumeSpecName "kube-api-access-twkvk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:17:34 crc kubenswrapper[4716]: I1207 16:17:34.296469 4716 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f98dfb6b-5a13-42e4-bca1-030f964b78ce-utilities\") on node \"crc\" DevicePath \"\"" Dec 07 16:17:34 crc kubenswrapper[4716]: I1207 16:17:34.296499 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-twkvk\" (UniqueName: \"kubernetes.io/projected/f98dfb6b-5a13-42e4-bca1-030f964b78ce-kube-api-access-twkvk\") on node \"crc\" DevicePath \"\"" Dec 07 16:17:34 crc kubenswrapper[4716]: I1207 16:17:34.322118 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f98dfb6b-5a13-42e4-bca1-030f964b78ce-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f98dfb6b-5a13-42e4-bca1-030f964b78ce" (UID: "f98dfb6b-5a13-42e4-bca1-030f964b78ce"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:17:34 crc kubenswrapper[4716]: I1207 16:17:34.398484 4716 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f98dfb6b-5a13-42e4-bca1-030f964b78ce-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 07 16:17:34 crc kubenswrapper[4716]: I1207 16:17:34.551755 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8phkg"] Dec 07 16:17:34 crc kubenswrapper[4716]: E1207 16:17:34.552121 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f98dfb6b-5a13-42e4-bca1-030f964b78ce" containerName="extract-utilities" Dec 07 16:17:34 crc kubenswrapper[4716]: I1207 16:17:34.552133 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="f98dfb6b-5a13-42e4-bca1-030f964b78ce" containerName="extract-utilities" Dec 07 16:17:34 crc kubenswrapper[4716]: E1207 16:17:34.552150 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f98dfb6b-5a13-42e4-bca1-030f964b78ce" containerName="registry-server" Dec 07 16:17:34 crc kubenswrapper[4716]: I1207 16:17:34.552156 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="f98dfb6b-5a13-42e4-bca1-030f964b78ce" containerName="registry-server" Dec 07 16:17:34 crc kubenswrapper[4716]: E1207 16:17:34.552172 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f98dfb6b-5a13-42e4-bca1-030f964b78ce" containerName="extract-content" Dec 07 16:17:34 crc kubenswrapper[4716]: I1207 16:17:34.552178 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="f98dfb6b-5a13-42e4-bca1-030f964b78ce" containerName="extract-content" Dec 07 16:17:34 crc kubenswrapper[4716]: I1207 16:17:34.552318 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="f98dfb6b-5a13-42e4-bca1-030f964b78ce" containerName="registry-server" Dec 07 16:17:34 crc kubenswrapper[4716]: I1207 16:17:34.553560 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8phkg" Dec 07 16:17:34 crc kubenswrapper[4716]: I1207 16:17:34.562320 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8phkg"] Dec 07 16:17:34 crc kubenswrapper[4716]: I1207 16:17:34.581282 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2q7l4" event={"ID":"f98dfb6b-5a13-42e4-bca1-030f964b78ce","Type":"ContainerDied","Data":"8c77248ccf81982da297e70d1f01df70fddcecc3bbe829e1707c28ff865951b7"} Dec 07 16:17:34 crc kubenswrapper[4716]: I1207 16:17:34.581341 4716 scope.go:117] "RemoveContainer" containerID="db98b29ab6b961e8d20058c659be7ccc34873b22ee13c03e7161937da0248b2b" Dec 07 16:17:34 crc kubenswrapper[4716]: I1207 16:17:34.581371 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2q7l4" Dec 07 16:17:34 crc kubenswrapper[4716]: I1207 16:17:34.615480 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2q7l4"] Dec 07 16:17:34 crc kubenswrapper[4716]: I1207 16:17:34.620995 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-2q7l4"] Dec 07 16:17:34 crc kubenswrapper[4716]: I1207 16:17:34.703744 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fef1e486-acca-4cc4-b2b1-3b20270d418a-utilities\") pod \"certified-operators-8phkg\" (UID: \"fef1e486-acca-4cc4-b2b1-3b20270d418a\") " pod="openshift-marketplace/certified-operators-8phkg" Dec 07 16:17:34 crc kubenswrapper[4716]: I1207 16:17:34.703799 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zz9t8\" (UniqueName: \"kubernetes.io/projected/fef1e486-acca-4cc4-b2b1-3b20270d418a-kube-api-access-zz9t8\") pod \"certified-operators-8phkg\" (UID: \"fef1e486-acca-4cc4-b2b1-3b20270d418a\") " pod="openshift-marketplace/certified-operators-8phkg" Dec 07 16:17:34 crc kubenswrapper[4716]: I1207 16:17:34.703825 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fef1e486-acca-4cc4-b2b1-3b20270d418a-catalog-content\") pod \"certified-operators-8phkg\" (UID: \"fef1e486-acca-4cc4-b2b1-3b20270d418a\") " pod="openshift-marketplace/certified-operators-8phkg" Dec 07 16:17:34 crc kubenswrapper[4716]: I1207 16:17:34.806293 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fef1e486-acca-4cc4-b2b1-3b20270d418a-utilities\") pod \"certified-operators-8phkg\" (UID: \"fef1e486-acca-4cc4-b2b1-3b20270d418a\") " pod="openshift-marketplace/certified-operators-8phkg" Dec 07 16:17:34 crc kubenswrapper[4716]: I1207 16:17:34.806359 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zz9t8\" (UniqueName: \"kubernetes.io/projected/fef1e486-acca-4cc4-b2b1-3b20270d418a-kube-api-access-zz9t8\") pod \"certified-operators-8phkg\" (UID: \"fef1e486-acca-4cc4-b2b1-3b20270d418a\") " pod="openshift-marketplace/certified-operators-8phkg" Dec 07 16:17:34 crc kubenswrapper[4716]: I1207 16:17:34.806383 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fef1e486-acca-4cc4-b2b1-3b20270d418a-catalog-content\") pod \"certified-operators-8phkg\" (UID: \"fef1e486-acca-4cc4-b2b1-3b20270d418a\") " pod="openshift-marketplace/certified-operators-8phkg" Dec 07 16:17:34 crc kubenswrapper[4716]: I1207 16:17:34.806958 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fef1e486-acca-4cc4-b2b1-3b20270d418a-catalog-content\") pod \"certified-operators-8phkg\" (UID: \"fef1e486-acca-4cc4-b2b1-3b20270d418a\") " pod="openshift-marketplace/certified-operators-8phkg" Dec 07 16:17:34 crc kubenswrapper[4716]: I1207 16:17:34.807219 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fef1e486-acca-4cc4-b2b1-3b20270d418a-utilities\") pod \"certified-operators-8phkg\" (UID: \"fef1e486-acca-4cc4-b2b1-3b20270d418a\") " pod="openshift-marketplace/certified-operators-8phkg" Dec 07 16:17:34 crc kubenswrapper[4716]: I1207 16:17:34.836008 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zz9t8\" (UniqueName: \"kubernetes.io/projected/fef1e486-acca-4cc4-b2b1-3b20270d418a-kube-api-access-zz9t8\") pod \"certified-operators-8phkg\" (UID: \"fef1e486-acca-4cc4-b2b1-3b20270d418a\") " pod="openshift-marketplace/certified-operators-8phkg" Dec 07 16:17:34 crc kubenswrapper[4716]: I1207 16:17:34.879770 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8phkg" Dec 07 16:17:35 crc kubenswrapper[4716]: I1207 16:17:35.675465 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f98dfb6b-5a13-42e4-bca1-030f964b78ce" path="/var/lib/kubelet/pods/f98dfb6b-5a13-42e4-bca1-030f964b78ce/volumes" Dec 07 16:17:35 crc kubenswrapper[4716]: I1207 16:17:35.866957 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-tgl27"] Dec 07 16:17:35 crc kubenswrapper[4716]: I1207 16:17:35.868275 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-tgl27" Dec 07 16:17:35 crc kubenswrapper[4716]: I1207 16:17:35.880412 4716 scope.go:117] "RemoveContainer" containerID="cf57458fb0cb2dccc8bdf4918c05ed2518feea9353396a4b3c291293f6f3ceae" Dec 07 16:17:35 crc kubenswrapper[4716]: I1207 16:17:35.882201 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-tgl27"] Dec 07 16:17:35 crc kubenswrapper[4716]: I1207 16:17:35.885817 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-xqg7n" Dec 07 16:17:35 crc kubenswrapper[4716]: I1207 16:17:35.886132 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Dec 07 16:17:35 crc kubenswrapper[4716]: I1207 16:17:35.886342 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Dec 07 16:17:35 crc kubenswrapper[4716]: I1207 16:17:35.900433 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-g6nwl"] Dec 07 16:17:35 crc kubenswrapper[4716]: I1207 16:17:35.904933 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-g6nwl" Dec 07 16:17:35 crc kubenswrapper[4716]: I1207 16:17:35.914987 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-g6nwl"] Dec 07 16:17:35 crc kubenswrapper[4716]: I1207 16:17:35.931992 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/399984bf-94af-4630-a7c3-9375a388159a-var-log-ovn\") pod \"ovn-controller-tgl27\" (UID: \"399984bf-94af-4630-a7c3-9375a388159a\") " pod="openstack/ovn-controller-tgl27" Dec 07 16:17:35 crc kubenswrapper[4716]: I1207 16:17:35.932086 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/399984bf-94af-4630-a7c3-9375a388159a-var-run-ovn\") pod \"ovn-controller-tgl27\" (UID: \"399984bf-94af-4630-a7c3-9375a388159a\") " pod="openstack/ovn-controller-tgl27" Dec 07 16:17:35 crc kubenswrapper[4716]: I1207 16:17:35.932103 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/399984bf-94af-4630-a7c3-9375a388159a-combined-ca-bundle\") pod \"ovn-controller-tgl27\" (UID: \"399984bf-94af-4630-a7c3-9375a388159a\") " pod="openstack/ovn-controller-tgl27" Dec 07 16:17:35 crc kubenswrapper[4716]: I1207 16:17:35.932122 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/399984bf-94af-4630-a7c3-9375a388159a-ovn-controller-tls-certs\") pod \"ovn-controller-tgl27\" (UID: \"399984bf-94af-4630-a7c3-9375a388159a\") " pod="openstack/ovn-controller-tgl27" Dec 07 16:17:35 crc kubenswrapper[4716]: I1207 16:17:35.932142 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/399984bf-94af-4630-a7c3-9375a388159a-scripts\") pod \"ovn-controller-tgl27\" (UID: \"399984bf-94af-4630-a7c3-9375a388159a\") " pod="openstack/ovn-controller-tgl27" Dec 07 16:17:35 crc kubenswrapper[4716]: I1207 16:17:35.932194 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bnxth\" (UniqueName: \"kubernetes.io/projected/399984bf-94af-4630-a7c3-9375a388159a-kube-api-access-bnxth\") pod \"ovn-controller-tgl27\" (UID: \"399984bf-94af-4630-a7c3-9375a388159a\") " pod="openstack/ovn-controller-tgl27" Dec 07 16:17:35 crc kubenswrapper[4716]: I1207 16:17:35.932222 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/399984bf-94af-4630-a7c3-9375a388159a-var-run\") pod \"ovn-controller-tgl27\" (UID: \"399984bf-94af-4630-a7c3-9375a388159a\") " pod="openstack/ovn-controller-tgl27" Dec 07 16:17:36 crc kubenswrapper[4716]: I1207 16:17:36.034129 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/0fc0f0b8-b60a-4aff-84dd-bbec37b12d81-etc-ovs\") pod \"ovn-controller-ovs-g6nwl\" (UID: \"0fc0f0b8-b60a-4aff-84dd-bbec37b12d81\") " pod="openstack/ovn-controller-ovs-g6nwl" Dec 07 16:17:36 crc kubenswrapper[4716]: I1207 16:17:36.034213 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bnxth\" (UniqueName: \"kubernetes.io/projected/399984bf-94af-4630-a7c3-9375a388159a-kube-api-access-bnxth\") pod \"ovn-controller-tgl27\" (UID: \"399984bf-94af-4630-a7c3-9375a388159a\") " pod="openstack/ovn-controller-tgl27" Dec 07 16:17:36 crc kubenswrapper[4716]: I1207 16:17:36.034245 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/0fc0f0b8-b60a-4aff-84dd-bbec37b12d81-var-lib\") pod \"ovn-controller-ovs-g6nwl\" (UID: \"0fc0f0b8-b60a-4aff-84dd-bbec37b12d81\") " pod="openstack/ovn-controller-ovs-g6nwl" Dec 07 16:17:36 crc kubenswrapper[4716]: I1207 16:17:36.034269 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/399984bf-94af-4630-a7c3-9375a388159a-var-run\") pod \"ovn-controller-tgl27\" (UID: \"399984bf-94af-4630-a7c3-9375a388159a\") " pod="openstack/ovn-controller-tgl27" Dec 07 16:17:36 crc kubenswrapper[4716]: I1207 16:17:36.034294 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/0fc0f0b8-b60a-4aff-84dd-bbec37b12d81-var-log\") pod \"ovn-controller-ovs-g6nwl\" (UID: \"0fc0f0b8-b60a-4aff-84dd-bbec37b12d81\") " pod="openstack/ovn-controller-ovs-g6nwl" Dec 07 16:17:36 crc kubenswrapper[4716]: I1207 16:17:36.034335 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0fc0f0b8-b60a-4aff-84dd-bbec37b12d81-scripts\") pod \"ovn-controller-ovs-g6nwl\" (UID: \"0fc0f0b8-b60a-4aff-84dd-bbec37b12d81\") " pod="openstack/ovn-controller-ovs-g6nwl" Dec 07 16:17:36 crc kubenswrapper[4716]: I1207 16:17:36.034377 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/399984bf-94af-4630-a7c3-9375a388159a-var-log-ovn\") pod \"ovn-controller-tgl27\" (UID: \"399984bf-94af-4630-a7c3-9375a388159a\") " pod="openstack/ovn-controller-tgl27" Dec 07 16:17:36 crc kubenswrapper[4716]: I1207 16:17:36.034397 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0fc0f0b8-b60a-4aff-84dd-bbec37b12d81-var-run\") pod \"ovn-controller-ovs-g6nwl\" (UID: \"0fc0f0b8-b60a-4aff-84dd-bbec37b12d81\") " pod="openstack/ovn-controller-ovs-g6nwl" Dec 07 16:17:36 crc kubenswrapper[4716]: I1207 16:17:36.034418 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqrq8\" (UniqueName: \"kubernetes.io/projected/0fc0f0b8-b60a-4aff-84dd-bbec37b12d81-kube-api-access-bqrq8\") pod \"ovn-controller-ovs-g6nwl\" (UID: \"0fc0f0b8-b60a-4aff-84dd-bbec37b12d81\") " pod="openstack/ovn-controller-ovs-g6nwl" Dec 07 16:17:36 crc kubenswrapper[4716]: I1207 16:17:36.034446 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/399984bf-94af-4630-a7c3-9375a388159a-var-run-ovn\") pod \"ovn-controller-tgl27\" (UID: \"399984bf-94af-4630-a7c3-9375a388159a\") " pod="openstack/ovn-controller-tgl27" Dec 07 16:17:36 crc kubenswrapper[4716]: I1207 16:17:36.034469 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/399984bf-94af-4630-a7c3-9375a388159a-combined-ca-bundle\") pod \"ovn-controller-tgl27\" (UID: \"399984bf-94af-4630-a7c3-9375a388159a\") " pod="openstack/ovn-controller-tgl27" Dec 07 16:17:36 crc kubenswrapper[4716]: I1207 16:17:36.034490 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/399984bf-94af-4630-a7c3-9375a388159a-ovn-controller-tls-certs\") pod \"ovn-controller-tgl27\" (UID: \"399984bf-94af-4630-a7c3-9375a388159a\") " pod="openstack/ovn-controller-tgl27" Dec 07 16:17:36 crc kubenswrapper[4716]: I1207 16:17:36.034512 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/399984bf-94af-4630-a7c3-9375a388159a-scripts\") pod \"ovn-controller-tgl27\" (UID: \"399984bf-94af-4630-a7c3-9375a388159a\") " pod="openstack/ovn-controller-tgl27" Dec 07 16:17:36 crc kubenswrapper[4716]: I1207 16:17:36.037060 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/399984bf-94af-4630-a7c3-9375a388159a-scripts\") pod \"ovn-controller-tgl27\" (UID: \"399984bf-94af-4630-a7c3-9375a388159a\") " pod="openstack/ovn-controller-tgl27" Dec 07 16:17:36 crc kubenswrapper[4716]: I1207 16:17:36.044810 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/399984bf-94af-4630-a7c3-9375a388159a-var-run\") pod \"ovn-controller-tgl27\" (UID: \"399984bf-94af-4630-a7c3-9375a388159a\") " pod="openstack/ovn-controller-tgl27" Dec 07 16:17:36 crc kubenswrapper[4716]: I1207 16:17:36.045208 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/399984bf-94af-4630-a7c3-9375a388159a-var-run-ovn\") pod \"ovn-controller-tgl27\" (UID: \"399984bf-94af-4630-a7c3-9375a388159a\") " pod="openstack/ovn-controller-tgl27" Dec 07 16:17:36 crc kubenswrapper[4716]: I1207 16:17:36.048131 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 07 16:17:36 crc kubenswrapper[4716]: I1207 16:17:36.051202 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 07 16:17:36 crc kubenswrapper[4716]: I1207 16:17:36.052713 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/399984bf-94af-4630-a7c3-9375a388159a-var-log-ovn\") pod \"ovn-controller-tgl27\" (UID: \"399984bf-94af-4630-a7c3-9375a388159a\") " pod="openstack/ovn-controller-tgl27" Dec 07 16:17:36 crc kubenswrapper[4716]: I1207 16:17:36.058841 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-d58mm" Dec 07 16:17:36 crc kubenswrapper[4716]: I1207 16:17:36.060582 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Dec 07 16:17:36 crc kubenswrapper[4716]: I1207 16:17:36.060898 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Dec 07 16:17:36 crc kubenswrapper[4716]: I1207 16:17:36.061095 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Dec 07 16:17:36 crc kubenswrapper[4716]: I1207 16:17:36.061248 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Dec 07 16:17:36 crc kubenswrapper[4716]: I1207 16:17:36.066144 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/399984bf-94af-4630-a7c3-9375a388159a-ovn-controller-tls-certs\") pod \"ovn-controller-tgl27\" (UID: \"399984bf-94af-4630-a7c3-9375a388159a\") " pod="openstack/ovn-controller-tgl27" Dec 07 16:17:36 crc kubenswrapper[4716]: I1207 16:17:36.067839 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/399984bf-94af-4630-a7c3-9375a388159a-combined-ca-bundle\") pod \"ovn-controller-tgl27\" (UID: \"399984bf-94af-4630-a7c3-9375a388159a\") " pod="openstack/ovn-controller-tgl27" Dec 07 16:17:36 crc kubenswrapper[4716]: I1207 16:17:36.072940 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bnxth\" (UniqueName: \"kubernetes.io/projected/399984bf-94af-4630-a7c3-9375a388159a-kube-api-access-bnxth\") pod \"ovn-controller-tgl27\" (UID: \"399984bf-94af-4630-a7c3-9375a388159a\") " pod="openstack/ovn-controller-tgl27" Dec 07 16:17:36 crc kubenswrapper[4716]: I1207 16:17:36.074676 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 07 16:17:36 crc kubenswrapper[4716]: I1207 16:17:36.135479 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/771a3c3e-34e4-43f6-a3e2-8e7156e081d1-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"771a3c3e-34e4-43f6-a3e2-8e7156e081d1\") " pod="openstack/ovsdbserver-nb-0" Dec 07 16:17:36 crc kubenswrapper[4716]: I1207 16:17:36.135879 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/0fc0f0b8-b60a-4aff-84dd-bbec37b12d81-var-lib\") pod \"ovn-controller-ovs-g6nwl\" (UID: \"0fc0f0b8-b60a-4aff-84dd-bbec37b12d81\") " pod="openstack/ovn-controller-ovs-g6nwl" Dec 07 16:17:36 crc kubenswrapper[4716]: I1207 16:17:36.135920 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/0fc0f0b8-b60a-4aff-84dd-bbec37b12d81-var-log\") pod \"ovn-controller-ovs-g6nwl\" (UID: \"0fc0f0b8-b60a-4aff-84dd-bbec37b12d81\") " pod="openstack/ovn-controller-ovs-g6nwl" Dec 07 16:17:36 crc kubenswrapper[4716]: I1207 16:17:36.135942 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-nb-0\" (UID: \"771a3c3e-34e4-43f6-a3e2-8e7156e081d1\") " pod="openstack/ovsdbserver-nb-0" Dec 07 16:17:36 crc kubenswrapper[4716]: I1207 16:17:36.135987 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0fc0f0b8-b60a-4aff-84dd-bbec37b12d81-scripts\") pod \"ovn-controller-ovs-g6nwl\" (UID: \"0fc0f0b8-b60a-4aff-84dd-bbec37b12d81\") " pod="openstack/ovn-controller-ovs-g6nwl" Dec 07 16:17:36 crc kubenswrapper[4716]: I1207 16:17:36.136011 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/771a3c3e-34e4-43f6-a3e2-8e7156e081d1-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"771a3c3e-34e4-43f6-a3e2-8e7156e081d1\") " pod="openstack/ovsdbserver-nb-0" Dec 07 16:17:36 crc kubenswrapper[4716]: I1207 16:17:36.136047 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ghbx\" (UniqueName: \"kubernetes.io/projected/771a3c3e-34e4-43f6-a3e2-8e7156e081d1-kube-api-access-9ghbx\") pod \"ovsdbserver-nb-0\" (UID: \"771a3c3e-34e4-43f6-a3e2-8e7156e081d1\") " pod="openstack/ovsdbserver-nb-0" Dec 07 16:17:36 crc kubenswrapper[4716]: I1207 16:17:36.136111 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0fc0f0b8-b60a-4aff-84dd-bbec37b12d81-var-run\") pod \"ovn-controller-ovs-g6nwl\" (UID: \"0fc0f0b8-b60a-4aff-84dd-bbec37b12d81\") " pod="openstack/ovn-controller-ovs-g6nwl" Dec 07 16:17:36 crc kubenswrapper[4716]: I1207 16:17:36.136145 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqrq8\" (UniqueName: \"kubernetes.io/projected/0fc0f0b8-b60a-4aff-84dd-bbec37b12d81-kube-api-access-bqrq8\") pod \"ovn-controller-ovs-g6nwl\" (UID: \"0fc0f0b8-b60a-4aff-84dd-bbec37b12d81\") " pod="openstack/ovn-controller-ovs-g6nwl" Dec 07 16:17:36 crc kubenswrapper[4716]: I1207 16:17:36.136178 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/771a3c3e-34e4-43f6-a3e2-8e7156e081d1-config\") pod \"ovsdbserver-nb-0\" (UID: \"771a3c3e-34e4-43f6-a3e2-8e7156e081d1\") " pod="openstack/ovsdbserver-nb-0" Dec 07 16:17:36 crc kubenswrapper[4716]: I1207 16:17:36.136204 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/771a3c3e-34e4-43f6-a3e2-8e7156e081d1-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"771a3c3e-34e4-43f6-a3e2-8e7156e081d1\") " pod="openstack/ovsdbserver-nb-0" Dec 07 16:17:36 crc kubenswrapper[4716]: I1207 16:17:36.136231 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/0fc0f0b8-b60a-4aff-84dd-bbec37b12d81-etc-ovs\") pod \"ovn-controller-ovs-g6nwl\" (UID: \"0fc0f0b8-b60a-4aff-84dd-bbec37b12d81\") " pod="openstack/ovn-controller-ovs-g6nwl" Dec 07 16:17:36 crc kubenswrapper[4716]: I1207 16:17:36.136250 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/771a3c3e-34e4-43f6-a3e2-8e7156e081d1-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"771a3c3e-34e4-43f6-a3e2-8e7156e081d1\") " pod="openstack/ovsdbserver-nb-0" Dec 07 16:17:36 crc kubenswrapper[4716]: I1207 16:17:36.136265 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/771a3c3e-34e4-43f6-a3e2-8e7156e081d1-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"771a3c3e-34e4-43f6-a3e2-8e7156e081d1\") " pod="openstack/ovsdbserver-nb-0" Dec 07 16:17:36 crc kubenswrapper[4716]: I1207 16:17:36.136259 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/0fc0f0b8-b60a-4aff-84dd-bbec37b12d81-var-lib\") pod \"ovn-controller-ovs-g6nwl\" (UID: \"0fc0f0b8-b60a-4aff-84dd-bbec37b12d81\") " pod="openstack/ovn-controller-ovs-g6nwl" Dec 07 16:17:36 crc kubenswrapper[4716]: I1207 16:17:36.136328 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0fc0f0b8-b60a-4aff-84dd-bbec37b12d81-var-run\") pod \"ovn-controller-ovs-g6nwl\" (UID: \"0fc0f0b8-b60a-4aff-84dd-bbec37b12d81\") " pod="openstack/ovn-controller-ovs-g6nwl" Dec 07 16:17:36 crc kubenswrapper[4716]: I1207 16:17:36.136417 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/0fc0f0b8-b60a-4aff-84dd-bbec37b12d81-var-log\") pod \"ovn-controller-ovs-g6nwl\" (UID: \"0fc0f0b8-b60a-4aff-84dd-bbec37b12d81\") " pod="openstack/ovn-controller-ovs-g6nwl" Dec 07 16:17:36 crc kubenswrapper[4716]: I1207 16:17:36.136922 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/0fc0f0b8-b60a-4aff-84dd-bbec37b12d81-etc-ovs\") pod \"ovn-controller-ovs-g6nwl\" (UID: \"0fc0f0b8-b60a-4aff-84dd-bbec37b12d81\") " pod="openstack/ovn-controller-ovs-g6nwl" Dec 07 16:17:36 crc kubenswrapper[4716]: I1207 16:17:36.138894 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0fc0f0b8-b60a-4aff-84dd-bbec37b12d81-scripts\") pod \"ovn-controller-ovs-g6nwl\" (UID: \"0fc0f0b8-b60a-4aff-84dd-bbec37b12d81\") " pod="openstack/ovn-controller-ovs-g6nwl" Dec 07 16:17:36 crc kubenswrapper[4716]: I1207 16:17:36.169327 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqrq8\" (UniqueName: \"kubernetes.io/projected/0fc0f0b8-b60a-4aff-84dd-bbec37b12d81-kube-api-access-bqrq8\") pod \"ovn-controller-ovs-g6nwl\" (UID: \"0fc0f0b8-b60a-4aff-84dd-bbec37b12d81\") " pod="openstack/ovn-controller-ovs-g6nwl" Dec 07 16:17:36 crc kubenswrapper[4716]: I1207 16:17:36.219765 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-tgl27" Dec 07 16:17:36 crc kubenswrapper[4716]: I1207 16:17:36.237666 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/771a3c3e-34e4-43f6-a3e2-8e7156e081d1-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"771a3c3e-34e4-43f6-a3e2-8e7156e081d1\") " pod="openstack/ovsdbserver-nb-0" Dec 07 16:17:36 crc kubenswrapper[4716]: I1207 16:17:36.237713 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ghbx\" (UniqueName: \"kubernetes.io/projected/771a3c3e-34e4-43f6-a3e2-8e7156e081d1-kube-api-access-9ghbx\") pod \"ovsdbserver-nb-0\" (UID: \"771a3c3e-34e4-43f6-a3e2-8e7156e081d1\") " pod="openstack/ovsdbserver-nb-0" Dec 07 16:17:36 crc kubenswrapper[4716]: I1207 16:17:36.237795 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/771a3c3e-34e4-43f6-a3e2-8e7156e081d1-config\") pod \"ovsdbserver-nb-0\" (UID: \"771a3c3e-34e4-43f6-a3e2-8e7156e081d1\") " pod="openstack/ovsdbserver-nb-0" Dec 07 16:17:36 crc kubenswrapper[4716]: I1207 16:17:36.237825 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/771a3c3e-34e4-43f6-a3e2-8e7156e081d1-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"771a3c3e-34e4-43f6-a3e2-8e7156e081d1\") " pod="openstack/ovsdbserver-nb-0" Dec 07 16:17:36 crc kubenswrapper[4716]: I1207 16:17:36.237853 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/771a3c3e-34e4-43f6-a3e2-8e7156e081d1-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"771a3c3e-34e4-43f6-a3e2-8e7156e081d1\") " pod="openstack/ovsdbserver-nb-0" Dec 07 16:17:36 crc kubenswrapper[4716]: I1207 16:17:36.237869 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/771a3c3e-34e4-43f6-a3e2-8e7156e081d1-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"771a3c3e-34e4-43f6-a3e2-8e7156e081d1\") " pod="openstack/ovsdbserver-nb-0" Dec 07 16:17:36 crc kubenswrapper[4716]: I1207 16:17:36.237899 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/771a3c3e-34e4-43f6-a3e2-8e7156e081d1-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"771a3c3e-34e4-43f6-a3e2-8e7156e081d1\") " pod="openstack/ovsdbserver-nb-0" Dec 07 16:17:36 crc kubenswrapper[4716]: I1207 16:17:36.237933 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-nb-0\" (UID: \"771a3c3e-34e4-43f6-a3e2-8e7156e081d1\") " pod="openstack/ovsdbserver-nb-0" Dec 07 16:17:36 crc kubenswrapper[4716]: I1207 16:17:36.238606 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/771a3c3e-34e4-43f6-a3e2-8e7156e081d1-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"771a3c3e-34e4-43f6-a3e2-8e7156e081d1\") " pod="openstack/ovsdbserver-nb-0" Dec 07 16:17:36 crc kubenswrapper[4716]: I1207 16:17:36.238750 4716 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-nb-0\" (UID: \"771a3c3e-34e4-43f6-a3e2-8e7156e081d1\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/ovsdbserver-nb-0" Dec 07 16:17:36 crc kubenswrapper[4716]: I1207 16:17:36.239474 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/771a3c3e-34e4-43f6-a3e2-8e7156e081d1-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"771a3c3e-34e4-43f6-a3e2-8e7156e081d1\") " pod="openstack/ovsdbserver-nb-0" Dec 07 16:17:36 crc kubenswrapper[4716]: I1207 16:17:36.239535 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/771a3c3e-34e4-43f6-a3e2-8e7156e081d1-config\") pod \"ovsdbserver-nb-0\" (UID: \"771a3c3e-34e4-43f6-a3e2-8e7156e081d1\") " pod="openstack/ovsdbserver-nb-0" Dec 07 16:17:36 crc kubenswrapper[4716]: I1207 16:17:36.246454 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/771a3c3e-34e4-43f6-a3e2-8e7156e081d1-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"771a3c3e-34e4-43f6-a3e2-8e7156e081d1\") " pod="openstack/ovsdbserver-nb-0" Dec 07 16:17:36 crc kubenswrapper[4716]: I1207 16:17:36.246512 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-g6nwl" Dec 07 16:17:36 crc kubenswrapper[4716]: I1207 16:17:36.250044 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/771a3c3e-34e4-43f6-a3e2-8e7156e081d1-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"771a3c3e-34e4-43f6-a3e2-8e7156e081d1\") " pod="openstack/ovsdbserver-nb-0" Dec 07 16:17:36 crc kubenswrapper[4716]: I1207 16:17:36.250253 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/771a3c3e-34e4-43f6-a3e2-8e7156e081d1-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"771a3c3e-34e4-43f6-a3e2-8e7156e081d1\") " pod="openstack/ovsdbserver-nb-0" Dec 07 16:17:36 crc kubenswrapper[4716]: I1207 16:17:36.259613 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ghbx\" (UniqueName: \"kubernetes.io/projected/771a3c3e-34e4-43f6-a3e2-8e7156e081d1-kube-api-access-9ghbx\") pod \"ovsdbserver-nb-0\" (UID: \"771a3c3e-34e4-43f6-a3e2-8e7156e081d1\") " pod="openstack/ovsdbserver-nb-0" Dec 07 16:17:36 crc kubenswrapper[4716]: I1207 16:17:36.270442 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-nb-0\" (UID: \"771a3c3e-34e4-43f6-a3e2-8e7156e081d1\") " pod="openstack/ovsdbserver-nb-0" Dec 07 16:17:36 crc kubenswrapper[4716]: I1207 16:17:36.454710 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 07 16:17:39 crc kubenswrapper[4716]: I1207 16:17:39.749927 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 07 16:17:39 crc kubenswrapper[4716]: I1207 16:17:39.751678 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 07 16:17:39 crc kubenswrapper[4716]: I1207 16:17:39.759804 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Dec 07 16:17:39 crc kubenswrapper[4716]: I1207 16:17:39.760068 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Dec 07 16:17:39 crc kubenswrapper[4716]: I1207 16:17:39.760338 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-jxnbs" Dec 07 16:17:39 crc kubenswrapper[4716]: I1207 16:17:39.761215 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Dec 07 16:17:39 crc kubenswrapper[4716]: I1207 16:17:39.761936 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 07 16:17:39 crc kubenswrapper[4716]: I1207 16:17:39.823395 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-sb-0\" (UID: \"5c11a9b9-34de-42f7-9399-3cddce1af7c9\") " pod="openstack/ovsdbserver-sb-0" Dec 07 16:17:39 crc kubenswrapper[4716]: I1207 16:17:39.823469 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/5c11a9b9-34de-42f7-9399-3cddce1af7c9-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"5c11a9b9-34de-42f7-9399-3cddce1af7c9\") " pod="openstack/ovsdbserver-sb-0" Dec 07 16:17:39 crc kubenswrapper[4716]: I1207 16:17:39.823517 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c9lbv\" (UniqueName: \"kubernetes.io/projected/5c11a9b9-34de-42f7-9399-3cddce1af7c9-kube-api-access-c9lbv\") pod \"ovsdbserver-sb-0\" (UID: \"5c11a9b9-34de-42f7-9399-3cddce1af7c9\") " pod="openstack/ovsdbserver-sb-0" Dec 07 16:17:39 crc kubenswrapper[4716]: I1207 16:17:39.823563 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c11a9b9-34de-42f7-9399-3cddce1af7c9-config\") pod \"ovsdbserver-sb-0\" (UID: \"5c11a9b9-34de-42f7-9399-3cddce1af7c9\") " pod="openstack/ovsdbserver-sb-0" Dec 07 16:17:39 crc kubenswrapper[4716]: I1207 16:17:39.823580 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c11a9b9-34de-42f7-9399-3cddce1af7c9-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"5c11a9b9-34de-42f7-9399-3cddce1af7c9\") " pod="openstack/ovsdbserver-sb-0" Dec 07 16:17:39 crc kubenswrapper[4716]: I1207 16:17:39.823605 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5c11a9b9-34de-42f7-9399-3cddce1af7c9-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"5c11a9b9-34de-42f7-9399-3cddce1af7c9\") " pod="openstack/ovsdbserver-sb-0" Dec 07 16:17:39 crc kubenswrapper[4716]: I1207 16:17:39.823625 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c11a9b9-34de-42f7-9399-3cddce1af7c9-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"5c11a9b9-34de-42f7-9399-3cddce1af7c9\") " pod="openstack/ovsdbserver-sb-0" Dec 07 16:17:39 crc kubenswrapper[4716]: I1207 16:17:39.823673 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c11a9b9-34de-42f7-9399-3cddce1af7c9-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"5c11a9b9-34de-42f7-9399-3cddce1af7c9\") " pod="openstack/ovsdbserver-sb-0" Dec 07 16:17:39 crc kubenswrapper[4716]: I1207 16:17:39.925263 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-sb-0\" (UID: \"5c11a9b9-34de-42f7-9399-3cddce1af7c9\") " pod="openstack/ovsdbserver-sb-0" Dec 07 16:17:39 crc kubenswrapper[4716]: I1207 16:17:39.925319 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/5c11a9b9-34de-42f7-9399-3cddce1af7c9-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"5c11a9b9-34de-42f7-9399-3cddce1af7c9\") " pod="openstack/ovsdbserver-sb-0" Dec 07 16:17:39 crc kubenswrapper[4716]: I1207 16:17:39.925356 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c9lbv\" (UniqueName: \"kubernetes.io/projected/5c11a9b9-34de-42f7-9399-3cddce1af7c9-kube-api-access-c9lbv\") pod \"ovsdbserver-sb-0\" (UID: \"5c11a9b9-34de-42f7-9399-3cddce1af7c9\") " pod="openstack/ovsdbserver-sb-0" Dec 07 16:17:39 crc kubenswrapper[4716]: I1207 16:17:39.925404 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c11a9b9-34de-42f7-9399-3cddce1af7c9-config\") pod \"ovsdbserver-sb-0\" (UID: \"5c11a9b9-34de-42f7-9399-3cddce1af7c9\") " pod="openstack/ovsdbserver-sb-0" Dec 07 16:17:39 crc kubenswrapper[4716]: I1207 16:17:39.925723 4716 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-sb-0\" (UID: \"5c11a9b9-34de-42f7-9399-3cddce1af7c9\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/ovsdbserver-sb-0" Dec 07 16:17:39 crc kubenswrapper[4716]: I1207 16:17:39.925424 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c11a9b9-34de-42f7-9399-3cddce1af7c9-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"5c11a9b9-34de-42f7-9399-3cddce1af7c9\") " pod="openstack/ovsdbserver-sb-0" Dec 07 16:17:39 crc kubenswrapper[4716]: I1207 16:17:39.926088 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5c11a9b9-34de-42f7-9399-3cddce1af7c9-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"5c11a9b9-34de-42f7-9399-3cddce1af7c9\") " pod="openstack/ovsdbserver-sb-0" Dec 07 16:17:39 crc kubenswrapper[4716]: I1207 16:17:39.926108 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c11a9b9-34de-42f7-9399-3cddce1af7c9-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"5c11a9b9-34de-42f7-9399-3cddce1af7c9\") " pod="openstack/ovsdbserver-sb-0" Dec 07 16:17:39 crc kubenswrapper[4716]: I1207 16:17:39.926131 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/5c11a9b9-34de-42f7-9399-3cddce1af7c9-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"5c11a9b9-34de-42f7-9399-3cddce1af7c9\") " pod="openstack/ovsdbserver-sb-0" Dec 07 16:17:39 crc kubenswrapper[4716]: I1207 16:17:39.926146 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c11a9b9-34de-42f7-9399-3cddce1af7c9-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"5c11a9b9-34de-42f7-9399-3cddce1af7c9\") " pod="openstack/ovsdbserver-sb-0" Dec 07 16:17:39 crc kubenswrapper[4716]: I1207 16:17:39.927156 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5c11a9b9-34de-42f7-9399-3cddce1af7c9-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"5c11a9b9-34de-42f7-9399-3cddce1af7c9\") " pod="openstack/ovsdbserver-sb-0" Dec 07 16:17:39 crc kubenswrapper[4716]: I1207 16:17:39.927295 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c11a9b9-34de-42f7-9399-3cddce1af7c9-config\") pod \"ovsdbserver-sb-0\" (UID: \"5c11a9b9-34de-42f7-9399-3cddce1af7c9\") " pod="openstack/ovsdbserver-sb-0" Dec 07 16:17:39 crc kubenswrapper[4716]: I1207 16:17:39.933443 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c11a9b9-34de-42f7-9399-3cddce1af7c9-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"5c11a9b9-34de-42f7-9399-3cddce1af7c9\") " pod="openstack/ovsdbserver-sb-0" Dec 07 16:17:39 crc kubenswrapper[4716]: I1207 16:17:39.934037 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c11a9b9-34de-42f7-9399-3cddce1af7c9-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"5c11a9b9-34de-42f7-9399-3cddce1af7c9\") " pod="openstack/ovsdbserver-sb-0" Dec 07 16:17:39 crc kubenswrapper[4716]: I1207 16:17:39.934301 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c11a9b9-34de-42f7-9399-3cddce1af7c9-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"5c11a9b9-34de-42f7-9399-3cddce1af7c9\") " pod="openstack/ovsdbserver-sb-0" Dec 07 16:17:39 crc kubenswrapper[4716]: I1207 16:17:39.942547 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c9lbv\" (UniqueName: \"kubernetes.io/projected/5c11a9b9-34de-42f7-9399-3cddce1af7c9-kube-api-access-c9lbv\") pod \"ovsdbserver-sb-0\" (UID: \"5c11a9b9-34de-42f7-9399-3cddce1af7c9\") " pod="openstack/ovsdbserver-sb-0" Dec 07 16:17:39 crc kubenswrapper[4716]: I1207 16:17:39.955385 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-sb-0\" (UID: \"5c11a9b9-34de-42f7-9399-3cddce1af7c9\") " pod="openstack/ovsdbserver-sb-0" Dec 07 16:17:40 crc kubenswrapper[4716]: I1207 16:17:40.092972 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 07 16:17:47 crc kubenswrapper[4716]: E1207 16:17:47.613808 4716 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-memcached:current-podified" Dec 07 16:17:47 crc kubenswrapper[4716]: E1207 16:17:47.614623 4716 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:memcached,Image:quay.io/podified-antelope-centos9/openstack-memcached:current-podified,Command:[/usr/bin/dumb-init -- /usr/local/bin/kolla_start],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:memcached,HostPort:0,ContainerPort:11211,Protocol:TCP,HostIP:,},ContainerPort{Name:memcached-tls,HostPort:0,ContainerPort:11212,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:POD_IPS,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIPs,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:CONFIG_HASH,Value:nbfhdh545h5fdh656h66bh58chb6hbbh577hb8h579h78h558h557hd9hc4h667h85hdfh5b4h8dh55ch5fbhcfh98h5cfhf4hfdh64hddh5b9q,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/src,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:memcached-tls-certs,ReadOnly:true,MountPath:/var/lib/config-data/tls/certs/memcached.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:memcached-tls-certs,ReadOnly:true,MountPath:/var/lib/config-data/tls/private/memcached.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xckdg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 11211 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 11211 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42457,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42457,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod memcached-0_openstack(0f4b5de9-9760-4930-bc8e-d7e47a13ad2e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 07 16:17:47 crc kubenswrapper[4716]: I1207 16:17:47.614815 4716 scope.go:117] "RemoveContainer" containerID="4f400069cb58c65bdc2c1a9f100546822c743a27bdf17cfdbba425cea44b7d4a" Dec 07 16:17:47 crc kubenswrapper[4716]: E1207 16:17:47.615709 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"memcached\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/memcached-0" podUID="0f4b5de9-9760-4930-bc8e-d7e47a13ad2e" Dec 07 16:17:47 crc kubenswrapper[4716]: E1207 16:17:47.736272 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"memcached\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-memcached:current-podified\\\"\"" pod="openstack/memcached-0" podUID="0f4b5de9-9760-4930-bc8e-d7e47a13ad2e" Dec 07 16:17:48 crc kubenswrapper[4716]: I1207 16:17:48.048791 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 07 16:17:48 crc kubenswrapper[4716]: W1207 16:17:48.629499 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc088cd1b_83a8_4f05_b7db_11c8ffb59fab.slice/crio-bb8e5f62eece4be6882fee91108453b363425e0c0ff925e89657243b1889ca2a WatchSource:0}: Error finding container bb8e5f62eece4be6882fee91108453b363425e0c0ff925e89657243b1889ca2a: Status 404 returned error can't find the container with id bb8e5f62eece4be6882fee91108453b363425e0c0ff925e89657243b1889ca2a Dec 07 16:17:48 crc kubenswrapper[4716]: E1207 16:17:48.655269 4716 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 07 16:17:48 crc kubenswrapper[4716]: E1207 16:17:48.655456 4716 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9258x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-n7br7_openstack(81351bb9-bf61-4c75-8a31-41769f3c09bf): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 07 16:17:48 crc kubenswrapper[4716]: E1207 16:17:48.656737 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-n7br7" podUID="81351bb9-bf61-4c75-8a31-41769f3c09bf" Dec 07 16:17:48 crc kubenswrapper[4716]: E1207 16:17:48.664296 4716 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 07 16:17:48 crc kubenswrapper[4716]: E1207 16:17:48.664470 4716 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-h79wx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-rjq6f_openstack(7e9041a0-0258-4e21-b0a0-07edb6549139): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 07 16:17:48 crc kubenswrapper[4716]: E1207 16:17:48.666595 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-rjq6f" podUID="7e9041a0-0258-4e21-b0a0-07edb6549139" Dec 07 16:17:48 crc kubenswrapper[4716]: I1207 16:17:48.789784 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"c088cd1b-83a8-4f05-b7db-11c8ffb59fab","Type":"ContainerStarted","Data":"bb8e5f62eece4be6882fee91108453b363425e0c0ff925e89657243b1889ca2a"} Dec 07 16:17:49 crc kubenswrapper[4716]: I1207 16:17:49.087472 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8phkg"] Dec 07 16:17:49 crc kubenswrapper[4716]: W1207 16:17:49.102194 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfef1e486_acca_4cc4_b2b1_3b20270d418a.slice/crio-fa224d91f9560174d80dbe06302487f4a62c5c3b531a4fe0fca57b4953e7abae WatchSource:0}: Error finding container fa224d91f9560174d80dbe06302487f4a62c5c3b531a4fe0fca57b4953e7abae: Status 404 returned error can't find the container with id fa224d91f9560174d80dbe06302487f4a62c5c3b531a4fe0fca57b4953e7abae Dec 07 16:17:49 crc kubenswrapper[4716]: I1207 16:17:49.314135 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 07 16:17:49 crc kubenswrapper[4716]: W1207 16:17:49.397489 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod771a3c3e_34e4_43f6_a3e2_8e7156e081d1.slice/crio-63ef1f6570fba867b35772905659f27828bd42edf9f1c7047a5adaf562213c42 WatchSource:0}: Error finding container 63ef1f6570fba867b35772905659f27828bd42edf9f1c7047a5adaf562213c42: Status 404 returned error can't find the container with id 63ef1f6570fba867b35772905659f27828bd42edf9f1c7047a5adaf562213c42 Dec 07 16:17:49 crc kubenswrapper[4716]: I1207 16:17:49.398730 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-tgl27"] Dec 07 16:17:49 crc kubenswrapper[4716]: I1207 16:17:49.412768 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-rjq6f" Dec 07 16:17:49 crc kubenswrapper[4716]: I1207 16:17:49.425840 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-n7br7" Dec 07 16:17:49 crc kubenswrapper[4716]: I1207 16:17:49.536773 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7e9041a0-0258-4e21-b0a0-07edb6549139-dns-svc\") pod \"7e9041a0-0258-4e21-b0a0-07edb6549139\" (UID: \"7e9041a0-0258-4e21-b0a0-07edb6549139\") " Dec 07 16:17:49 crc kubenswrapper[4716]: I1207 16:17:49.537191 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81351bb9-bf61-4c75-8a31-41769f3c09bf-config\") pod \"81351bb9-bf61-4c75-8a31-41769f3c09bf\" (UID: \"81351bb9-bf61-4c75-8a31-41769f3c09bf\") " Dec 07 16:17:49 crc kubenswrapper[4716]: I1207 16:17:49.537266 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9258x\" (UniqueName: \"kubernetes.io/projected/81351bb9-bf61-4c75-8a31-41769f3c09bf-kube-api-access-9258x\") pod \"81351bb9-bf61-4c75-8a31-41769f3c09bf\" (UID: \"81351bb9-bf61-4c75-8a31-41769f3c09bf\") " Dec 07 16:17:49 crc kubenswrapper[4716]: I1207 16:17:49.537350 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h79wx\" (UniqueName: \"kubernetes.io/projected/7e9041a0-0258-4e21-b0a0-07edb6549139-kube-api-access-h79wx\") pod \"7e9041a0-0258-4e21-b0a0-07edb6549139\" (UID: \"7e9041a0-0258-4e21-b0a0-07edb6549139\") " Dec 07 16:17:49 crc kubenswrapper[4716]: I1207 16:17:49.537390 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e9041a0-0258-4e21-b0a0-07edb6549139-config\") pod \"7e9041a0-0258-4e21-b0a0-07edb6549139\" (UID: \"7e9041a0-0258-4e21-b0a0-07edb6549139\") " Dec 07 16:17:49 crc kubenswrapper[4716]: I1207 16:17:49.537381 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e9041a0-0258-4e21-b0a0-07edb6549139-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7e9041a0-0258-4e21-b0a0-07edb6549139" (UID: "7e9041a0-0258-4e21-b0a0-07edb6549139"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:17:49 crc kubenswrapper[4716]: I1207 16:17:49.537776 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/81351bb9-bf61-4c75-8a31-41769f3c09bf-config" (OuterVolumeSpecName: "config") pod "81351bb9-bf61-4c75-8a31-41769f3c09bf" (UID: "81351bb9-bf61-4c75-8a31-41769f3c09bf"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:17:49 crc kubenswrapper[4716]: I1207 16:17:49.537859 4716 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7e9041a0-0258-4e21-b0a0-07edb6549139-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 07 16:17:49 crc kubenswrapper[4716]: I1207 16:17:49.538460 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e9041a0-0258-4e21-b0a0-07edb6549139-config" (OuterVolumeSpecName: "config") pod "7e9041a0-0258-4e21-b0a0-07edb6549139" (UID: "7e9041a0-0258-4e21-b0a0-07edb6549139"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:17:49 crc kubenswrapper[4716]: I1207 16:17:49.544281 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81351bb9-bf61-4c75-8a31-41769f3c09bf-kube-api-access-9258x" (OuterVolumeSpecName: "kube-api-access-9258x") pod "81351bb9-bf61-4c75-8a31-41769f3c09bf" (UID: "81351bb9-bf61-4c75-8a31-41769f3c09bf"). InnerVolumeSpecName "kube-api-access-9258x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:17:49 crc kubenswrapper[4716]: I1207 16:17:49.544421 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e9041a0-0258-4e21-b0a0-07edb6549139-kube-api-access-h79wx" (OuterVolumeSpecName: "kube-api-access-h79wx") pod "7e9041a0-0258-4e21-b0a0-07edb6549139" (UID: "7e9041a0-0258-4e21-b0a0-07edb6549139"). InnerVolumeSpecName "kube-api-access-h79wx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:17:49 crc kubenswrapper[4716]: I1207 16:17:49.601386 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 07 16:17:49 crc kubenswrapper[4716]: I1207 16:17:49.638988 4716 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81351bb9-bf61-4c75-8a31-41769f3c09bf-config\") on node \"crc\" DevicePath \"\"" Dec 07 16:17:49 crc kubenswrapper[4716]: I1207 16:17:49.639032 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9258x\" (UniqueName: \"kubernetes.io/projected/81351bb9-bf61-4c75-8a31-41769f3c09bf-kube-api-access-9258x\") on node \"crc\" DevicePath \"\"" Dec 07 16:17:49 crc kubenswrapper[4716]: I1207 16:17:49.639043 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h79wx\" (UniqueName: \"kubernetes.io/projected/7e9041a0-0258-4e21-b0a0-07edb6549139-kube-api-access-h79wx\") on node \"crc\" DevicePath \"\"" Dec 07 16:17:49 crc kubenswrapper[4716]: I1207 16:17:49.639056 4716 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e9041a0-0258-4e21-b0a0-07edb6549139-config\") on node \"crc\" DevicePath \"\"" Dec 07 16:17:49 crc kubenswrapper[4716]: I1207 16:17:49.804699 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-n7br7" event={"ID":"81351bb9-bf61-4c75-8a31-41769f3c09bf","Type":"ContainerDied","Data":"8e65c04758f9dcfd7a9d71a96b034416c28e4aea2a6d20b2c01413c17912c587"} Dec 07 16:17:49 crc kubenswrapper[4716]: I1207 16:17:49.804742 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-n7br7" Dec 07 16:17:49 crc kubenswrapper[4716]: I1207 16:17:49.806819 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-tgl27" event={"ID":"399984bf-94af-4630-a7c3-9375a388159a","Type":"ContainerStarted","Data":"ee1f99c1e6e16e454014e34b9b94601b206a246e7e1293cea525ee156427936b"} Dec 07 16:17:49 crc kubenswrapper[4716]: I1207 16:17:49.808771 4716 generic.go:334] "Generic (PLEG): container finished" podID="240ebe8a-eb59-4726-a66f-4ce1e2900e4c" containerID="ed81504e8f9d8db9e2c7703a8dc46d342a761904bb7ad96a3c70efde8df007f8" exitCode=0 Dec 07 16:17:49 crc kubenswrapper[4716]: I1207 16:17:49.808935 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-ktd8n" event={"ID":"240ebe8a-eb59-4726-a66f-4ce1e2900e4c","Type":"ContainerDied","Data":"ed81504e8f9d8db9e2c7703a8dc46d342a761904bb7ad96a3c70efde8df007f8"} Dec 07 16:17:49 crc kubenswrapper[4716]: I1207 16:17:49.812790 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"f5bcaa65-381c-40e7-97bd-4fc0eafc572a","Type":"ContainerStarted","Data":"0e3dcb20fb6d192ecc8827bcf14b4510e6752b48c8abed425637efed55b553a2"} Dec 07 16:17:49 crc kubenswrapper[4716]: I1207 16:17:49.822635 4716 generic.go:334] "Generic (PLEG): container finished" podID="b4f34fda-3133-46e3-99d6-d49f0ae032d5" containerID="a19d8f171cbaee1ba0e8568864d5f7b86ce1134b769dd03dc3eac12098f1ce13" exitCode=0 Dec 07 16:17:49 crc kubenswrapper[4716]: I1207 16:17:49.822754 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-dwp2b" event={"ID":"b4f34fda-3133-46e3-99d6-d49f0ae032d5","Type":"ContainerDied","Data":"a19d8f171cbaee1ba0e8568864d5f7b86ce1134b769dd03dc3eac12098f1ce13"} Dec 07 16:17:49 crc kubenswrapper[4716]: I1207 16:17:49.824599 4716 generic.go:334] "Generic (PLEG): container finished" podID="fef1e486-acca-4cc4-b2b1-3b20270d418a" containerID="1fd1e50c076ddfcac3629fa3995fb943661bf547ba40904201364e1f2d459134" exitCode=0 Dec 07 16:17:49 crc kubenswrapper[4716]: I1207 16:17:49.824748 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8phkg" event={"ID":"fef1e486-acca-4cc4-b2b1-3b20270d418a","Type":"ContainerDied","Data":"1fd1e50c076ddfcac3629fa3995fb943661bf547ba40904201364e1f2d459134"} Dec 07 16:17:49 crc kubenswrapper[4716]: I1207 16:17:49.824855 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8phkg" event={"ID":"fef1e486-acca-4cc4-b2b1-3b20270d418a","Type":"ContainerStarted","Data":"fa224d91f9560174d80dbe06302487f4a62c5c3b531a4fe0fca57b4953e7abae"} Dec 07 16:17:49 crc kubenswrapper[4716]: I1207 16:17:49.826399 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"c450ef40-3a23-4bab-9d64-4475f12d0490","Type":"ContainerStarted","Data":"95846aaf937c0d79a3f2b350cffb908f16a9cb60a3f248fea62ec2508774e87d"} Dec 07 16:17:49 crc kubenswrapper[4716]: I1207 16:17:49.831419 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-rjq6f" event={"ID":"7e9041a0-0258-4e21-b0a0-07edb6549139","Type":"ContainerDied","Data":"6cf210dcb88d6dfbf796d41105cc4c88595ad110329ba0146984755d1ee60b31"} Dec 07 16:17:49 crc kubenswrapper[4716]: I1207 16:17:49.831452 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-rjq6f" Dec 07 16:17:49 crc kubenswrapper[4716]: I1207 16:17:49.832734 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"771a3c3e-34e4-43f6-a3e2-8e7156e081d1","Type":"ContainerStarted","Data":"63ef1f6570fba867b35772905659f27828bd42edf9f1c7047a5adaf562213c42"} Dec 07 16:17:49 crc kubenswrapper[4716]: I1207 16:17:49.951144 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-n7br7"] Dec 07 16:17:49 crc kubenswrapper[4716]: I1207 16:17:49.959058 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-n7br7"] Dec 07 16:17:49 crc kubenswrapper[4716]: I1207 16:17:49.995286 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-rjq6f"] Dec 07 16:17:49 crc kubenswrapper[4716]: I1207 16:17:49.999928 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-rjq6f"] Dec 07 16:17:50 crc kubenswrapper[4716]: I1207 16:17:50.163346 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-g6nwl"] Dec 07 16:17:50 crc kubenswrapper[4716]: I1207 16:17:50.849775 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe","Type":"ContainerStarted","Data":"dec9f36ef119ff579ca865b84eed4ecbc4e8e9b8f0a6ad017672916770251362"} Dec 07 16:17:53 crc kubenswrapper[4716]: I1207 16:17:51.665993 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e9041a0-0258-4e21-b0a0-07edb6549139" path="/var/lib/kubelet/pods/7e9041a0-0258-4e21-b0a0-07edb6549139/volumes" Dec 07 16:17:53 crc kubenswrapper[4716]: I1207 16:17:51.666391 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81351bb9-bf61-4c75-8a31-41769f3c09bf" path="/var/lib/kubelet/pods/81351bb9-bf61-4c75-8a31-41769f3c09bf/volumes" Dec 07 16:17:53 crc kubenswrapper[4716]: W1207 16:17:53.332713 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0fc0f0b8_b60a_4aff_84dd_bbec37b12d81.slice/crio-b4cefd0af927b436846dd35e8379a743bbffcc3e5cd0f69fb347904ee1c52857 WatchSource:0}: Error finding container b4cefd0af927b436846dd35e8379a743bbffcc3e5cd0f69fb347904ee1c52857: Status 404 returned error can't find the container with id b4cefd0af927b436846dd35e8379a743bbffcc3e5cd0f69fb347904ee1c52857 Dec 07 16:17:53 crc kubenswrapper[4716]: I1207 16:17:53.879478 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f653df87-c234-4cb6-8ee8-6b16b4fe0044","Type":"ContainerStarted","Data":"c4250e9cb4d0c6bd1c2d3e3bc3ca8c43b9a53e5cd186b5bd154837621d4a55f1"} Dec 07 16:17:53 crc kubenswrapper[4716]: I1207 16:17:53.886811 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"5c11a9b9-34de-42f7-9399-3cddce1af7c9","Type":"ContainerStarted","Data":"e78473ff37a5c1f0b3a1e2c2fabb60422785da4630aed6893a7cb9b76f5b1c43"} Dec 07 16:17:53 crc kubenswrapper[4716]: I1207 16:17:53.890479 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-666b6646f7-dwp2b" Dec 07 16:17:53 crc kubenswrapper[4716]: I1207 16:17:53.892470 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-g6nwl" event={"ID":"0fc0f0b8-b60a-4aff-84dd-bbec37b12d81","Type":"ContainerStarted","Data":"b4cefd0af927b436846dd35e8379a743bbffcc3e5cd0f69fb347904ee1c52857"} Dec 07 16:17:53 crc kubenswrapper[4716]: I1207 16:17:53.972254 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-666b6646f7-dwp2b" podStartSLOduration=7.039451646 podStartE2EDuration="28.972232306s" podCreationTimestamp="2025-12-07 16:17:25 +0000 UTC" firstStartedPulling="2025-12-07 16:17:26.833530642 +0000 UTC m=+909.523815554" lastFinishedPulling="2025-12-07 16:17:48.766311302 +0000 UTC m=+931.456596214" observedRunningTime="2025-12-07 16:17:53.966052713 +0000 UTC m=+936.656337635" watchObservedRunningTime="2025-12-07 16:17:53.972232306 +0000 UTC m=+936.662517228" Dec 07 16:17:54 crc kubenswrapper[4716]: I1207 16:17:54.901783 4716 generic.go:334] "Generic (PLEG): container finished" podID="fef1e486-acca-4cc4-b2b1-3b20270d418a" containerID="22ee7211c116eb84a19b1d9da6c8e22a81b40c2415e1456a23b462893c1c4dc4" exitCode=0 Dec 07 16:17:54 crc kubenswrapper[4716]: I1207 16:17:54.901865 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8phkg" event={"ID":"fef1e486-acca-4cc4-b2b1-3b20270d418a","Type":"ContainerDied","Data":"22ee7211c116eb84a19b1d9da6c8e22a81b40c2415e1456a23b462893c1c4dc4"} Dec 07 16:17:54 crc kubenswrapper[4716]: I1207 16:17:54.905046 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-ktd8n" event={"ID":"240ebe8a-eb59-4726-a66f-4ce1e2900e4c","Type":"ContainerStarted","Data":"7bd18f80e3166b1c08642a65ba11421da6ac8e3311a8c296181cfabdf597a7cd"} Dec 07 16:17:54 crc kubenswrapper[4716]: I1207 16:17:54.905127 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-ktd8n" Dec 07 16:17:54 crc kubenswrapper[4716]: I1207 16:17:54.909417 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-dwp2b" event={"ID":"b4f34fda-3133-46e3-99d6-d49f0ae032d5","Type":"ContainerStarted","Data":"f73a048dd2a4ed04fd2f0093add7ef4b8399c0526e4795314492d284d756017f"} Dec 07 16:17:54 crc kubenswrapper[4716]: I1207 16:17:54.937328 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-ktd8n" podStartSLOduration=7.196720227 podStartE2EDuration="28.937311784s" podCreationTimestamp="2025-12-07 16:17:26 +0000 UTC" firstStartedPulling="2025-12-07 16:17:27.119223002 +0000 UTC m=+909.809507914" lastFinishedPulling="2025-12-07 16:17:48.859814559 +0000 UTC m=+931.550099471" observedRunningTime="2025-12-07 16:17:54.936414059 +0000 UTC m=+937.626698971" watchObservedRunningTime="2025-12-07 16:17:54.937311784 +0000 UTC m=+937.627596696" Dec 07 16:17:56 crc kubenswrapper[4716]: I1207 16:17:56.922857 4716 generic.go:334] "Generic (PLEG): container finished" podID="f5bcaa65-381c-40e7-97bd-4fc0eafc572a" containerID="0e3dcb20fb6d192ecc8827bcf14b4510e6752b48c8abed425637efed55b553a2" exitCode=0 Dec 07 16:17:56 crc kubenswrapper[4716]: I1207 16:17:56.922955 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"f5bcaa65-381c-40e7-97bd-4fc0eafc572a","Type":"ContainerDied","Data":"0e3dcb20fb6d192ecc8827bcf14b4510e6752b48c8abed425637efed55b553a2"} Dec 07 16:17:57 crc kubenswrapper[4716]: I1207 16:17:57.935508 4716 generic.go:334] "Generic (PLEG): container finished" podID="c450ef40-3a23-4bab-9d64-4475f12d0490" containerID="95846aaf937c0d79a3f2b350cffb908f16a9cb60a3f248fea62ec2508774e87d" exitCode=0 Dec 07 16:17:57 crc kubenswrapper[4716]: I1207 16:17:57.935582 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"c450ef40-3a23-4bab-9d64-4475f12d0490","Type":"ContainerDied","Data":"95846aaf937c0d79a3f2b350cffb908f16a9cb60a3f248fea62ec2508774e87d"} Dec 07 16:17:57 crc kubenswrapper[4716]: I1207 16:17:57.942567 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"f5bcaa65-381c-40e7-97bd-4fc0eafc572a","Type":"ContainerStarted","Data":"3506aa2e42adc63bce9f926c9bfc8bccdb8ad1a7059823596eac31ea1d88eb13"} Dec 07 16:17:57 crc kubenswrapper[4716]: I1207 16:17:57.947281 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"771a3c3e-34e4-43f6-a3e2-8e7156e081d1","Type":"ContainerStarted","Data":"423934aa979d26e9d582850d27a2c4818bcd59b44ae3983585789d2569b13e85"} Dec 07 16:17:57 crc kubenswrapper[4716]: I1207 16:17:57.953347 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8phkg" event={"ID":"fef1e486-acca-4cc4-b2b1-3b20270d418a","Type":"ContainerStarted","Data":"53639bea8fec33f2cf8928b14b9c0d9e9e956716b32781f207c37b1f9e3054b1"} Dec 07 16:17:57 crc kubenswrapper[4716]: I1207 16:17:57.983921 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=11.033935564 podStartE2EDuration="28.98390599s" podCreationTimestamp="2025-12-07 16:17:29 +0000 UTC" firstStartedPulling="2025-12-07 16:17:30.753932876 +0000 UTC m=+913.444217788" lastFinishedPulling="2025-12-07 16:17:48.703903302 +0000 UTC m=+931.394188214" observedRunningTime="2025-12-07 16:17:57.98140089 +0000 UTC m=+940.671685802" watchObservedRunningTime="2025-12-07 16:17:57.98390599 +0000 UTC m=+940.674190902" Dec 07 16:17:58 crc kubenswrapper[4716]: I1207 16:17:58.003739 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8phkg" podStartSLOduration=19.614659036 podStartE2EDuration="24.003723133s" podCreationTimestamp="2025-12-07 16:17:34 +0000 UTC" firstStartedPulling="2025-12-07 16:17:53.323196719 +0000 UTC m=+936.013481631" lastFinishedPulling="2025-12-07 16:17:57.712260816 +0000 UTC m=+940.402545728" observedRunningTime="2025-12-07 16:17:58.002095497 +0000 UTC m=+940.692380419" watchObservedRunningTime="2025-12-07 16:17:58.003723133 +0000 UTC m=+940.694008045" Dec 07 16:17:58 crc kubenswrapper[4716]: I1207 16:17:58.961995 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"c450ef40-3a23-4bab-9d64-4475f12d0490","Type":"ContainerStarted","Data":"2eb63db55440339314c12b54bc9dc5cfdee8ba3f6847df718cf07c94ae00f8eb"} Dec 07 16:17:58 crc kubenswrapper[4716]: I1207 16:17:58.964333 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"c088cd1b-83a8-4f05-b7db-11c8ffb59fab","Type":"ContainerStarted","Data":"5695d3e0ddfd3193c3f76ef6bfbe23181c8fd9bc6f004904a15d8d152feadcd5"} Dec 07 16:17:58 crc kubenswrapper[4716]: I1207 16:17:58.964410 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 07 16:17:58 crc kubenswrapper[4716]: I1207 16:17:58.965733 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"5c11a9b9-34de-42f7-9399-3cddce1af7c9","Type":"ContainerStarted","Data":"4226a0a65c486a99b712f6a2b568b1d1d39c2fb92cbc8a78d9b88f5e18f38226"} Dec 07 16:17:58 crc kubenswrapper[4716]: I1207 16:17:58.967376 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-tgl27" event={"ID":"399984bf-94af-4630-a7c3-9375a388159a","Type":"ContainerStarted","Data":"95a1c793111a82bf7a1f9a4f900828c2951044e575a2e3129ae35d80c0bd3ebf"} Dec 07 16:17:58 crc kubenswrapper[4716]: I1207 16:17:58.967764 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-tgl27" Dec 07 16:17:58 crc kubenswrapper[4716]: I1207 16:17:58.968923 4716 generic.go:334] "Generic (PLEG): container finished" podID="0fc0f0b8-b60a-4aff-84dd-bbec37b12d81" containerID="7ecdd877d56da5feffeb0f29c08ba96a788a6857d2e4d71378dca96abf1f9256" exitCode=0 Dec 07 16:17:58 crc kubenswrapper[4716]: I1207 16:17:58.969709 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-g6nwl" event={"ID":"0fc0f0b8-b60a-4aff-84dd-bbec37b12d81","Type":"ContainerDied","Data":"7ecdd877d56da5feffeb0f29c08ba96a788a6857d2e4d71378dca96abf1f9256"} Dec 07 16:17:58 crc kubenswrapper[4716]: I1207 16:17:58.983628 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=13.351942821 podStartE2EDuration="31.983604834s" podCreationTimestamp="2025-12-07 16:17:27 +0000 UTC" firstStartedPulling="2025-12-07 16:17:30.115842045 +0000 UTC m=+912.806126957" lastFinishedPulling="2025-12-07 16:17:48.747504058 +0000 UTC m=+931.437788970" observedRunningTime="2025-12-07 16:17:58.9784434 +0000 UTC m=+941.668728312" watchObservedRunningTime="2025-12-07 16:17:58.983604834 +0000 UTC m=+941.673889746" Dec 07 16:17:59 crc kubenswrapper[4716]: I1207 16:17:59.021259 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=17.909428085000002 podStartE2EDuration="27.021237353s" podCreationTimestamp="2025-12-07 16:17:32 +0000 UTC" firstStartedPulling="2025-12-07 16:17:48.660848062 +0000 UTC m=+931.351132974" lastFinishedPulling="2025-12-07 16:17:57.77265731 +0000 UTC m=+940.462942242" observedRunningTime="2025-12-07 16:17:59.017567631 +0000 UTC m=+941.707852553" watchObservedRunningTime="2025-12-07 16:17:59.021237353 +0000 UTC m=+941.711522265" Dec 07 16:17:59 crc kubenswrapper[4716]: I1207 16:17:59.039149 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-tgl27" podStartSLOduration=15.694286568999999 podStartE2EDuration="24.039135513s" podCreationTimestamp="2025-12-07 16:17:35 +0000 UTC" firstStartedPulling="2025-12-07 16:17:49.40353869 +0000 UTC m=+932.093823592" lastFinishedPulling="2025-12-07 16:17:57.748387624 +0000 UTC m=+940.438672536" observedRunningTime="2025-12-07 16:17:59.034724769 +0000 UTC m=+941.725009681" watchObservedRunningTime="2025-12-07 16:17:59.039135513 +0000 UTC m=+941.729420425" Dec 07 16:17:59 crc kubenswrapper[4716]: I1207 16:17:59.119272 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Dec 07 16:17:59 crc kubenswrapper[4716]: I1207 16:17:59.119336 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Dec 07 16:17:59 crc kubenswrapper[4716]: I1207 16:17:59.155151 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-dnnsf"] Dec 07 16:17:59 crc kubenswrapper[4716]: I1207 16:17:59.156827 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dnnsf" Dec 07 16:17:59 crc kubenswrapper[4716]: I1207 16:17:59.160519 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dnnsf"] Dec 07 16:17:59 crc kubenswrapper[4716]: I1207 16:17:59.295013 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b870e60-acde-4d14-be71-bee41e0c53bd-utilities\") pod \"community-operators-dnnsf\" (UID: \"9b870e60-acde-4d14-be71-bee41e0c53bd\") " pod="openshift-marketplace/community-operators-dnnsf" Dec 07 16:17:59 crc kubenswrapper[4716]: I1207 16:17:59.295134 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b870e60-acde-4d14-be71-bee41e0c53bd-catalog-content\") pod \"community-operators-dnnsf\" (UID: \"9b870e60-acde-4d14-be71-bee41e0c53bd\") " pod="openshift-marketplace/community-operators-dnnsf" Dec 07 16:17:59 crc kubenswrapper[4716]: I1207 16:17:59.295185 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zx7cc\" (UniqueName: \"kubernetes.io/projected/9b870e60-acde-4d14-be71-bee41e0c53bd-kube-api-access-zx7cc\") pod \"community-operators-dnnsf\" (UID: \"9b870e60-acde-4d14-be71-bee41e0c53bd\") " pod="openshift-marketplace/community-operators-dnnsf" Dec 07 16:17:59 crc kubenswrapper[4716]: I1207 16:17:59.396531 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b870e60-acde-4d14-be71-bee41e0c53bd-utilities\") pod \"community-operators-dnnsf\" (UID: \"9b870e60-acde-4d14-be71-bee41e0c53bd\") " pod="openshift-marketplace/community-operators-dnnsf" Dec 07 16:17:59 crc kubenswrapper[4716]: I1207 16:17:59.396894 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b870e60-acde-4d14-be71-bee41e0c53bd-catalog-content\") pod \"community-operators-dnnsf\" (UID: \"9b870e60-acde-4d14-be71-bee41e0c53bd\") " pod="openshift-marketplace/community-operators-dnnsf" Dec 07 16:17:59 crc kubenswrapper[4716]: I1207 16:17:59.396939 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zx7cc\" (UniqueName: \"kubernetes.io/projected/9b870e60-acde-4d14-be71-bee41e0c53bd-kube-api-access-zx7cc\") pod \"community-operators-dnnsf\" (UID: \"9b870e60-acde-4d14-be71-bee41e0c53bd\") " pod="openshift-marketplace/community-operators-dnnsf" Dec 07 16:17:59 crc kubenswrapper[4716]: I1207 16:17:59.397008 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b870e60-acde-4d14-be71-bee41e0c53bd-utilities\") pod \"community-operators-dnnsf\" (UID: \"9b870e60-acde-4d14-be71-bee41e0c53bd\") " pod="openshift-marketplace/community-operators-dnnsf" Dec 07 16:17:59 crc kubenswrapper[4716]: I1207 16:17:59.397258 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b870e60-acde-4d14-be71-bee41e0c53bd-catalog-content\") pod \"community-operators-dnnsf\" (UID: \"9b870e60-acde-4d14-be71-bee41e0c53bd\") " pod="openshift-marketplace/community-operators-dnnsf" Dec 07 16:17:59 crc kubenswrapper[4716]: I1207 16:17:59.416822 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zx7cc\" (UniqueName: \"kubernetes.io/projected/9b870e60-acde-4d14-be71-bee41e0c53bd-kube-api-access-zx7cc\") pod \"community-operators-dnnsf\" (UID: \"9b870e60-acde-4d14-be71-bee41e0c53bd\") " pod="openshift-marketplace/community-operators-dnnsf" Dec 07 16:17:59 crc kubenswrapper[4716]: I1207 16:17:59.500756 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dnnsf" Dec 07 16:17:59 crc kubenswrapper[4716]: I1207 16:17:59.979651 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"0f4b5de9-9760-4930-bc8e-d7e47a13ad2e","Type":"ContainerStarted","Data":"34685ed9b53b6edf3fa0cfbf84e3ff40896360a1892b3817de8c2fe662d590f2"} Dec 07 16:17:59 crc kubenswrapper[4716]: I1207 16:17:59.982056 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Dec 07 16:17:59 crc kubenswrapper[4716]: I1207 16:17:59.986176 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-g6nwl" event={"ID":"0fc0f0b8-b60a-4aff-84dd-bbec37b12d81","Type":"ContainerStarted","Data":"55e8f06e1a7815ec94f16fe041b2b4d28bf86e763e8c85938842acef15b4157c"} Dec 07 16:17:59 crc kubenswrapper[4716]: I1207 16:17:59.986377 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-g6nwl" event={"ID":"0fc0f0b8-b60a-4aff-84dd-bbec37b12d81","Type":"ContainerStarted","Data":"b52830bca207b2d26cbff53cc51ef9cfd80ae04fb8c24d15460f505401a09c4e"} Dec 07 16:17:59 crc kubenswrapper[4716]: I1207 16:17:59.986987 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-g6nwl" Dec 07 16:17:59 crc kubenswrapper[4716]: I1207 16:17:59.987518 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-g6nwl" Dec 07 16:18:00 crc kubenswrapper[4716]: I1207 16:18:00.002021 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=1.97494474 podStartE2EDuration="30.001999669s" podCreationTimestamp="2025-12-07 16:17:30 +0000 UTC" firstStartedPulling="2025-12-07 16:17:31.220910776 +0000 UTC m=+913.911195708" lastFinishedPulling="2025-12-07 16:17:59.247965725 +0000 UTC m=+941.938250637" observedRunningTime="2025-12-07 16:17:59.99808897 +0000 UTC m=+942.688373882" watchObservedRunningTime="2025-12-07 16:18:00.001999669 +0000 UTC m=+942.692284581" Dec 07 16:18:00 crc kubenswrapper[4716]: I1207 16:18:00.029487 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-g6nwl" podStartSLOduration=20.942436689 podStartE2EDuration="25.029469585s" podCreationTimestamp="2025-12-07 16:17:35 +0000 UTC" firstStartedPulling="2025-12-07 16:17:53.335322656 +0000 UTC m=+936.025607568" lastFinishedPulling="2025-12-07 16:17:57.422355552 +0000 UTC m=+940.112640464" observedRunningTime="2025-12-07 16:18:00.023290093 +0000 UTC m=+942.713575015" watchObservedRunningTime="2025-12-07 16:18:00.029469585 +0000 UTC m=+942.719754507" Dec 07 16:18:00 crc kubenswrapper[4716]: I1207 16:18:00.046551 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dnnsf"] Dec 07 16:18:00 crc kubenswrapper[4716]: W1207 16:18:00.073114 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9b870e60_acde_4d14_be71_bee41e0c53bd.slice/crio-d8b43898810fc04e8bd86c6076b7dc4ab287555ff3642b84fa38d231c67fd6cb WatchSource:0}: Error finding container d8b43898810fc04e8bd86c6076b7dc4ab287555ff3642b84fa38d231c67fd6cb: Status 404 returned error can't find the container with id d8b43898810fc04e8bd86c6076b7dc4ab287555ff3642b84fa38d231c67fd6cb Dec 07 16:18:00 crc kubenswrapper[4716]: I1207 16:18:00.400351 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Dec 07 16:18:00 crc kubenswrapper[4716]: I1207 16:18:00.400443 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Dec 07 16:18:00 crc kubenswrapper[4716]: I1207 16:18:00.994424 4716 generic.go:334] "Generic (PLEG): container finished" podID="9b870e60-acde-4d14-be71-bee41e0c53bd" containerID="aedfaa33234d36f11ded958b11e6191ff13fb623d0c2d3262e693feca0e1a0a9" exitCode=0 Dec 07 16:18:00 crc kubenswrapper[4716]: I1207 16:18:00.994573 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dnnsf" event={"ID":"9b870e60-acde-4d14-be71-bee41e0c53bd","Type":"ContainerDied","Data":"aedfaa33234d36f11ded958b11e6191ff13fb623d0c2d3262e693feca0e1a0a9"} Dec 07 16:18:00 crc kubenswrapper[4716]: I1207 16:18:00.994794 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dnnsf" event={"ID":"9b870e60-acde-4d14-be71-bee41e0c53bd","Type":"ContainerStarted","Data":"d8b43898810fc04e8bd86c6076b7dc4ab287555ff3642b84fa38d231c67fd6cb"} Dec 07 16:18:01 crc kubenswrapper[4716]: I1207 16:18:01.248375 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-666b6646f7-dwp2b" Dec 07 16:18:01 crc kubenswrapper[4716]: I1207 16:18:01.834855 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57d769cc4f-ktd8n" Dec 07 16:18:01 crc kubenswrapper[4716]: I1207 16:18:01.887509 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-dwp2b"] Dec 07 16:18:02 crc kubenswrapper[4716]: I1207 16:18:01.999986 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-666b6646f7-dwp2b" podUID="b4f34fda-3133-46e3-99d6-d49f0ae032d5" containerName="dnsmasq-dns" containerID="cri-o://f73a048dd2a4ed04fd2f0093add7ef4b8399c0526e4795314492d284d756017f" gracePeriod=10 Dec 07 16:18:02 crc kubenswrapper[4716]: I1207 16:18:02.488543 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Dec 07 16:18:02 crc kubenswrapper[4716]: I1207 16:18:02.561960 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Dec 07 16:18:02 crc kubenswrapper[4716]: I1207 16:18:02.664009 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-dwp2b" Dec 07 16:18:02 crc kubenswrapper[4716]: I1207 16:18:02.786767 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-plj7t\" (UniqueName: \"kubernetes.io/projected/b4f34fda-3133-46e3-99d6-d49f0ae032d5-kube-api-access-plj7t\") pod \"b4f34fda-3133-46e3-99d6-d49f0ae032d5\" (UID: \"b4f34fda-3133-46e3-99d6-d49f0ae032d5\") " Dec 07 16:18:02 crc kubenswrapper[4716]: I1207 16:18:02.786806 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4f34fda-3133-46e3-99d6-d49f0ae032d5-config\") pod \"b4f34fda-3133-46e3-99d6-d49f0ae032d5\" (UID: \"b4f34fda-3133-46e3-99d6-d49f0ae032d5\") " Dec 07 16:18:02 crc kubenswrapper[4716]: I1207 16:18:02.786837 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b4f34fda-3133-46e3-99d6-d49f0ae032d5-dns-svc\") pod \"b4f34fda-3133-46e3-99d6-d49f0ae032d5\" (UID: \"b4f34fda-3133-46e3-99d6-d49f0ae032d5\") " Dec 07 16:18:02 crc kubenswrapper[4716]: I1207 16:18:02.790338 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4f34fda-3133-46e3-99d6-d49f0ae032d5-kube-api-access-plj7t" (OuterVolumeSpecName: "kube-api-access-plj7t") pod "b4f34fda-3133-46e3-99d6-d49f0ae032d5" (UID: "b4f34fda-3133-46e3-99d6-d49f0ae032d5"). InnerVolumeSpecName "kube-api-access-plj7t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:18:02 crc kubenswrapper[4716]: I1207 16:18:02.821805 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b4f34fda-3133-46e3-99d6-d49f0ae032d5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b4f34fda-3133-46e3-99d6-d49f0ae032d5" (UID: "b4f34fda-3133-46e3-99d6-d49f0ae032d5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:18:02 crc kubenswrapper[4716]: I1207 16:18:02.822358 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b4f34fda-3133-46e3-99d6-d49f0ae032d5-config" (OuterVolumeSpecName: "config") pod "b4f34fda-3133-46e3-99d6-d49f0ae032d5" (UID: "b4f34fda-3133-46e3-99d6-d49f0ae032d5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:18:02 crc kubenswrapper[4716]: I1207 16:18:02.888805 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-plj7t\" (UniqueName: \"kubernetes.io/projected/b4f34fda-3133-46e3-99d6-d49f0ae032d5-kube-api-access-plj7t\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:02 crc kubenswrapper[4716]: I1207 16:18:02.888834 4716 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4f34fda-3133-46e3-99d6-d49f0ae032d5-config\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:02 crc kubenswrapper[4716]: I1207 16:18:02.888844 4716 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b4f34fda-3133-46e3-99d6-d49f0ae032d5-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:03 crc kubenswrapper[4716]: I1207 16:18:03.007635 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dnnsf" event={"ID":"9b870e60-acde-4d14-be71-bee41e0c53bd","Type":"ContainerStarted","Data":"a388ac549a0d8efe723c4e06f780c553909ba14958e61473baa94c584fd2b5ba"} Dec 07 16:18:03 crc kubenswrapper[4716]: I1207 16:18:03.011605 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"771a3c3e-34e4-43f6-a3e2-8e7156e081d1","Type":"ContainerStarted","Data":"f8bc481b01c0c9190ef1cc042d0f449e76f9754f2e44a6cd76a6b11e50122e7f"} Dec 07 16:18:03 crc kubenswrapper[4716]: I1207 16:18:03.013450 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"5c11a9b9-34de-42f7-9399-3cddce1af7c9","Type":"ContainerStarted","Data":"5ab3af991b2239c89e8bdb0324202812c1e90964c52136d3cd6656d54b6fdc7a"} Dec 07 16:18:03 crc kubenswrapper[4716]: I1207 16:18:03.017960 4716 generic.go:334] "Generic (PLEG): container finished" podID="b4f34fda-3133-46e3-99d6-d49f0ae032d5" containerID="f73a048dd2a4ed04fd2f0093add7ef4b8399c0526e4795314492d284d756017f" exitCode=0 Dec 07 16:18:03 crc kubenswrapper[4716]: I1207 16:18:03.018044 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-dwp2b" Dec 07 16:18:03 crc kubenswrapper[4716]: I1207 16:18:03.018043 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-dwp2b" event={"ID":"b4f34fda-3133-46e3-99d6-d49f0ae032d5","Type":"ContainerDied","Data":"f73a048dd2a4ed04fd2f0093add7ef4b8399c0526e4795314492d284d756017f"} Dec 07 16:18:03 crc kubenswrapper[4716]: I1207 16:18:03.018131 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-dwp2b" event={"ID":"b4f34fda-3133-46e3-99d6-d49f0ae032d5","Type":"ContainerDied","Data":"13f05e456fc8da842a3bc47f553afcaa206a2a51512c74c494302cee89f3f6b4"} Dec 07 16:18:03 crc kubenswrapper[4716]: I1207 16:18:03.018155 4716 scope.go:117] "RemoveContainer" containerID="f73a048dd2a4ed04fd2f0093add7ef4b8399c0526e4795314492d284d756017f" Dec 07 16:18:03 crc kubenswrapper[4716]: I1207 16:18:03.043147 4716 scope.go:117] "RemoveContainer" containerID="a19d8f171cbaee1ba0e8568864d5f7b86ce1134b769dd03dc3eac12098f1ce13" Dec 07 16:18:03 crc kubenswrapper[4716]: I1207 16:18:03.078597 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=15.865228121 podStartE2EDuration="25.078579431s" podCreationTimestamp="2025-12-07 16:17:38 +0000 UTC" firstStartedPulling="2025-12-07 16:17:53.332470507 +0000 UTC m=+936.022755419" lastFinishedPulling="2025-12-07 16:18:02.545821827 +0000 UTC m=+945.236106729" observedRunningTime="2025-12-07 16:18:03.073264844 +0000 UTC m=+945.763549776" watchObservedRunningTime="2025-12-07 16:18:03.078579431 +0000 UTC m=+945.768864353" Dec 07 16:18:03 crc kubenswrapper[4716]: I1207 16:18:03.080761 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=14.866753275 podStartE2EDuration="28.080752842s" podCreationTimestamp="2025-12-07 16:17:35 +0000 UTC" firstStartedPulling="2025-12-07 16:17:49.402017357 +0000 UTC m=+932.092302269" lastFinishedPulling="2025-12-07 16:18:02.616016914 +0000 UTC m=+945.306301836" observedRunningTime="2025-12-07 16:18:03.056455125 +0000 UTC m=+945.746740047" watchObservedRunningTime="2025-12-07 16:18:03.080752842 +0000 UTC m=+945.771037774" Dec 07 16:18:03 crc kubenswrapper[4716]: I1207 16:18:03.090787 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-dwp2b"] Dec 07 16:18:03 crc kubenswrapper[4716]: I1207 16:18:03.097799 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-dwp2b"] Dec 07 16:18:03 crc kubenswrapper[4716]: I1207 16:18:03.122288 4716 scope.go:117] "RemoveContainer" containerID="f73a048dd2a4ed04fd2f0093add7ef4b8399c0526e4795314492d284d756017f" Dec 07 16:18:03 crc kubenswrapper[4716]: E1207 16:18:03.122888 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f73a048dd2a4ed04fd2f0093add7ef4b8399c0526e4795314492d284d756017f\": container with ID starting with f73a048dd2a4ed04fd2f0093add7ef4b8399c0526e4795314492d284d756017f not found: ID does not exist" containerID="f73a048dd2a4ed04fd2f0093add7ef4b8399c0526e4795314492d284d756017f" Dec 07 16:18:03 crc kubenswrapper[4716]: I1207 16:18:03.122931 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f73a048dd2a4ed04fd2f0093add7ef4b8399c0526e4795314492d284d756017f"} err="failed to get container status \"f73a048dd2a4ed04fd2f0093add7ef4b8399c0526e4795314492d284d756017f\": rpc error: code = NotFound desc = could not find container \"f73a048dd2a4ed04fd2f0093add7ef4b8399c0526e4795314492d284d756017f\": container with ID starting with f73a048dd2a4ed04fd2f0093add7ef4b8399c0526e4795314492d284d756017f not found: ID does not exist" Dec 07 16:18:03 crc kubenswrapper[4716]: I1207 16:18:03.122959 4716 scope.go:117] "RemoveContainer" containerID="a19d8f171cbaee1ba0e8568864d5f7b86ce1134b769dd03dc3eac12098f1ce13" Dec 07 16:18:03 crc kubenswrapper[4716]: E1207 16:18:03.123396 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a19d8f171cbaee1ba0e8568864d5f7b86ce1134b769dd03dc3eac12098f1ce13\": container with ID starting with a19d8f171cbaee1ba0e8568864d5f7b86ce1134b769dd03dc3eac12098f1ce13 not found: ID does not exist" containerID="a19d8f171cbaee1ba0e8568864d5f7b86ce1134b769dd03dc3eac12098f1ce13" Dec 07 16:18:03 crc kubenswrapper[4716]: I1207 16:18:03.123487 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a19d8f171cbaee1ba0e8568864d5f7b86ce1134b769dd03dc3eac12098f1ce13"} err="failed to get container status \"a19d8f171cbaee1ba0e8568864d5f7b86ce1134b769dd03dc3eac12098f1ce13\": rpc error: code = NotFound desc = could not find container \"a19d8f171cbaee1ba0e8568864d5f7b86ce1134b769dd03dc3eac12098f1ce13\": container with ID starting with a19d8f171cbaee1ba0e8568864d5f7b86ce1134b769dd03dc3eac12098f1ce13 not found: ID does not exist" Dec 07 16:18:03 crc kubenswrapper[4716]: I1207 16:18:03.455453 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Dec 07 16:18:03 crc kubenswrapper[4716]: I1207 16:18:03.514523 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Dec 07 16:18:03 crc kubenswrapper[4716]: I1207 16:18:03.675641 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b4f34fda-3133-46e3-99d6-d49f0ae032d5" path="/var/lib/kubelet/pods/b4f34fda-3133-46e3-99d6-d49f0ae032d5/volumes" Dec 07 16:18:04 crc kubenswrapper[4716]: I1207 16:18:04.031012 4716 generic.go:334] "Generic (PLEG): container finished" podID="9b870e60-acde-4d14-be71-bee41e0c53bd" containerID="a388ac549a0d8efe723c4e06f780c553909ba14958e61473baa94c584fd2b5ba" exitCode=0 Dec 07 16:18:04 crc kubenswrapper[4716]: I1207 16:18:04.031073 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dnnsf" event={"ID":"9b870e60-acde-4d14-be71-bee41e0c53bd","Type":"ContainerDied","Data":"a388ac549a0d8efe723c4e06f780c553909ba14958e61473baa94c584fd2b5ba"} Dec 07 16:18:04 crc kubenswrapper[4716]: I1207 16:18:04.034225 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Dec 07 16:18:04 crc kubenswrapper[4716]: I1207 16:18:04.083282 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Dec 07 16:18:04 crc kubenswrapper[4716]: I1207 16:18:04.094904 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Dec 07 16:18:04 crc kubenswrapper[4716]: I1207 16:18:04.163884 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Dec 07 16:18:04 crc kubenswrapper[4716]: I1207 16:18:04.297674 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-2fj94"] Dec 07 16:18:04 crc kubenswrapper[4716]: E1207 16:18:04.298006 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4f34fda-3133-46e3-99d6-d49f0ae032d5" containerName="init" Dec 07 16:18:04 crc kubenswrapper[4716]: I1207 16:18:04.298022 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4f34fda-3133-46e3-99d6-d49f0ae032d5" containerName="init" Dec 07 16:18:04 crc kubenswrapper[4716]: E1207 16:18:04.298046 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4f34fda-3133-46e3-99d6-d49f0ae032d5" containerName="dnsmasq-dns" Dec 07 16:18:04 crc kubenswrapper[4716]: I1207 16:18:04.298053 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4f34fda-3133-46e3-99d6-d49f0ae032d5" containerName="dnsmasq-dns" Dec 07 16:18:04 crc kubenswrapper[4716]: I1207 16:18:04.298235 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4f34fda-3133-46e3-99d6-d49f0ae032d5" containerName="dnsmasq-dns" Dec 07 16:18:04 crc kubenswrapper[4716]: I1207 16:18:04.299098 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-2fj94" Dec 07 16:18:04 crc kubenswrapper[4716]: I1207 16:18:04.300636 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Dec 07 16:18:04 crc kubenswrapper[4716]: I1207 16:18:04.312195 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-2fj94"] Dec 07 16:18:04 crc kubenswrapper[4716]: I1207 16:18:04.380960 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-8r8gk"] Dec 07 16:18:04 crc kubenswrapper[4716]: I1207 16:18:04.382060 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-8r8gk" Dec 07 16:18:04 crc kubenswrapper[4716]: I1207 16:18:04.383795 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Dec 07 16:18:04 crc kubenswrapper[4716]: I1207 16:18:04.394064 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-8r8gk"] Dec 07 16:18:04 crc kubenswrapper[4716]: I1207 16:18:04.426330 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0a2f2b59-b71e-416d-bdc6-2a69ed43b920-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-2fj94\" (UID: \"0a2f2b59-b71e-416d-bdc6-2a69ed43b920\") " pod="openstack/dnsmasq-dns-7fd796d7df-2fj94" Dec 07 16:18:04 crc kubenswrapper[4716]: I1207 16:18:04.426442 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a2f2b59-b71e-416d-bdc6-2a69ed43b920-config\") pod \"dnsmasq-dns-7fd796d7df-2fj94\" (UID: \"0a2f2b59-b71e-416d-bdc6-2a69ed43b920\") " pod="openstack/dnsmasq-dns-7fd796d7df-2fj94" Dec 07 16:18:04 crc kubenswrapper[4716]: I1207 16:18:04.426480 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-82sq7\" (UniqueName: \"kubernetes.io/projected/0a2f2b59-b71e-416d-bdc6-2a69ed43b920-kube-api-access-82sq7\") pod \"dnsmasq-dns-7fd796d7df-2fj94\" (UID: \"0a2f2b59-b71e-416d-bdc6-2a69ed43b920\") " pod="openstack/dnsmasq-dns-7fd796d7df-2fj94" Dec 07 16:18:04 crc kubenswrapper[4716]: I1207 16:18:04.426555 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0a2f2b59-b71e-416d-bdc6-2a69ed43b920-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-2fj94\" (UID: \"0a2f2b59-b71e-416d-bdc6-2a69ed43b920\") " pod="openstack/dnsmasq-dns-7fd796d7df-2fj94" Dec 07 16:18:04 crc kubenswrapper[4716]: I1207 16:18:04.536007 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1fcf6527-b54d-4239-a4c4-a9cd6fe35bd2-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-8r8gk\" (UID: \"1fcf6527-b54d-4239-a4c4-a9cd6fe35bd2\") " pod="openstack/ovn-controller-metrics-8r8gk" Dec 07 16:18:04 crc kubenswrapper[4716]: I1207 16:18:04.536090 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/1fcf6527-b54d-4239-a4c4-a9cd6fe35bd2-ovs-rundir\") pod \"ovn-controller-metrics-8r8gk\" (UID: \"1fcf6527-b54d-4239-a4c4-a9cd6fe35bd2\") " pod="openstack/ovn-controller-metrics-8r8gk" Dec 07 16:18:04 crc kubenswrapper[4716]: I1207 16:18:04.536126 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1fcf6527-b54d-4239-a4c4-a9cd6fe35bd2-config\") pod \"ovn-controller-metrics-8r8gk\" (UID: \"1fcf6527-b54d-4239-a4c4-a9cd6fe35bd2\") " pod="openstack/ovn-controller-metrics-8r8gk" Dec 07 16:18:04 crc kubenswrapper[4716]: I1207 16:18:04.536153 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0a2f2b59-b71e-416d-bdc6-2a69ed43b920-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-2fj94\" (UID: \"0a2f2b59-b71e-416d-bdc6-2a69ed43b920\") " pod="openstack/dnsmasq-dns-7fd796d7df-2fj94" Dec 07 16:18:04 crc kubenswrapper[4716]: I1207 16:18:04.536266 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/1fcf6527-b54d-4239-a4c4-a9cd6fe35bd2-ovn-rundir\") pod \"ovn-controller-metrics-8r8gk\" (UID: \"1fcf6527-b54d-4239-a4c4-a9cd6fe35bd2\") " pod="openstack/ovn-controller-metrics-8r8gk" Dec 07 16:18:04 crc kubenswrapper[4716]: I1207 16:18:04.536301 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmx9z\" (UniqueName: \"kubernetes.io/projected/1fcf6527-b54d-4239-a4c4-a9cd6fe35bd2-kube-api-access-pmx9z\") pod \"ovn-controller-metrics-8r8gk\" (UID: \"1fcf6527-b54d-4239-a4c4-a9cd6fe35bd2\") " pod="openstack/ovn-controller-metrics-8r8gk" Dec 07 16:18:04 crc kubenswrapper[4716]: I1207 16:18:04.536327 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a2f2b59-b71e-416d-bdc6-2a69ed43b920-config\") pod \"dnsmasq-dns-7fd796d7df-2fj94\" (UID: \"0a2f2b59-b71e-416d-bdc6-2a69ed43b920\") " pod="openstack/dnsmasq-dns-7fd796d7df-2fj94" Dec 07 16:18:04 crc kubenswrapper[4716]: I1207 16:18:04.536348 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-82sq7\" (UniqueName: \"kubernetes.io/projected/0a2f2b59-b71e-416d-bdc6-2a69ed43b920-kube-api-access-82sq7\") pod \"dnsmasq-dns-7fd796d7df-2fj94\" (UID: \"0a2f2b59-b71e-416d-bdc6-2a69ed43b920\") " pod="openstack/dnsmasq-dns-7fd796d7df-2fj94" Dec 07 16:18:04 crc kubenswrapper[4716]: I1207 16:18:04.536374 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1fcf6527-b54d-4239-a4c4-a9cd6fe35bd2-combined-ca-bundle\") pod \"ovn-controller-metrics-8r8gk\" (UID: \"1fcf6527-b54d-4239-a4c4-a9cd6fe35bd2\") " pod="openstack/ovn-controller-metrics-8r8gk" Dec 07 16:18:04 crc kubenswrapper[4716]: I1207 16:18:04.536397 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0a2f2b59-b71e-416d-bdc6-2a69ed43b920-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-2fj94\" (UID: \"0a2f2b59-b71e-416d-bdc6-2a69ed43b920\") " pod="openstack/dnsmasq-dns-7fd796d7df-2fj94" Dec 07 16:18:04 crc kubenswrapper[4716]: I1207 16:18:04.536952 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0a2f2b59-b71e-416d-bdc6-2a69ed43b920-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-2fj94\" (UID: \"0a2f2b59-b71e-416d-bdc6-2a69ed43b920\") " pod="openstack/dnsmasq-dns-7fd796d7df-2fj94" Dec 07 16:18:04 crc kubenswrapper[4716]: I1207 16:18:04.537155 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a2f2b59-b71e-416d-bdc6-2a69ed43b920-config\") pod \"dnsmasq-dns-7fd796d7df-2fj94\" (UID: \"0a2f2b59-b71e-416d-bdc6-2a69ed43b920\") " pod="openstack/dnsmasq-dns-7fd796d7df-2fj94" Dec 07 16:18:04 crc kubenswrapper[4716]: I1207 16:18:04.539194 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0a2f2b59-b71e-416d-bdc6-2a69ed43b920-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-2fj94\" (UID: \"0a2f2b59-b71e-416d-bdc6-2a69ed43b920\") " pod="openstack/dnsmasq-dns-7fd796d7df-2fj94" Dec 07 16:18:04 crc kubenswrapper[4716]: I1207 16:18:04.557789 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-82sq7\" (UniqueName: \"kubernetes.io/projected/0a2f2b59-b71e-416d-bdc6-2a69ed43b920-kube-api-access-82sq7\") pod \"dnsmasq-dns-7fd796d7df-2fj94\" (UID: \"0a2f2b59-b71e-416d-bdc6-2a69ed43b920\") " pod="openstack/dnsmasq-dns-7fd796d7df-2fj94" Dec 07 16:18:04 crc kubenswrapper[4716]: I1207 16:18:04.615964 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-2fj94" Dec 07 16:18:04 crc kubenswrapper[4716]: I1207 16:18:04.638239 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1fcf6527-b54d-4239-a4c4-a9cd6fe35bd2-combined-ca-bundle\") pod \"ovn-controller-metrics-8r8gk\" (UID: \"1fcf6527-b54d-4239-a4c4-a9cd6fe35bd2\") " pod="openstack/ovn-controller-metrics-8r8gk" Dec 07 16:18:04 crc kubenswrapper[4716]: I1207 16:18:04.638329 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1fcf6527-b54d-4239-a4c4-a9cd6fe35bd2-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-8r8gk\" (UID: \"1fcf6527-b54d-4239-a4c4-a9cd6fe35bd2\") " pod="openstack/ovn-controller-metrics-8r8gk" Dec 07 16:18:04 crc kubenswrapper[4716]: I1207 16:18:04.638379 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/1fcf6527-b54d-4239-a4c4-a9cd6fe35bd2-ovs-rundir\") pod \"ovn-controller-metrics-8r8gk\" (UID: \"1fcf6527-b54d-4239-a4c4-a9cd6fe35bd2\") " pod="openstack/ovn-controller-metrics-8r8gk" Dec 07 16:18:04 crc kubenswrapper[4716]: I1207 16:18:04.638417 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1fcf6527-b54d-4239-a4c4-a9cd6fe35bd2-config\") pod \"ovn-controller-metrics-8r8gk\" (UID: \"1fcf6527-b54d-4239-a4c4-a9cd6fe35bd2\") " pod="openstack/ovn-controller-metrics-8r8gk" Dec 07 16:18:04 crc kubenswrapper[4716]: I1207 16:18:04.638462 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/1fcf6527-b54d-4239-a4c4-a9cd6fe35bd2-ovn-rundir\") pod \"ovn-controller-metrics-8r8gk\" (UID: \"1fcf6527-b54d-4239-a4c4-a9cd6fe35bd2\") " pod="openstack/ovn-controller-metrics-8r8gk" Dec 07 16:18:04 crc kubenswrapper[4716]: I1207 16:18:04.638488 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmx9z\" (UniqueName: \"kubernetes.io/projected/1fcf6527-b54d-4239-a4c4-a9cd6fe35bd2-kube-api-access-pmx9z\") pod \"ovn-controller-metrics-8r8gk\" (UID: \"1fcf6527-b54d-4239-a4c4-a9cd6fe35bd2\") " pod="openstack/ovn-controller-metrics-8r8gk" Dec 07 16:18:04 crc kubenswrapper[4716]: I1207 16:18:04.639532 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/1fcf6527-b54d-4239-a4c4-a9cd6fe35bd2-ovs-rundir\") pod \"ovn-controller-metrics-8r8gk\" (UID: \"1fcf6527-b54d-4239-a4c4-a9cd6fe35bd2\") " pod="openstack/ovn-controller-metrics-8r8gk" Dec 07 16:18:04 crc kubenswrapper[4716]: I1207 16:18:04.640223 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/1fcf6527-b54d-4239-a4c4-a9cd6fe35bd2-ovn-rundir\") pod \"ovn-controller-metrics-8r8gk\" (UID: \"1fcf6527-b54d-4239-a4c4-a9cd6fe35bd2\") " pod="openstack/ovn-controller-metrics-8r8gk" Dec 07 16:18:04 crc kubenswrapper[4716]: I1207 16:18:04.641346 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1fcf6527-b54d-4239-a4c4-a9cd6fe35bd2-config\") pod \"ovn-controller-metrics-8r8gk\" (UID: \"1fcf6527-b54d-4239-a4c4-a9cd6fe35bd2\") " pod="openstack/ovn-controller-metrics-8r8gk" Dec 07 16:18:04 crc kubenswrapper[4716]: I1207 16:18:04.655588 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1fcf6527-b54d-4239-a4c4-a9cd6fe35bd2-combined-ca-bundle\") pod \"ovn-controller-metrics-8r8gk\" (UID: \"1fcf6527-b54d-4239-a4c4-a9cd6fe35bd2\") " pod="openstack/ovn-controller-metrics-8r8gk" Dec 07 16:18:04 crc kubenswrapper[4716]: I1207 16:18:04.658656 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1fcf6527-b54d-4239-a4c4-a9cd6fe35bd2-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-8r8gk\" (UID: \"1fcf6527-b54d-4239-a4c4-a9cd6fe35bd2\") " pod="openstack/ovn-controller-metrics-8r8gk" Dec 07 16:18:04 crc kubenswrapper[4716]: I1207 16:18:04.671975 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmx9z\" (UniqueName: \"kubernetes.io/projected/1fcf6527-b54d-4239-a4c4-a9cd6fe35bd2-kube-api-access-pmx9z\") pod \"ovn-controller-metrics-8r8gk\" (UID: \"1fcf6527-b54d-4239-a4c4-a9cd6fe35bd2\") " pod="openstack/ovn-controller-metrics-8r8gk" Dec 07 16:18:04 crc kubenswrapper[4716]: I1207 16:18:04.684822 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-2fj94"] Dec 07 16:18:04 crc kubenswrapper[4716]: I1207 16:18:04.700823 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-8r8gk" Dec 07 16:18:04 crc kubenswrapper[4716]: I1207 16:18:04.712744 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-tsr96"] Dec 07 16:18:04 crc kubenswrapper[4716]: I1207 16:18:04.713955 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-tsr96" Dec 07 16:18:04 crc kubenswrapper[4716]: I1207 16:18:04.725322 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Dec 07 16:18:04 crc kubenswrapper[4716]: I1207 16:18:04.728966 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-tsr96"] Dec 07 16:18:04 crc kubenswrapper[4716]: I1207 16:18:04.840582 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/61f22dc6-45b1-4307-8eee-fd493fc96408-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-tsr96\" (UID: \"61f22dc6-45b1-4307-8eee-fd493fc96408\") " pod="openstack/dnsmasq-dns-86db49b7ff-tsr96" Dec 07 16:18:04 crc kubenswrapper[4716]: I1207 16:18:04.840700 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/61f22dc6-45b1-4307-8eee-fd493fc96408-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-tsr96\" (UID: \"61f22dc6-45b1-4307-8eee-fd493fc96408\") " pod="openstack/dnsmasq-dns-86db49b7ff-tsr96" Dec 07 16:18:04 crc kubenswrapper[4716]: I1207 16:18:04.840754 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c9wjv\" (UniqueName: \"kubernetes.io/projected/61f22dc6-45b1-4307-8eee-fd493fc96408-kube-api-access-c9wjv\") pod \"dnsmasq-dns-86db49b7ff-tsr96\" (UID: \"61f22dc6-45b1-4307-8eee-fd493fc96408\") " pod="openstack/dnsmasq-dns-86db49b7ff-tsr96" Dec 07 16:18:04 crc kubenswrapper[4716]: I1207 16:18:04.840774 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/61f22dc6-45b1-4307-8eee-fd493fc96408-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-tsr96\" (UID: \"61f22dc6-45b1-4307-8eee-fd493fc96408\") " pod="openstack/dnsmasq-dns-86db49b7ff-tsr96" Dec 07 16:18:04 crc kubenswrapper[4716]: I1207 16:18:04.840807 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61f22dc6-45b1-4307-8eee-fd493fc96408-config\") pod \"dnsmasq-dns-86db49b7ff-tsr96\" (UID: \"61f22dc6-45b1-4307-8eee-fd493fc96408\") " pod="openstack/dnsmasq-dns-86db49b7ff-tsr96" Dec 07 16:18:04 crc kubenswrapper[4716]: I1207 16:18:04.882121 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8phkg" Dec 07 16:18:04 crc kubenswrapper[4716]: I1207 16:18:04.882158 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8phkg" Dec 07 16:18:04 crc kubenswrapper[4716]: I1207 16:18:04.942628 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61f22dc6-45b1-4307-8eee-fd493fc96408-config\") pod \"dnsmasq-dns-86db49b7ff-tsr96\" (UID: \"61f22dc6-45b1-4307-8eee-fd493fc96408\") " pod="openstack/dnsmasq-dns-86db49b7ff-tsr96" Dec 07 16:18:04 crc kubenswrapper[4716]: I1207 16:18:04.942666 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/61f22dc6-45b1-4307-8eee-fd493fc96408-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-tsr96\" (UID: \"61f22dc6-45b1-4307-8eee-fd493fc96408\") " pod="openstack/dnsmasq-dns-86db49b7ff-tsr96" Dec 07 16:18:04 crc kubenswrapper[4716]: I1207 16:18:04.942766 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/61f22dc6-45b1-4307-8eee-fd493fc96408-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-tsr96\" (UID: \"61f22dc6-45b1-4307-8eee-fd493fc96408\") " pod="openstack/dnsmasq-dns-86db49b7ff-tsr96" Dec 07 16:18:04 crc kubenswrapper[4716]: I1207 16:18:04.942810 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c9wjv\" (UniqueName: \"kubernetes.io/projected/61f22dc6-45b1-4307-8eee-fd493fc96408-kube-api-access-c9wjv\") pod \"dnsmasq-dns-86db49b7ff-tsr96\" (UID: \"61f22dc6-45b1-4307-8eee-fd493fc96408\") " pod="openstack/dnsmasq-dns-86db49b7ff-tsr96" Dec 07 16:18:04 crc kubenswrapper[4716]: I1207 16:18:04.942829 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/61f22dc6-45b1-4307-8eee-fd493fc96408-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-tsr96\" (UID: \"61f22dc6-45b1-4307-8eee-fd493fc96408\") " pod="openstack/dnsmasq-dns-86db49b7ff-tsr96" Dec 07 16:18:04 crc kubenswrapper[4716]: I1207 16:18:04.943742 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/61f22dc6-45b1-4307-8eee-fd493fc96408-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-tsr96\" (UID: \"61f22dc6-45b1-4307-8eee-fd493fc96408\") " pod="openstack/dnsmasq-dns-86db49b7ff-tsr96" Dec 07 16:18:04 crc kubenswrapper[4716]: I1207 16:18:04.944289 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61f22dc6-45b1-4307-8eee-fd493fc96408-config\") pod \"dnsmasq-dns-86db49b7ff-tsr96\" (UID: \"61f22dc6-45b1-4307-8eee-fd493fc96408\") " pod="openstack/dnsmasq-dns-86db49b7ff-tsr96" Dec 07 16:18:04 crc kubenswrapper[4716]: I1207 16:18:04.944841 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/61f22dc6-45b1-4307-8eee-fd493fc96408-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-tsr96\" (UID: \"61f22dc6-45b1-4307-8eee-fd493fc96408\") " pod="openstack/dnsmasq-dns-86db49b7ff-tsr96" Dec 07 16:18:04 crc kubenswrapper[4716]: I1207 16:18:04.948658 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/61f22dc6-45b1-4307-8eee-fd493fc96408-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-tsr96\" (UID: \"61f22dc6-45b1-4307-8eee-fd493fc96408\") " pod="openstack/dnsmasq-dns-86db49b7ff-tsr96" Dec 07 16:18:04 crc kubenswrapper[4716]: I1207 16:18:04.969510 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c9wjv\" (UniqueName: \"kubernetes.io/projected/61f22dc6-45b1-4307-8eee-fd493fc96408-kube-api-access-c9wjv\") pod \"dnsmasq-dns-86db49b7ff-tsr96\" (UID: \"61f22dc6-45b1-4307-8eee-fd493fc96408\") " pod="openstack/dnsmasq-dns-86db49b7ff-tsr96" Dec 07 16:18:04 crc kubenswrapper[4716]: I1207 16:18:04.988438 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8phkg" Dec 07 16:18:05 crc kubenswrapper[4716]: I1207 16:18:05.048589 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dnnsf" event={"ID":"9b870e60-acde-4d14-be71-bee41e0c53bd","Type":"ContainerStarted","Data":"4e61d65f1c66e3295fbc6e46d16e92fe2d66a11ff291bd2c1fff6fb4654b99d1"} Dec 07 16:18:05 crc kubenswrapper[4716]: I1207 16:18:05.049463 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Dec 07 16:18:05 crc kubenswrapper[4716]: I1207 16:18:05.078442 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-dnnsf" podStartSLOduration=3.48138976 podStartE2EDuration="6.078418941s" podCreationTimestamp="2025-12-07 16:17:59 +0000 UTC" firstStartedPulling="2025-12-07 16:18:01.889370144 +0000 UTC m=+944.579655056" lastFinishedPulling="2025-12-07 16:18:04.486399335 +0000 UTC m=+947.176684237" observedRunningTime="2025-12-07 16:18:05.071619292 +0000 UTC m=+947.761904204" watchObservedRunningTime="2025-12-07 16:18:05.078418941 +0000 UTC m=+947.768703873" Dec 07 16:18:05 crc kubenswrapper[4716]: I1207 16:18:05.102570 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Dec 07 16:18:05 crc kubenswrapper[4716]: I1207 16:18:05.121065 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8phkg" Dec 07 16:18:05 crc kubenswrapper[4716]: I1207 16:18:05.137433 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-tsr96" Dec 07 16:18:05 crc kubenswrapper[4716]: I1207 16:18:05.280379 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Dec 07 16:18:05 crc kubenswrapper[4716]: I1207 16:18:05.281835 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 07 16:18:05 crc kubenswrapper[4716]: I1207 16:18:05.287585 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Dec 07 16:18:05 crc kubenswrapper[4716]: I1207 16:18:05.287683 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Dec 07 16:18:05 crc kubenswrapper[4716]: I1207 16:18:05.287797 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-bh7c2" Dec 07 16:18:05 crc kubenswrapper[4716]: I1207 16:18:05.288409 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Dec 07 16:18:05 crc kubenswrapper[4716]: I1207 16:18:05.291595 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Dec 07 16:18:05 crc kubenswrapper[4716]: I1207 16:18:05.297127 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 07 16:18:05 crc kubenswrapper[4716]: I1207 16:18:05.339953 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-2fj94"] Dec 07 16:18:05 crc kubenswrapper[4716]: I1207 16:18:05.354134 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/3ae6bded-8084-4e07-afcc-a7fafb718e3a-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"3ae6bded-8084-4e07-afcc-a7fafb718e3a\") " pod="openstack/ovn-northd-0" Dec 07 16:18:05 crc kubenswrapper[4716]: I1207 16:18:05.354468 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ae6bded-8084-4e07-afcc-a7fafb718e3a-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"3ae6bded-8084-4e07-afcc-a7fafb718e3a\") " pod="openstack/ovn-northd-0" Dec 07 16:18:05 crc kubenswrapper[4716]: I1207 16:18:05.354492 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ae6bded-8084-4e07-afcc-a7fafb718e3a-config\") pod \"ovn-northd-0\" (UID: \"3ae6bded-8084-4e07-afcc-a7fafb718e3a\") " pod="openstack/ovn-northd-0" Dec 07 16:18:05 crc kubenswrapper[4716]: I1207 16:18:05.354514 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ae6bded-8084-4e07-afcc-a7fafb718e3a-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"3ae6bded-8084-4e07-afcc-a7fafb718e3a\") " pod="openstack/ovn-northd-0" Dec 07 16:18:05 crc kubenswrapper[4716]: I1207 16:18:05.354537 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ae6bded-8084-4e07-afcc-a7fafb718e3a-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"3ae6bded-8084-4e07-afcc-a7fafb718e3a\") " pod="openstack/ovn-northd-0" Dec 07 16:18:05 crc kubenswrapper[4716]: I1207 16:18:05.354556 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tf2ck\" (UniqueName: \"kubernetes.io/projected/3ae6bded-8084-4e07-afcc-a7fafb718e3a-kube-api-access-tf2ck\") pod \"ovn-northd-0\" (UID: \"3ae6bded-8084-4e07-afcc-a7fafb718e3a\") " pod="openstack/ovn-northd-0" Dec 07 16:18:05 crc kubenswrapper[4716]: I1207 16:18:05.354582 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3ae6bded-8084-4e07-afcc-a7fafb718e3a-scripts\") pod \"ovn-northd-0\" (UID: \"3ae6bded-8084-4e07-afcc-a7fafb718e3a\") " pod="openstack/ovn-northd-0" Dec 07 16:18:05 crc kubenswrapper[4716]: I1207 16:18:05.386503 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Dec 07 16:18:05 crc kubenswrapper[4716]: W1207 16:18:05.424723 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1fcf6527_b54d_4239_a4c4_a9cd6fe35bd2.slice/crio-8373ca8a5c0f0d3075002f9d1363514f6a15f397319e3a8f4ff8622439b70820 WatchSource:0}: Error finding container 8373ca8a5c0f0d3075002f9d1363514f6a15f397319e3a8f4ff8622439b70820: Status 404 returned error can't find the container with id 8373ca8a5c0f0d3075002f9d1363514f6a15f397319e3a8f4ff8622439b70820 Dec 07 16:18:05 crc kubenswrapper[4716]: I1207 16:18:05.428023 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-8r8gk"] Dec 07 16:18:05 crc kubenswrapper[4716]: I1207 16:18:05.455674 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ae6bded-8084-4e07-afcc-a7fafb718e3a-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"3ae6bded-8084-4e07-afcc-a7fafb718e3a\") " pod="openstack/ovn-northd-0" Dec 07 16:18:05 crc kubenswrapper[4716]: I1207 16:18:05.455734 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ae6bded-8084-4e07-afcc-a7fafb718e3a-config\") pod \"ovn-northd-0\" (UID: \"3ae6bded-8084-4e07-afcc-a7fafb718e3a\") " pod="openstack/ovn-northd-0" Dec 07 16:18:05 crc kubenswrapper[4716]: I1207 16:18:05.455783 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ae6bded-8084-4e07-afcc-a7fafb718e3a-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"3ae6bded-8084-4e07-afcc-a7fafb718e3a\") " pod="openstack/ovn-northd-0" Dec 07 16:18:05 crc kubenswrapper[4716]: I1207 16:18:05.455808 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ae6bded-8084-4e07-afcc-a7fafb718e3a-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"3ae6bded-8084-4e07-afcc-a7fafb718e3a\") " pod="openstack/ovn-northd-0" Dec 07 16:18:05 crc kubenswrapper[4716]: I1207 16:18:05.455827 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tf2ck\" (UniqueName: \"kubernetes.io/projected/3ae6bded-8084-4e07-afcc-a7fafb718e3a-kube-api-access-tf2ck\") pod \"ovn-northd-0\" (UID: \"3ae6bded-8084-4e07-afcc-a7fafb718e3a\") " pod="openstack/ovn-northd-0" Dec 07 16:18:05 crc kubenswrapper[4716]: I1207 16:18:05.455879 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3ae6bded-8084-4e07-afcc-a7fafb718e3a-scripts\") pod \"ovn-northd-0\" (UID: \"3ae6bded-8084-4e07-afcc-a7fafb718e3a\") " pod="openstack/ovn-northd-0" Dec 07 16:18:05 crc kubenswrapper[4716]: I1207 16:18:05.456028 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/3ae6bded-8084-4e07-afcc-a7fafb718e3a-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"3ae6bded-8084-4e07-afcc-a7fafb718e3a\") " pod="openstack/ovn-northd-0" Dec 07 16:18:05 crc kubenswrapper[4716]: I1207 16:18:05.456797 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/3ae6bded-8084-4e07-afcc-a7fafb718e3a-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"3ae6bded-8084-4e07-afcc-a7fafb718e3a\") " pod="openstack/ovn-northd-0" Dec 07 16:18:05 crc kubenswrapper[4716]: I1207 16:18:05.456847 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ae6bded-8084-4e07-afcc-a7fafb718e3a-config\") pod \"ovn-northd-0\" (UID: \"3ae6bded-8084-4e07-afcc-a7fafb718e3a\") " pod="openstack/ovn-northd-0" Dec 07 16:18:05 crc kubenswrapper[4716]: I1207 16:18:05.459588 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3ae6bded-8084-4e07-afcc-a7fafb718e3a-scripts\") pod \"ovn-northd-0\" (UID: \"3ae6bded-8084-4e07-afcc-a7fafb718e3a\") " pod="openstack/ovn-northd-0" Dec 07 16:18:05 crc kubenswrapper[4716]: I1207 16:18:05.477728 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tf2ck\" (UniqueName: \"kubernetes.io/projected/3ae6bded-8084-4e07-afcc-a7fafb718e3a-kube-api-access-tf2ck\") pod \"ovn-northd-0\" (UID: \"3ae6bded-8084-4e07-afcc-a7fafb718e3a\") " pod="openstack/ovn-northd-0" Dec 07 16:18:05 crc kubenswrapper[4716]: I1207 16:18:05.478444 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ae6bded-8084-4e07-afcc-a7fafb718e3a-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"3ae6bded-8084-4e07-afcc-a7fafb718e3a\") " pod="openstack/ovn-northd-0" Dec 07 16:18:05 crc kubenswrapper[4716]: I1207 16:18:05.479062 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ae6bded-8084-4e07-afcc-a7fafb718e3a-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"3ae6bded-8084-4e07-afcc-a7fafb718e3a\") " pod="openstack/ovn-northd-0" Dec 07 16:18:05 crc kubenswrapper[4716]: I1207 16:18:05.480899 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ae6bded-8084-4e07-afcc-a7fafb718e3a-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"3ae6bded-8084-4e07-afcc-a7fafb718e3a\") " pod="openstack/ovn-northd-0" Dec 07 16:18:05 crc kubenswrapper[4716]: I1207 16:18:05.607904 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 07 16:18:05 crc kubenswrapper[4716]: W1207 16:18:05.666392 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod61f22dc6_45b1_4307_8eee_fd493fc96408.slice/crio-86af27eb39f5d5e4e2abd2ef9fc47d0c36e0f61f95a94e4e99f102b13a975d29 WatchSource:0}: Error finding container 86af27eb39f5d5e4e2abd2ef9fc47d0c36e0f61f95a94e4e99f102b13a975d29: Status 404 returned error can't find the container with id 86af27eb39f5d5e4e2abd2ef9fc47d0c36e0f61f95a94e4e99f102b13a975d29 Dec 07 16:18:05 crc kubenswrapper[4716]: I1207 16:18:05.667156 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-tsr96"] Dec 07 16:18:05 crc kubenswrapper[4716]: I1207 16:18:05.813240 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Dec 07 16:18:05 crc kubenswrapper[4716]: I1207 16:18:05.918021 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8phkg"] Dec 07 16:18:06 crc kubenswrapper[4716]: I1207 16:18:06.066277 4716 generic.go:334] "Generic (PLEG): container finished" podID="0a2f2b59-b71e-416d-bdc6-2a69ed43b920" containerID="1f791d3911a76ca9ff7033df05bef650990f28ec02c5858b69fc86f2a7f76d19" exitCode=0 Dec 07 16:18:06 crc kubenswrapper[4716]: I1207 16:18:06.066515 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-2fj94" event={"ID":"0a2f2b59-b71e-416d-bdc6-2a69ed43b920","Type":"ContainerDied","Data":"1f791d3911a76ca9ff7033df05bef650990f28ec02c5858b69fc86f2a7f76d19"} Dec 07 16:18:06 crc kubenswrapper[4716]: I1207 16:18:06.066549 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-2fj94" event={"ID":"0a2f2b59-b71e-416d-bdc6-2a69ed43b920","Type":"ContainerStarted","Data":"197d58eb9607d2e4dc6ae0543d244d9cec5ce76c9a80a4b4030a580b56003f90"} Dec 07 16:18:06 crc kubenswrapper[4716]: I1207 16:18:06.074902 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-8r8gk" event={"ID":"1fcf6527-b54d-4239-a4c4-a9cd6fe35bd2","Type":"ContainerStarted","Data":"39dba03aeb25f57b729ba5248bca2d2cd009c76447cae6404cfebcadf3c52cf2"} Dec 07 16:18:06 crc kubenswrapper[4716]: I1207 16:18:06.075255 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-8r8gk" event={"ID":"1fcf6527-b54d-4239-a4c4-a9cd6fe35bd2","Type":"ContainerStarted","Data":"8373ca8a5c0f0d3075002f9d1363514f6a15f397319e3a8f4ff8622439b70820"} Dec 07 16:18:06 crc kubenswrapper[4716]: I1207 16:18:06.088863 4716 generic.go:334] "Generic (PLEG): container finished" podID="61f22dc6-45b1-4307-8eee-fd493fc96408" containerID="426ee91ecd278e5884292d74ef2d1931368271bc3d735ed09ca634e3000eb534" exitCode=0 Dec 07 16:18:06 crc kubenswrapper[4716]: I1207 16:18:06.088984 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-tsr96" event={"ID":"61f22dc6-45b1-4307-8eee-fd493fc96408","Type":"ContainerDied","Data":"426ee91ecd278e5884292d74ef2d1931368271bc3d735ed09ca634e3000eb534"} Dec 07 16:18:06 crc kubenswrapper[4716]: I1207 16:18:06.089016 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-tsr96" event={"ID":"61f22dc6-45b1-4307-8eee-fd493fc96408","Type":"ContainerStarted","Data":"86af27eb39f5d5e4e2abd2ef9fc47d0c36e0f61f95a94e4e99f102b13a975d29"} Dec 07 16:18:06 crc kubenswrapper[4716]: I1207 16:18:06.104501 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-zkgtv"] Dec 07 16:18:06 crc kubenswrapper[4716]: I1207 16:18:06.126835 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-zkgtv" Dec 07 16:18:06 crc kubenswrapper[4716]: I1207 16:18:06.169801 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-92a4-account-create-update-pzqm6"] Dec 07 16:18:06 crc kubenswrapper[4716]: I1207 16:18:06.179481 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-92a4-account-create-update-pzqm6" Dec 07 16:18:06 crc kubenswrapper[4716]: I1207 16:18:06.190469 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Dec 07 16:18:06 crc kubenswrapper[4716]: I1207 16:18:06.210748 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-zkgtv"] Dec 07 16:18:06 crc kubenswrapper[4716]: I1207 16:18:06.234489 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-92a4-account-create-update-pzqm6"] Dec 07 16:18:06 crc kubenswrapper[4716]: I1207 16:18:06.244827 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 07 16:18:06 crc kubenswrapper[4716]: I1207 16:18:06.247256 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-8r8gk" podStartSLOduration=2.247235041 podStartE2EDuration="2.247235041s" podCreationTimestamp="2025-12-07 16:18:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:18:06.168712402 +0000 UTC m=+948.858997314" watchObservedRunningTime="2025-12-07 16:18:06.247235041 +0000 UTC m=+948.937519943" Dec 07 16:18:06 crc kubenswrapper[4716]: I1207 16:18:06.290696 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cm8dl\" (UniqueName: \"kubernetes.io/projected/3a8a9543-3bd8-4f32-8e70-82eea76874df-kube-api-access-cm8dl\") pod \"glance-db-create-zkgtv\" (UID: \"3a8a9543-3bd8-4f32-8e70-82eea76874df\") " pod="openstack/glance-db-create-zkgtv" Dec 07 16:18:06 crc kubenswrapper[4716]: I1207 16:18:06.290751 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zcwrc\" (UniqueName: \"kubernetes.io/projected/708e4fff-969d-4753-9223-7ecbe84890a9-kube-api-access-zcwrc\") pod \"glance-92a4-account-create-update-pzqm6\" (UID: \"708e4fff-969d-4753-9223-7ecbe84890a9\") " pod="openstack/glance-92a4-account-create-update-pzqm6" Dec 07 16:18:06 crc kubenswrapper[4716]: I1207 16:18:06.290776 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3a8a9543-3bd8-4f32-8e70-82eea76874df-operator-scripts\") pod \"glance-db-create-zkgtv\" (UID: \"3a8a9543-3bd8-4f32-8e70-82eea76874df\") " pod="openstack/glance-db-create-zkgtv" Dec 07 16:18:06 crc kubenswrapper[4716]: I1207 16:18:06.290827 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/708e4fff-969d-4753-9223-7ecbe84890a9-operator-scripts\") pod \"glance-92a4-account-create-update-pzqm6\" (UID: \"708e4fff-969d-4753-9223-7ecbe84890a9\") " pod="openstack/glance-92a4-account-create-update-pzqm6" Dec 07 16:18:06 crc kubenswrapper[4716]: I1207 16:18:06.397988 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zcwrc\" (UniqueName: \"kubernetes.io/projected/708e4fff-969d-4753-9223-7ecbe84890a9-kube-api-access-zcwrc\") pod \"glance-92a4-account-create-update-pzqm6\" (UID: \"708e4fff-969d-4753-9223-7ecbe84890a9\") " pod="openstack/glance-92a4-account-create-update-pzqm6" Dec 07 16:18:06 crc kubenswrapper[4716]: I1207 16:18:06.399306 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3a8a9543-3bd8-4f32-8e70-82eea76874df-operator-scripts\") pod \"glance-db-create-zkgtv\" (UID: \"3a8a9543-3bd8-4f32-8e70-82eea76874df\") " pod="openstack/glance-db-create-zkgtv" Dec 07 16:18:06 crc kubenswrapper[4716]: I1207 16:18:06.399617 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/708e4fff-969d-4753-9223-7ecbe84890a9-operator-scripts\") pod \"glance-92a4-account-create-update-pzqm6\" (UID: \"708e4fff-969d-4753-9223-7ecbe84890a9\") " pod="openstack/glance-92a4-account-create-update-pzqm6" Dec 07 16:18:06 crc kubenswrapper[4716]: I1207 16:18:06.399830 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cm8dl\" (UniqueName: \"kubernetes.io/projected/3a8a9543-3bd8-4f32-8e70-82eea76874df-kube-api-access-cm8dl\") pod \"glance-db-create-zkgtv\" (UID: \"3a8a9543-3bd8-4f32-8e70-82eea76874df\") " pod="openstack/glance-db-create-zkgtv" Dec 07 16:18:06 crc kubenswrapper[4716]: I1207 16:18:06.400489 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3a8a9543-3bd8-4f32-8e70-82eea76874df-operator-scripts\") pod \"glance-db-create-zkgtv\" (UID: \"3a8a9543-3bd8-4f32-8e70-82eea76874df\") " pod="openstack/glance-db-create-zkgtv" Dec 07 16:18:06 crc kubenswrapper[4716]: I1207 16:18:06.400742 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/708e4fff-969d-4753-9223-7ecbe84890a9-operator-scripts\") pod \"glance-92a4-account-create-update-pzqm6\" (UID: \"708e4fff-969d-4753-9223-7ecbe84890a9\") " pod="openstack/glance-92a4-account-create-update-pzqm6" Dec 07 16:18:06 crc kubenswrapper[4716]: I1207 16:18:06.422636 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cm8dl\" (UniqueName: \"kubernetes.io/projected/3a8a9543-3bd8-4f32-8e70-82eea76874df-kube-api-access-cm8dl\") pod \"glance-db-create-zkgtv\" (UID: \"3a8a9543-3bd8-4f32-8e70-82eea76874df\") " pod="openstack/glance-db-create-zkgtv" Dec 07 16:18:06 crc kubenswrapper[4716]: I1207 16:18:06.423986 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zcwrc\" (UniqueName: \"kubernetes.io/projected/708e4fff-969d-4753-9223-7ecbe84890a9-kube-api-access-zcwrc\") pod \"glance-92a4-account-create-update-pzqm6\" (UID: \"708e4fff-969d-4753-9223-7ecbe84890a9\") " pod="openstack/glance-92a4-account-create-update-pzqm6" Dec 07 16:18:06 crc kubenswrapper[4716]: I1207 16:18:06.449676 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-zkgtv" Dec 07 16:18:06 crc kubenswrapper[4716]: I1207 16:18:06.455181 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-2fj94" Dec 07 16:18:06 crc kubenswrapper[4716]: I1207 16:18:06.501658 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0a2f2b59-b71e-416d-bdc6-2a69ed43b920-dns-svc\") pod \"0a2f2b59-b71e-416d-bdc6-2a69ed43b920\" (UID: \"0a2f2b59-b71e-416d-bdc6-2a69ed43b920\") " Dec 07 16:18:06 crc kubenswrapper[4716]: I1207 16:18:06.502179 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-82sq7\" (UniqueName: \"kubernetes.io/projected/0a2f2b59-b71e-416d-bdc6-2a69ed43b920-kube-api-access-82sq7\") pod \"0a2f2b59-b71e-416d-bdc6-2a69ed43b920\" (UID: \"0a2f2b59-b71e-416d-bdc6-2a69ed43b920\") " Dec 07 16:18:06 crc kubenswrapper[4716]: I1207 16:18:06.502206 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0a2f2b59-b71e-416d-bdc6-2a69ed43b920-ovsdbserver-nb\") pod \"0a2f2b59-b71e-416d-bdc6-2a69ed43b920\" (UID: \"0a2f2b59-b71e-416d-bdc6-2a69ed43b920\") " Dec 07 16:18:06 crc kubenswrapper[4716]: I1207 16:18:06.502264 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a2f2b59-b71e-416d-bdc6-2a69ed43b920-config\") pod \"0a2f2b59-b71e-416d-bdc6-2a69ed43b920\" (UID: \"0a2f2b59-b71e-416d-bdc6-2a69ed43b920\") " Dec 07 16:18:06 crc kubenswrapper[4716]: I1207 16:18:06.510641 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a2f2b59-b71e-416d-bdc6-2a69ed43b920-kube-api-access-82sq7" (OuterVolumeSpecName: "kube-api-access-82sq7") pod "0a2f2b59-b71e-416d-bdc6-2a69ed43b920" (UID: "0a2f2b59-b71e-416d-bdc6-2a69ed43b920"). InnerVolumeSpecName "kube-api-access-82sq7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:18:06 crc kubenswrapper[4716]: I1207 16:18:06.524604 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a2f2b59-b71e-416d-bdc6-2a69ed43b920-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0a2f2b59-b71e-416d-bdc6-2a69ed43b920" (UID: "0a2f2b59-b71e-416d-bdc6-2a69ed43b920"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:18:06 crc kubenswrapper[4716]: I1207 16:18:06.530138 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a2f2b59-b71e-416d-bdc6-2a69ed43b920-config" (OuterVolumeSpecName: "config") pod "0a2f2b59-b71e-416d-bdc6-2a69ed43b920" (UID: "0a2f2b59-b71e-416d-bdc6-2a69ed43b920"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:18:06 crc kubenswrapper[4716]: I1207 16:18:06.543967 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a2f2b59-b71e-416d-bdc6-2a69ed43b920-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0a2f2b59-b71e-416d-bdc6-2a69ed43b920" (UID: "0a2f2b59-b71e-416d-bdc6-2a69ed43b920"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:18:06 crc kubenswrapper[4716]: I1207 16:18:06.547062 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-92a4-account-create-update-pzqm6" Dec 07 16:18:06 crc kubenswrapper[4716]: I1207 16:18:06.604947 4716 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0a2f2b59-b71e-416d-bdc6-2a69ed43b920-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:06 crc kubenswrapper[4716]: I1207 16:18:06.604983 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-82sq7\" (UniqueName: \"kubernetes.io/projected/0a2f2b59-b71e-416d-bdc6-2a69ed43b920-kube-api-access-82sq7\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:06 crc kubenswrapper[4716]: I1207 16:18:06.605000 4716 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0a2f2b59-b71e-416d-bdc6-2a69ed43b920-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:06 crc kubenswrapper[4716]: I1207 16:18:06.605014 4716 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a2f2b59-b71e-416d-bdc6-2a69ed43b920-config\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:06 crc kubenswrapper[4716]: I1207 16:18:06.906748 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-zkgtv"] Dec 07 16:18:06 crc kubenswrapper[4716]: W1207 16:18:06.914455 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3a8a9543_3bd8_4f32_8e70_82eea76874df.slice/crio-f14e9fa0653cbd6ea824f3abf48488eda3f2788125c3ee776ad616581fb61b6b WatchSource:0}: Error finding container f14e9fa0653cbd6ea824f3abf48488eda3f2788125c3ee776ad616581fb61b6b: Status 404 returned error can't find the container with id f14e9fa0653cbd6ea824f3abf48488eda3f2788125c3ee776ad616581fb61b6b Dec 07 16:18:07 crc kubenswrapper[4716]: I1207 16:18:07.008220 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-92a4-account-create-update-pzqm6"] Dec 07 16:18:07 crc kubenswrapper[4716]: W1207 16:18:07.014528 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod708e4fff_969d_4753_9223_7ecbe84890a9.slice/crio-54f514be2bf681ff11feba56429478ec243c6c6533c5160ec7a24f3277c9da7e WatchSource:0}: Error finding container 54f514be2bf681ff11feba56429478ec243c6c6533c5160ec7a24f3277c9da7e: Status 404 returned error can't find the container with id 54f514be2bf681ff11feba56429478ec243c6c6533c5160ec7a24f3277c9da7e Dec 07 16:18:07 crc kubenswrapper[4716]: I1207 16:18:07.098303 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-zkgtv" event={"ID":"3a8a9543-3bd8-4f32-8e70-82eea76874df","Type":"ContainerStarted","Data":"f14e9fa0653cbd6ea824f3abf48488eda3f2788125c3ee776ad616581fb61b6b"} Dec 07 16:18:07 crc kubenswrapper[4716]: I1207 16:18:07.100851 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-tsr96" event={"ID":"61f22dc6-45b1-4307-8eee-fd493fc96408","Type":"ContainerStarted","Data":"8c8965356f476766caa898f00d666ba948d54ca797e2859835959799e1a859a0"} Dec 07 16:18:07 crc kubenswrapper[4716]: I1207 16:18:07.101828 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-86db49b7ff-tsr96" Dec 07 16:18:07 crc kubenswrapper[4716]: I1207 16:18:07.103163 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-2fj94" event={"ID":"0a2f2b59-b71e-416d-bdc6-2a69ed43b920","Type":"ContainerDied","Data":"197d58eb9607d2e4dc6ae0543d244d9cec5ce76c9a80a4b4030a580b56003f90"} Dec 07 16:18:07 crc kubenswrapper[4716]: I1207 16:18:07.103201 4716 scope.go:117] "RemoveContainer" containerID="1f791d3911a76ca9ff7033df05bef650990f28ec02c5858b69fc86f2a7f76d19" Dec 07 16:18:07 crc kubenswrapper[4716]: I1207 16:18:07.103302 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-2fj94" Dec 07 16:18:07 crc kubenswrapper[4716]: I1207 16:18:07.108919 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-92a4-account-create-update-pzqm6" event={"ID":"708e4fff-969d-4753-9223-7ecbe84890a9","Type":"ContainerStarted","Data":"54f514be2bf681ff11feba56429478ec243c6c6533c5160ec7a24f3277c9da7e"} Dec 07 16:18:07 crc kubenswrapper[4716]: I1207 16:18:07.118154 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"3ae6bded-8084-4e07-afcc-a7fafb718e3a","Type":"ContainerStarted","Data":"045ecdf6fd981186edba366fc98e1e2f653d76c2a117ffb6a5039d2fd23afd22"} Dec 07 16:18:07 crc kubenswrapper[4716]: I1207 16:18:07.118183 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-8phkg" podUID="fef1e486-acca-4cc4-b2b1-3b20270d418a" containerName="registry-server" containerID="cri-o://53639bea8fec33f2cf8928b14b9c0d9e9e956716b32781f207c37b1f9e3054b1" gracePeriod=2 Dec 07 16:18:07 crc kubenswrapper[4716]: I1207 16:18:07.132634 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-86db49b7ff-tsr96" podStartSLOduration=3.132618788 podStartE2EDuration="3.132618788s" podCreationTimestamp="2025-12-07 16:18:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:18:07.128876913 +0000 UTC m=+949.819161845" watchObservedRunningTime="2025-12-07 16:18:07.132618788 +0000 UTC m=+949.822903700" Dec 07 16:18:07 crc kubenswrapper[4716]: I1207 16:18:07.221472 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-2fj94"] Dec 07 16:18:07 crc kubenswrapper[4716]: I1207 16:18:07.244526 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-2fj94"] Dec 07 16:18:07 crc kubenswrapper[4716]: I1207 16:18:07.671171 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a2f2b59-b71e-416d-bdc6-2a69ed43b920" path="/var/lib/kubelet/pods/0a2f2b59-b71e-416d-bdc6-2a69ed43b920/volumes" Dec 07 16:18:07 crc kubenswrapper[4716]: I1207 16:18:07.840389 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8phkg" Dec 07 16:18:08 crc kubenswrapper[4716]: I1207 16:18:08.028093 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fef1e486-acca-4cc4-b2b1-3b20270d418a-catalog-content\") pod \"fef1e486-acca-4cc4-b2b1-3b20270d418a\" (UID: \"fef1e486-acca-4cc4-b2b1-3b20270d418a\") " Dec 07 16:18:08 crc kubenswrapper[4716]: I1207 16:18:08.028262 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zz9t8\" (UniqueName: \"kubernetes.io/projected/fef1e486-acca-4cc4-b2b1-3b20270d418a-kube-api-access-zz9t8\") pod \"fef1e486-acca-4cc4-b2b1-3b20270d418a\" (UID: \"fef1e486-acca-4cc4-b2b1-3b20270d418a\") " Dec 07 16:18:08 crc kubenswrapper[4716]: I1207 16:18:08.028312 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fef1e486-acca-4cc4-b2b1-3b20270d418a-utilities\") pod \"fef1e486-acca-4cc4-b2b1-3b20270d418a\" (UID: \"fef1e486-acca-4cc4-b2b1-3b20270d418a\") " Dec 07 16:18:08 crc kubenswrapper[4716]: I1207 16:18:08.029544 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fef1e486-acca-4cc4-b2b1-3b20270d418a-utilities" (OuterVolumeSpecName: "utilities") pod "fef1e486-acca-4cc4-b2b1-3b20270d418a" (UID: "fef1e486-acca-4cc4-b2b1-3b20270d418a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:18:08 crc kubenswrapper[4716]: I1207 16:18:08.034023 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fef1e486-acca-4cc4-b2b1-3b20270d418a-kube-api-access-zz9t8" (OuterVolumeSpecName: "kube-api-access-zz9t8") pod "fef1e486-acca-4cc4-b2b1-3b20270d418a" (UID: "fef1e486-acca-4cc4-b2b1-3b20270d418a"). InnerVolumeSpecName "kube-api-access-zz9t8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:18:08 crc kubenswrapper[4716]: I1207 16:18:08.080914 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fef1e486-acca-4cc4-b2b1-3b20270d418a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fef1e486-acca-4cc4-b2b1-3b20270d418a" (UID: "fef1e486-acca-4cc4-b2b1-3b20270d418a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:18:08 crc kubenswrapper[4716]: I1207 16:18:08.127109 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"3ae6bded-8084-4e07-afcc-a7fafb718e3a","Type":"ContainerStarted","Data":"5fac6b504cd9e6e7fcecbfaf2ce111e3ce0c3c292c94616689de1868e23836f7"} Dec 07 16:18:08 crc kubenswrapper[4716]: I1207 16:18:08.127151 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"3ae6bded-8084-4e07-afcc-a7fafb718e3a","Type":"ContainerStarted","Data":"5d6cd32d8a0c86206adaa34c9c102dc7529ac2e0bb90c226607f56aea326cfc1"} Dec 07 16:18:08 crc kubenswrapper[4716]: I1207 16:18:08.127260 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Dec 07 16:18:08 crc kubenswrapper[4716]: I1207 16:18:08.128899 4716 generic.go:334] "Generic (PLEG): container finished" podID="3a8a9543-3bd8-4f32-8e70-82eea76874df" containerID="30d3d72035887670236040e883c7c2e5b42e0a519c96902c5b92be3276ab017a" exitCode=0 Dec 07 16:18:08 crc kubenswrapper[4716]: I1207 16:18:08.128949 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-zkgtv" event={"ID":"3a8a9543-3bd8-4f32-8e70-82eea76874df","Type":"ContainerDied","Data":"30d3d72035887670236040e883c7c2e5b42e0a519c96902c5b92be3276ab017a"} Dec 07 16:18:08 crc kubenswrapper[4716]: I1207 16:18:08.129326 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zz9t8\" (UniqueName: \"kubernetes.io/projected/fef1e486-acca-4cc4-b2b1-3b20270d418a-kube-api-access-zz9t8\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:08 crc kubenswrapper[4716]: I1207 16:18:08.129374 4716 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fef1e486-acca-4cc4-b2b1-3b20270d418a-utilities\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:08 crc kubenswrapper[4716]: I1207 16:18:08.129385 4716 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fef1e486-acca-4cc4-b2b1-3b20270d418a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:08 crc kubenswrapper[4716]: I1207 16:18:08.139722 4716 generic.go:334] "Generic (PLEG): container finished" podID="fef1e486-acca-4cc4-b2b1-3b20270d418a" containerID="53639bea8fec33f2cf8928b14b9c0d9e9e956716b32781f207c37b1f9e3054b1" exitCode=0 Dec 07 16:18:08 crc kubenswrapper[4716]: I1207 16:18:08.139778 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8phkg" event={"ID":"fef1e486-acca-4cc4-b2b1-3b20270d418a","Type":"ContainerDied","Data":"53639bea8fec33f2cf8928b14b9c0d9e9e956716b32781f207c37b1f9e3054b1"} Dec 07 16:18:08 crc kubenswrapper[4716]: I1207 16:18:08.139804 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8phkg" event={"ID":"fef1e486-acca-4cc4-b2b1-3b20270d418a","Type":"ContainerDied","Data":"fa224d91f9560174d80dbe06302487f4a62c5c3b531a4fe0fca57b4953e7abae"} Dec 07 16:18:08 crc kubenswrapper[4716]: I1207 16:18:08.139826 4716 scope.go:117] "RemoveContainer" containerID="53639bea8fec33f2cf8928b14b9c0d9e9e956716b32781f207c37b1f9e3054b1" Dec 07 16:18:08 crc kubenswrapper[4716]: I1207 16:18:08.139918 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8phkg" Dec 07 16:18:08 crc kubenswrapper[4716]: I1207 16:18:08.145248 4716 generic.go:334] "Generic (PLEG): container finished" podID="708e4fff-969d-4753-9223-7ecbe84890a9" containerID="b5f5cd456586aa7da6430b4cbeb4d206921c062c830d0665cccb4657c7bcf4f5" exitCode=0 Dec 07 16:18:08 crc kubenswrapper[4716]: I1207 16:18:08.145312 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-92a4-account-create-update-pzqm6" event={"ID":"708e4fff-969d-4753-9223-7ecbe84890a9","Type":"ContainerDied","Data":"b5f5cd456586aa7da6430b4cbeb4d206921c062c830d0665cccb4657c7bcf4f5"} Dec 07 16:18:08 crc kubenswrapper[4716]: I1207 16:18:08.161897 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=1.889387315 podStartE2EDuration="3.161866415s" podCreationTimestamp="2025-12-07 16:18:05 +0000 UTC" firstStartedPulling="2025-12-07 16:18:06.201024712 +0000 UTC m=+948.891309634" lastFinishedPulling="2025-12-07 16:18:07.473503822 +0000 UTC m=+950.163788734" observedRunningTime="2025-12-07 16:18:08.150250752 +0000 UTC m=+950.840535674" watchObservedRunningTime="2025-12-07 16:18:08.161866415 +0000 UTC m=+950.852151337" Dec 07 16:18:08 crc kubenswrapper[4716]: I1207 16:18:08.193646 4716 scope.go:117] "RemoveContainer" containerID="22ee7211c116eb84a19b1d9da6c8e22a81b40c2415e1456a23b462893c1c4dc4" Dec 07 16:18:08 crc kubenswrapper[4716]: I1207 16:18:08.210193 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8phkg"] Dec 07 16:18:08 crc kubenswrapper[4716]: I1207 16:18:08.215267 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-8phkg"] Dec 07 16:18:08 crc kubenswrapper[4716]: I1207 16:18:08.218653 4716 scope.go:117] "RemoveContainer" containerID="1fd1e50c076ddfcac3629fa3995fb943661bf547ba40904201364e1f2d459134" Dec 07 16:18:08 crc kubenswrapper[4716]: I1207 16:18:08.236456 4716 scope.go:117] "RemoveContainer" containerID="53639bea8fec33f2cf8928b14b9c0d9e9e956716b32781f207c37b1f9e3054b1" Dec 07 16:18:08 crc kubenswrapper[4716]: E1207 16:18:08.236903 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53639bea8fec33f2cf8928b14b9c0d9e9e956716b32781f207c37b1f9e3054b1\": container with ID starting with 53639bea8fec33f2cf8928b14b9c0d9e9e956716b32781f207c37b1f9e3054b1 not found: ID does not exist" containerID="53639bea8fec33f2cf8928b14b9c0d9e9e956716b32781f207c37b1f9e3054b1" Dec 07 16:18:08 crc kubenswrapper[4716]: I1207 16:18:08.236958 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53639bea8fec33f2cf8928b14b9c0d9e9e956716b32781f207c37b1f9e3054b1"} err="failed to get container status \"53639bea8fec33f2cf8928b14b9c0d9e9e956716b32781f207c37b1f9e3054b1\": rpc error: code = NotFound desc = could not find container \"53639bea8fec33f2cf8928b14b9c0d9e9e956716b32781f207c37b1f9e3054b1\": container with ID starting with 53639bea8fec33f2cf8928b14b9c0d9e9e956716b32781f207c37b1f9e3054b1 not found: ID does not exist" Dec 07 16:18:08 crc kubenswrapper[4716]: I1207 16:18:08.236998 4716 scope.go:117] "RemoveContainer" containerID="22ee7211c116eb84a19b1d9da6c8e22a81b40c2415e1456a23b462893c1c4dc4" Dec 07 16:18:08 crc kubenswrapper[4716]: E1207 16:18:08.237444 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"22ee7211c116eb84a19b1d9da6c8e22a81b40c2415e1456a23b462893c1c4dc4\": container with ID starting with 22ee7211c116eb84a19b1d9da6c8e22a81b40c2415e1456a23b462893c1c4dc4 not found: ID does not exist" containerID="22ee7211c116eb84a19b1d9da6c8e22a81b40c2415e1456a23b462893c1c4dc4" Dec 07 16:18:08 crc kubenswrapper[4716]: I1207 16:18:08.237478 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"22ee7211c116eb84a19b1d9da6c8e22a81b40c2415e1456a23b462893c1c4dc4"} err="failed to get container status \"22ee7211c116eb84a19b1d9da6c8e22a81b40c2415e1456a23b462893c1c4dc4\": rpc error: code = NotFound desc = could not find container \"22ee7211c116eb84a19b1d9da6c8e22a81b40c2415e1456a23b462893c1c4dc4\": container with ID starting with 22ee7211c116eb84a19b1d9da6c8e22a81b40c2415e1456a23b462893c1c4dc4 not found: ID does not exist" Dec 07 16:18:08 crc kubenswrapper[4716]: I1207 16:18:08.237502 4716 scope.go:117] "RemoveContainer" containerID="1fd1e50c076ddfcac3629fa3995fb943661bf547ba40904201364e1f2d459134" Dec 07 16:18:08 crc kubenswrapper[4716]: E1207 16:18:08.237806 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1fd1e50c076ddfcac3629fa3995fb943661bf547ba40904201364e1f2d459134\": container with ID starting with 1fd1e50c076ddfcac3629fa3995fb943661bf547ba40904201364e1f2d459134 not found: ID does not exist" containerID="1fd1e50c076ddfcac3629fa3995fb943661bf547ba40904201364e1f2d459134" Dec 07 16:18:08 crc kubenswrapper[4716]: I1207 16:18:08.237843 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1fd1e50c076ddfcac3629fa3995fb943661bf547ba40904201364e1f2d459134"} err="failed to get container status \"1fd1e50c076ddfcac3629fa3995fb943661bf547ba40904201364e1f2d459134\": rpc error: code = NotFound desc = could not find container \"1fd1e50c076ddfcac3629fa3995fb943661bf547ba40904201364e1f2d459134\": container with ID starting with 1fd1e50c076ddfcac3629fa3995fb943661bf547ba40904201364e1f2d459134 not found: ID does not exist" Dec 07 16:18:09 crc kubenswrapper[4716]: I1207 16:18:09.501559 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-dnnsf" Dec 07 16:18:09 crc kubenswrapper[4716]: I1207 16:18:09.501951 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-dnnsf" Dec 07 16:18:09 crc kubenswrapper[4716]: I1207 16:18:09.555860 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-dnnsf" Dec 07 16:18:09 crc kubenswrapper[4716]: I1207 16:18:09.596322 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-92a4-account-create-update-pzqm6" Dec 07 16:18:09 crc kubenswrapper[4716]: I1207 16:18:09.600812 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-zkgtv" Dec 07 16:18:09 crc kubenswrapper[4716]: I1207 16:18:09.671447 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fef1e486-acca-4cc4-b2b1-3b20270d418a" path="/var/lib/kubelet/pods/fef1e486-acca-4cc4-b2b1-3b20270d418a/volumes" Dec 07 16:18:09 crc kubenswrapper[4716]: I1207 16:18:09.760123 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zcwrc\" (UniqueName: \"kubernetes.io/projected/708e4fff-969d-4753-9223-7ecbe84890a9-kube-api-access-zcwrc\") pod \"708e4fff-969d-4753-9223-7ecbe84890a9\" (UID: \"708e4fff-969d-4753-9223-7ecbe84890a9\") " Dec 07 16:18:09 crc kubenswrapper[4716]: I1207 16:18:09.760259 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3a8a9543-3bd8-4f32-8e70-82eea76874df-operator-scripts\") pod \"3a8a9543-3bd8-4f32-8e70-82eea76874df\" (UID: \"3a8a9543-3bd8-4f32-8e70-82eea76874df\") " Dec 07 16:18:09 crc kubenswrapper[4716]: I1207 16:18:09.760393 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/708e4fff-969d-4753-9223-7ecbe84890a9-operator-scripts\") pod \"708e4fff-969d-4753-9223-7ecbe84890a9\" (UID: \"708e4fff-969d-4753-9223-7ecbe84890a9\") " Dec 07 16:18:09 crc kubenswrapper[4716]: I1207 16:18:09.760470 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cm8dl\" (UniqueName: \"kubernetes.io/projected/3a8a9543-3bd8-4f32-8e70-82eea76874df-kube-api-access-cm8dl\") pod \"3a8a9543-3bd8-4f32-8e70-82eea76874df\" (UID: \"3a8a9543-3bd8-4f32-8e70-82eea76874df\") " Dec 07 16:18:09 crc kubenswrapper[4716]: I1207 16:18:09.761673 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3a8a9543-3bd8-4f32-8e70-82eea76874df-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3a8a9543-3bd8-4f32-8e70-82eea76874df" (UID: "3a8a9543-3bd8-4f32-8e70-82eea76874df"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:18:09 crc kubenswrapper[4716]: I1207 16:18:09.761869 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/708e4fff-969d-4753-9223-7ecbe84890a9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "708e4fff-969d-4753-9223-7ecbe84890a9" (UID: "708e4fff-969d-4753-9223-7ecbe84890a9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:18:09 crc kubenswrapper[4716]: I1207 16:18:09.767366 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/708e4fff-969d-4753-9223-7ecbe84890a9-kube-api-access-zcwrc" (OuterVolumeSpecName: "kube-api-access-zcwrc") pod "708e4fff-969d-4753-9223-7ecbe84890a9" (UID: "708e4fff-969d-4753-9223-7ecbe84890a9"). InnerVolumeSpecName "kube-api-access-zcwrc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:18:09 crc kubenswrapper[4716]: I1207 16:18:09.781853 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a8a9543-3bd8-4f32-8e70-82eea76874df-kube-api-access-cm8dl" (OuterVolumeSpecName: "kube-api-access-cm8dl") pod "3a8a9543-3bd8-4f32-8e70-82eea76874df" (UID: "3a8a9543-3bd8-4f32-8e70-82eea76874df"). InnerVolumeSpecName "kube-api-access-cm8dl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:18:09 crc kubenswrapper[4716]: I1207 16:18:09.862620 4716 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/708e4fff-969d-4753-9223-7ecbe84890a9-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:09 crc kubenswrapper[4716]: I1207 16:18:09.862672 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cm8dl\" (UniqueName: \"kubernetes.io/projected/3a8a9543-3bd8-4f32-8e70-82eea76874df-kube-api-access-cm8dl\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:09 crc kubenswrapper[4716]: I1207 16:18:09.862686 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zcwrc\" (UniqueName: \"kubernetes.io/projected/708e4fff-969d-4753-9223-7ecbe84890a9-kube-api-access-zcwrc\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:09 crc kubenswrapper[4716]: I1207 16:18:09.862695 4716 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3a8a9543-3bd8-4f32-8e70-82eea76874df-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:10 crc kubenswrapper[4716]: I1207 16:18:10.169674 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-zkgtv" event={"ID":"3a8a9543-3bd8-4f32-8e70-82eea76874df","Type":"ContainerDied","Data":"f14e9fa0653cbd6ea824f3abf48488eda3f2788125c3ee776ad616581fb61b6b"} Dec 07 16:18:10 crc kubenswrapper[4716]: I1207 16:18:10.169750 4716 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f14e9fa0653cbd6ea824f3abf48488eda3f2788125c3ee776ad616581fb61b6b" Dec 07 16:18:10 crc kubenswrapper[4716]: I1207 16:18:10.169911 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-zkgtv" Dec 07 16:18:10 crc kubenswrapper[4716]: I1207 16:18:10.171357 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-92a4-account-create-update-pzqm6" event={"ID":"708e4fff-969d-4753-9223-7ecbe84890a9","Type":"ContainerDied","Data":"54f514be2bf681ff11feba56429478ec243c6c6533c5160ec7a24f3277c9da7e"} Dec 07 16:18:10 crc kubenswrapper[4716]: I1207 16:18:10.171391 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-92a4-account-create-update-pzqm6" Dec 07 16:18:10 crc kubenswrapper[4716]: I1207 16:18:10.171404 4716 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="54f514be2bf681ff11feba56429478ec243c6c6533c5160ec7a24f3277c9da7e" Dec 07 16:18:10 crc kubenswrapper[4716]: I1207 16:18:10.248299 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-dnnsf" Dec 07 16:18:10 crc kubenswrapper[4716]: I1207 16:18:10.316030 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-c4d5d"] Dec 07 16:18:10 crc kubenswrapper[4716]: E1207 16:18:10.316432 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fef1e486-acca-4cc4-b2b1-3b20270d418a" containerName="extract-content" Dec 07 16:18:10 crc kubenswrapper[4716]: I1207 16:18:10.316451 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="fef1e486-acca-4cc4-b2b1-3b20270d418a" containerName="extract-content" Dec 07 16:18:10 crc kubenswrapper[4716]: E1207 16:18:10.316464 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a8a9543-3bd8-4f32-8e70-82eea76874df" containerName="mariadb-database-create" Dec 07 16:18:10 crc kubenswrapper[4716]: I1207 16:18:10.316472 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a8a9543-3bd8-4f32-8e70-82eea76874df" containerName="mariadb-database-create" Dec 07 16:18:10 crc kubenswrapper[4716]: E1207 16:18:10.316482 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a2f2b59-b71e-416d-bdc6-2a69ed43b920" containerName="init" Dec 07 16:18:10 crc kubenswrapper[4716]: I1207 16:18:10.316490 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a2f2b59-b71e-416d-bdc6-2a69ed43b920" containerName="init" Dec 07 16:18:10 crc kubenswrapper[4716]: E1207 16:18:10.316515 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fef1e486-acca-4cc4-b2b1-3b20270d418a" containerName="extract-utilities" Dec 07 16:18:10 crc kubenswrapper[4716]: I1207 16:18:10.316522 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="fef1e486-acca-4cc4-b2b1-3b20270d418a" containerName="extract-utilities" Dec 07 16:18:10 crc kubenswrapper[4716]: E1207 16:18:10.316536 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fef1e486-acca-4cc4-b2b1-3b20270d418a" containerName="registry-server" Dec 07 16:18:10 crc kubenswrapper[4716]: I1207 16:18:10.316543 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="fef1e486-acca-4cc4-b2b1-3b20270d418a" containerName="registry-server" Dec 07 16:18:10 crc kubenswrapper[4716]: E1207 16:18:10.316561 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="708e4fff-969d-4753-9223-7ecbe84890a9" containerName="mariadb-account-create-update" Dec 07 16:18:10 crc kubenswrapper[4716]: I1207 16:18:10.316570 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="708e4fff-969d-4753-9223-7ecbe84890a9" containerName="mariadb-account-create-update" Dec 07 16:18:10 crc kubenswrapper[4716]: I1207 16:18:10.316756 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a2f2b59-b71e-416d-bdc6-2a69ed43b920" containerName="init" Dec 07 16:18:10 crc kubenswrapper[4716]: I1207 16:18:10.316772 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a8a9543-3bd8-4f32-8e70-82eea76874df" containerName="mariadb-database-create" Dec 07 16:18:10 crc kubenswrapper[4716]: I1207 16:18:10.316788 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="708e4fff-969d-4753-9223-7ecbe84890a9" containerName="mariadb-account-create-update" Dec 07 16:18:10 crc kubenswrapper[4716]: I1207 16:18:10.316800 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="fef1e486-acca-4cc4-b2b1-3b20270d418a" containerName="registry-server" Dec 07 16:18:10 crc kubenswrapper[4716]: I1207 16:18:10.317441 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-c4d5d" Dec 07 16:18:10 crc kubenswrapper[4716]: I1207 16:18:10.326235 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-c4d5d"] Dec 07 16:18:10 crc kubenswrapper[4716]: I1207 16:18:10.417486 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-f5b6-account-create-update-b7hp7"] Dec 07 16:18:10 crc kubenswrapper[4716]: I1207 16:18:10.418563 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-f5b6-account-create-update-b7hp7" Dec 07 16:18:10 crc kubenswrapper[4716]: I1207 16:18:10.427696 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Dec 07 16:18:10 crc kubenswrapper[4716]: I1207 16:18:10.450774 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-f5b6-account-create-update-b7hp7"] Dec 07 16:18:10 crc kubenswrapper[4716]: I1207 16:18:10.475182 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1ce6b31a-2beb-4fc8-ba79-07cb3f4a7ec8-operator-scripts\") pod \"keystone-db-create-c4d5d\" (UID: \"1ce6b31a-2beb-4fc8-ba79-07cb3f4a7ec8\") " pod="openstack/keystone-db-create-c4d5d" Dec 07 16:18:10 crc kubenswrapper[4716]: I1207 16:18:10.475262 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x7zbj\" (UniqueName: \"kubernetes.io/projected/1ce6b31a-2beb-4fc8-ba79-07cb3f4a7ec8-kube-api-access-x7zbj\") pod \"keystone-db-create-c4d5d\" (UID: \"1ce6b31a-2beb-4fc8-ba79-07cb3f4a7ec8\") " pod="openstack/keystone-db-create-c4d5d" Dec 07 16:18:10 crc kubenswrapper[4716]: I1207 16:18:10.576933 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mc9m8\" (UniqueName: \"kubernetes.io/projected/db19870c-0848-4366-8e07-f6028a151c86-kube-api-access-mc9m8\") pod \"keystone-f5b6-account-create-update-b7hp7\" (UID: \"db19870c-0848-4366-8e07-f6028a151c86\") " pod="openstack/keystone-f5b6-account-create-update-b7hp7" Dec 07 16:18:10 crc kubenswrapper[4716]: I1207 16:18:10.578138 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1ce6b31a-2beb-4fc8-ba79-07cb3f4a7ec8-operator-scripts\") pod \"keystone-db-create-c4d5d\" (UID: \"1ce6b31a-2beb-4fc8-ba79-07cb3f4a7ec8\") " pod="openstack/keystone-db-create-c4d5d" Dec 07 16:18:10 crc kubenswrapper[4716]: I1207 16:18:10.578223 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x7zbj\" (UniqueName: \"kubernetes.io/projected/1ce6b31a-2beb-4fc8-ba79-07cb3f4a7ec8-kube-api-access-x7zbj\") pod \"keystone-db-create-c4d5d\" (UID: \"1ce6b31a-2beb-4fc8-ba79-07cb3f4a7ec8\") " pod="openstack/keystone-db-create-c4d5d" Dec 07 16:18:10 crc kubenswrapper[4716]: I1207 16:18:10.578293 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/db19870c-0848-4366-8e07-f6028a151c86-operator-scripts\") pod \"keystone-f5b6-account-create-update-b7hp7\" (UID: \"db19870c-0848-4366-8e07-f6028a151c86\") " pod="openstack/keystone-f5b6-account-create-update-b7hp7" Dec 07 16:18:10 crc kubenswrapper[4716]: I1207 16:18:10.578892 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1ce6b31a-2beb-4fc8-ba79-07cb3f4a7ec8-operator-scripts\") pod \"keystone-db-create-c4d5d\" (UID: \"1ce6b31a-2beb-4fc8-ba79-07cb3f4a7ec8\") " pod="openstack/keystone-db-create-c4d5d" Dec 07 16:18:10 crc kubenswrapper[4716]: I1207 16:18:10.603431 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x7zbj\" (UniqueName: \"kubernetes.io/projected/1ce6b31a-2beb-4fc8-ba79-07cb3f4a7ec8-kube-api-access-x7zbj\") pod \"keystone-db-create-c4d5d\" (UID: \"1ce6b31a-2beb-4fc8-ba79-07cb3f4a7ec8\") " pod="openstack/keystone-db-create-c4d5d" Dec 07 16:18:10 crc kubenswrapper[4716]: I1207 16:18:10.650019 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-c4d5d" Dec 07 16:18:10 crc kubenswrapper[4716]: I1207 16:18:10.680731 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/db19870c-0848-4366-8e07-f6028a151c86-operator-scripts\") pod \"keystone-f5b6-account-create-update-b7hp7\" (UID: \"db19870c-0848-4366-8e07-f6028a151c86\") " pod="openstack/keystone-f5b6-account-create-update-b7hp7" Dec 07 16:18:10 crc kubenswrapper[4716]: I1207 16:18:10.681372 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mc9m8\" (UniqueName: \"kubernetes.io/projected/db19870c-0848-4366-8e07-f6028a151c86-kube-api-access-mc9m8\") pod \"keystone-f5b6-account-create-update-b7hp7\" (UID: \"db19870c-0848-4366-8e07-f6028a151c86\") " pod="openstack/keystone-f5b6-account-create-update-b7hp7" Dec 07 16:18:10 crc kubenswrapper[4716]: I1207 16:18:10.681911 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/db19870c-0848-4366-8e07-f6028a151c86-operator-scripts\") pod \"keystone-f5b6-account-create-update-b7hp7\" (UID: \"db19870c-0848-4366-8e07-f6028a151c86\") " pod="openstack/keystone-f5b6-account-create-update-b7hp7" Dec 07 16:18:10 crc kubenswrapper[4716]: I1207 16:18:10.695208 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-7w7t8"] Dec 07 16:18:10 crc kubenswrapper[4716]: I1207 16:18:10.697537 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-7w7t8" Dec 07 16:18:10 crc kubenswrapper[4716]: I1207 16:18:10.703514 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mc9m8\" (UniqueName: \"kubernetes.io/projected/db19870c-0848-4366-8e07-f6028a151c86-kube-api-access-mc9m8\") pod \"keystone-f5b6-account-create-update-b7hp7\" (UID: \"db19870c-0848-4366-8e07-f6028a151c86\") " pod="openstack/keystone-f5b6-account-create-update-b7hp7" Dec 07 16:18:10 crc kubenswrapper[4716]: I1207 16:18:10.704250 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-7w7t8"] Dec 07 16:18:10 crc kubenswrapper[4716]: I1207 16:18:10.783343 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4dxzz\" (UniqueName: \"kubernetes.io/projected/9f2a2677-eb1a-40c1-a947-c3347f6c8c5a-kube-api-access-4dxzz\") pod \"placement-db-create-7w7t8\" (UID: \"9f2a2677-eb1a-40c1-a947-c3347f6c8c5a\") " pod="openstack/placement-db-create-7w7t8" Dec 07 16:18:10 crc kubenswrapper[4716]: I1207 16:18:10.783412 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9f2a2677-eb1a-40c1-a947-c3347f6c8c5a-operator-scripts\") pod \"placement-db-create-7w7t8\" (UID: \"9f2a2677-eb1a-40c1-a947-c3347f6c8c5a\") " pod="openstack/placement-db-create-7w7t8" Dec 07 16:18:10 crc kubenswrapper[4716]: I1207 16:18:10.789135 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-f5b6-account-create-update-b7hp7" Dec 07 16:18:10 crc kubenswrapper[4716]: I1207 16:18:10.789757 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-ce39-account-create-update-5f2ml"] Dec 07 16:18:10 crc kubenswrapper[4716]: I1207 16:18:10.790811 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-ce39-account-create-update-5f2ml" Dec 07 16:18:10 crc kubenswrapper[4716]: I1207 16:18:10.792766 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Dec 07 16:18:10 crc kubenswrapper[4716]: I1207 16:18:10.803656 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-ce39-account-create-update-5f2ml"] Dec 07 16:18:10 crc kubenswrapper[4716]: I1207 16:18:10.884802 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9f2a2677-eb1a-40c1-a947-c3347f6c8c5a-operator-scripts\") pod \"placement-db-create-7w7t8\" (UID: \"9f2a2677-eb1a-40c1-a947-c3347f6c8c5a\") " pod="openstack/placement-db-create-7w7t8" Dec 07 16:18:10 crc kubenswrapper[4716]: I1207 16:18:10.884870 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1c2f879d-d624-4537-9ab4-f4076614d82e-operator-scripts\") pod \"placement-ce39-account-create-update-5f2ml\" (UID: \"1c2f879d-d624-4537-9ab4-f4076614d82e\") " pod="openstack/placement-ce39-account-create-update-5f2ml" Dec 07 16:18:10 crc kubenswrapper[4716]: I1207 16:18:10.885301 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swsbh\" (UniqueName: \"kubernetes.io/projected/1c2f879d-d624-4537-9ab4-f4076614d82e-kube-api-access-swsbh\") pod \"placement-ce39-account-create-update-5f2ml\" (UID: \"1c2f879d-d624-4537-9ab4-f4076614d82e\") " pod="openstack/placement-ce39-account-create-update-5f2ml" Dec 07 16:18:10 crc kubenswrapper[4716]: I1207 16:18:10.885357 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4dxzz\" (UniqueName: \"kubernetes.io/projected/9f2a2677-eb1a-40c1-a947-c3347f6c8c5a-kube-api-access-4dxzz\") pod \"placement-db-create-7w7t8\" (UID: \"9f2a2677-eb1a-40c1-a947-c3347f6c8c5a\") " pod="openstack/placement-db-create-7w7t8" Dec 07 16:18:10 crc kubenswrapper[4716]: I1207 16:18:10.886155 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9f2a2677-eb1a-40c1-a947-c3347f6c8c5a-operator-scripts\") pod \"placement-db-create-7w7t8\" (UID: \"9f2a2677-eb1a-40c1-a947-c3347f6c8c5a\") " pod="openstack/placement-db-create-7w7t8" Dec 07 16:18:10 crc kubenswrapper[4716]: I1207 16:18:10.906034 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4dxzz\" (UniqueName: \"kubernetes.io/projected/9f2a2677-eb1a-40c1-a947-c3347f6c8c5a-kube-api-access-4dxzz\") pod \"placement-db-create-7w7t8\" (UID: \"9f2a2677-eb1a-40c1-a947-c3347f6c8c5a\") " pod="openstack/placement-db-create-7w7t8" Dec 07 16:18:10 crc kubenswrapper[4716]: I1207 16:18:10.920362 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dnnsf"] Dec 07 16:18:10 crc kubenswrapper[4716]: I1207 16:18:10.985994 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swsbh\" (UniqueName: \"kubernetes.io/projected/1c2f879d-d624-4537-9ab4-f4076614d82e-kube-api-access-swsbh\") pod \"placement-ce39-account-create-update-5f2ml\" (UID: \"1c2f879d-d624-4537-9ab4-f4076614d82e\") " pod="openstack/placement-ce39-account-create-update-5f2ml" Dec 07 16:18:10 crc kubenswrapper[4716]: I1207 16:18:10.986363 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1c2f879d-d624-4537-9ab4-f4076614d82e-operator-scripts\") pod \"placement-ce39-account-create-update-5f2ml\" (UID: \"1c2f879d-d624-4537-9ab4-f4076614d82e\") " pod="openstack/placement-ce39-account-create-update-5f2ml" Dec 07 16:18:10 crc kubenswrapper[4716]: I1207 16:18:10.987035 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1c2f879d-d624-4537-9ab4-f4076614d82e-operator-scripts\") pod \"placement-ce39-account-create-update-5f2ml\" (UID: \"1c2f879d-d624-4537-9ab4-f4076614d82e\") " pod="openstack/placement-ce39-account-create-update-5f2ml" Dec 07 16:18:11 crc kubenswrapper[4716]: I1207 16:18:11.002508 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swsbh\" (UniqueName: \"kubernetes.io/projected/1c2f879d-d624-4537-9ab4-f4076614d82e-kube-api-access-swsbh\") pod \"placement-ce39-account-create-update-5f2ml\" (UID: \"1c2f879d-d624-4537-9ab4-f4076614d82e\") " pod="openstack/placement-ce39-account-create-update-5f2ml" Dec 07 16:18:11 crc kubenswrapper[4716]: I1207 16:18:11.056186 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-7w7t8" Dec 07 16:18:11 crc kubenswrapper[4716]: I1207 16:18:11.118044 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-ce39-account-create-update-5f2ml" Dec 07 16:18:11 crc kubenswrapper[4716]: I1207 16:18:11.181204 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-c4d5d"] Dec 07 16:18:11 crc kubenswrapper[4716]: I1207 16:18:11.259298 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-f5b6-account-create-update-b7hp7"] Dec 07 16:18:11 crc kubenswrapper[4716]: I1207 16:18:11.280025 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-gq62d"] Dec 07 16:18:11 crc kubenswrapper[4716]: I1207 16:18:11.281159 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-gq62d" Dec 07 16:18:11 crc kubenswrapper[4716]: I1207 16:18:11.283634 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Dec 07 16:18:11 crc kubenswrapper[4716]: I1207 16:18:11.283855 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-2hfps" Dec 07 16:18:11 crc kubenswrapper[4716]: I1207 16:18:11.292738 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a17893a0-50ef-4d05-8aaf-c6a9e3f8d20d-config-data\") pod \"glance-db-sync-gq62d\" (UID: \"a17893a0-50ef-4d05-8aaf-c6a9e3f8d20d\") " pod="openstack/glance-db-sync-gq62d" Dec 07 16:18:11 crc kubenswrapper[4716]: I1207 16:18:11.292793 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sp8kc\" (UniqueName: \"kubernetes.io/projected/a17893a0-50ef-4d05-8aaf-c6a9e3f8d20d-kube-api-access-sp8kc\") pod \"glance-db-sync-gq62d\" (UID: \"a17893a0-50ef-4d05-8aaf-c6a9e3f8d20d\") " pod="openstack/glance-db-sync-gq62d" Dec 07 16:18:11 crc kubenswrapper[4716]: I1207 16:18:11.292829 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a17893a0-50ef-4d05-8aaf-c6a9e3f8d20d-db-sync-config-data\") pod \"glance-db-sync-gq62d\" (UID: \"a17893a0-50ef-4d05-8aaf-c6a9e3f8d20d\") " pod="openstack/glance-db-sync-gq62d" Dec 07 16:18:11 crc kubenswrapper[4716]: I1207 16:18:11.292878 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a17893a0-50ef-4d05-8aaf-c6a9e3f8d20d-combined-ca-bundle\") pod \"glance-db-sync-gq62d\" (UID: \"a17893a0-50ef-4d05-8aaf-c6a9e3f8d20d\") " pod="openstack/glance-db-sync-gq62d" Dec 07 16:18:11 crc kubenswrapper[4716]: I1207 16:18:11.307447 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-gq62d"] Dec 07 16:18:11 crc kubenswrapper[4716]: I1207 16:18:11.393638 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a17893a0-50ef-4d05-8aaf-c6a9e3f8d20d-combined-ca-bundle\") pod \"glance-db-sync-gq62d\" (UID: \"a17893a0-50ef-4d05-8aaf-c6a9e3f8d20d\") " pod="openstack/glance-db-sync-gq62d" Dec 07 16:18:11 crc kubenswrapper[4716]: I1207 16:18:11.394139 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a17893a0-50ef-4d05-8aaf-c6a9e3f8d20d-config-data\") pod \"glance-db-sync-gq62d\" (UID: \"a17893a0-50ef-4d05-8aaf-c6a9e3f8d20d\") " pod="openstack/glance-db-sync-gq62d" Dec 07 16:18:11 crc kubenswrapper[4716]: I1207 16:18:11.394238 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sp8kc\" (UniqueName: \"kubernetes.io/projected/a17893a0-50ef-4d05-8aaf-c6a9e3f8d20d-kube-api-access-sp8kc\") pod \"glance-db-sync-gq62d\" (UID: \"a17893a0-50ef-4d05-8aaf-c6a9e3f8d20d\") " pod="openstack/glance-db-sync-gq62d" Dec 07 16:18:11 crc kubenswrapper[4716]: I1207 16:18:11.394355 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a17893a0-50ef-4d05-8aaf-c6a9e3f8d20d-db-sync-config-data\") pod \"glance-db-sync-gq62d\" (UID: \"a17893a0-50ef-4d05-8aaf-c6a9e3f8d20d\") " pod="openstack/glance-db-sync-gq62d" Dec 07 16:18:11 crc kubenswrapper[4716]: I1207 16:18:11.398842 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a17893a0-50ef-4d05-8aaf-c6a9e3f8d20d-combined-ca-bundle\") pod \"glance-db-sync-gq62d\" (UID: \"a17893a0-50ef-4d05-8aaf-c6a9e3f8d20d\") " pod="openstack/glance-db-sync-gq62d" Dec 07 16:18:11 crc kubenswrapper[4716]: I1207 16:18:11.400475 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a17893a0-50ef-4d05-8aaf-c6a9e3f8d20d-config-data\") pod \"glance-db-sync-gq62d\" (UID: \"a17893a0-50ef-4d05-8aaf-c6a9e3f8d20d\") " pod="openstack/glance-db-sync-gq62d" Dec 07 16:18:11 crc kubenswrapper[4716]: I1207 16:18:11.403137 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a17893a0-50ef-4d05-8aaf-c6a9e3f8d20d-db-sync-config-data\") pod \"glance-db-sync-gq62d\" (UID: \"a17893a0-50ef-4d05-8aaf-c6a9e3f8d20d\") " pod="openstack/glance-db-sync-gq62d" Dec 07 16:18:11 crc kubenswrapper[4716]: I1207 16:18:11.415867 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sp8kc\" (UniqueName: \"kubernetes.io/projected/a17893a0-50ef-4d05-8aaf-c6a9e3f8d20d-kube-api-access-sp8kc\") pod \"glance-db-sync-gq62d\" (UID: \"a17893a0-50ef-4d05-8aaf-c6a9e3f8d20d\") " pod="openstack/glance-db-sync-gq62d" Dec 07 16:18:11 crc kubenswrapper[4716]: I1207 16:18:11.512755 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-7w7t8"] Dec 07 16:18:11 crc kubenswrapper[4716]: W1207 16:18:11.522728 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9f2a2677_eb1a_40c1_a947_c3347f6c8c5a.slice/crio-616cb9cb90f8b90ac477f676661e63812b79055fa3dda227c7a9f37a024a8465 WatchSource:0}: Error finding container 616cb9cb90f8b90ac477f676661e63812b79055fa3dda227c7a9f37a024a8465: Status 404 returned error can't find the container with id 616cb9cb90f8b90ac477f676661e63812b79055fa3dda227c7a9f37a024a8465 Dec 07 16:18:11 crc kubenswrapper[4716]: I1207 16:18:11.613106 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-gq62d" Dec 07 16:18:11 crc kubenswrapper[4716]: W1207 16:18:11.624569 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1c2f879d_d624_4537_9ab4_f4076614d82e.slice/crio-3c8b3281a14fc08f10c40ea9ea4a1f68669481fb09f47334b9c08045f8422fe4 WatchSource:0}: Error finding container 3c8b3281a14fc08f10c40ea9ea4a1f68669481fb09f47334b9c08045f8422fe4: Status 404 returned error can't find the container with id 3c8b3281a14fc08f10c40ea9ea4a1f68669481fb09f47334b9c08045f8422fe4 Dec 07 16:18:11 crc kubenswrapper[4716]: I1207 16:18:11.628624 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-ce39-account-create-update-5f2ml"] Dec 07 16:18:12 crc kubenswrapper[4716]: I1207 16:18:12.189679 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-f5b6-account-create-update-b7hp7" event={"ID":"db19870c-0848-4366-8e07-f6028a151c86","Type":"ContainerStarted","Data":"510e1c2305c4e0f7b4156bbad7d847d8690ba86d7fd7e2cd035b8d2aaa5a1421"} Dec 07 16:18:12 crc kubenswrapper[4716]: I1207 16:18:12.190060 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-f5b6-account-create-update-b7hp7" event={"ID":"db19870c-0848-4366-8e07-f6028a151c86","Type":"ContainerStarted","Data":"e1c5b50c560d2a9661cfe8b19ac4030ef8ea65998a8d6695aa5335ae7fed5d38"} Dec 07 16:18:12 crc kubenswrapper[4716]: I1207 16:18:12.192013 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-7w7t8" event={"ID":"9f2a2677-eb1a-40c1-a947-c3347f6c8c5a","Type":"ContainerStarted","Data":"97495ae9bc006e7a09396ed08e1ff1a2212153c8be2a793c8ea9ff8c6182ddf1"} Dec 07 16:18:12 crc kubenswrapper[4716]: I1207 16:18:12.192038 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-7w7t8" event={"ID":"9f2a2677-eb1a-40c1-a947-c3347f6c8c5a","Type":"ContainerStarted","Data":"616cb9cb90f8b90ac477f676661e63812b79055fa3dda227c7a9f37a024a8465"} Dec 07 16:18:12 crc kubenswrapper[4716]: I1207 16:18:12.194440 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-ce39-account-create-update-5f2ml" event={"ID":"1c2f879d-d624-4537-9ab4-f4076614d82e","Type":"ContainerStarted","Data":"7ee46a0bc54b0c8ca6cfa63abfeb8f724ab05bbea382bf6025caf8c59b98af4e"} Dec 07 16:18:12 crc kubenswrapper[4716]: I1207 16:18:12.194470 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-ce39-account-create-update-5f2ml" event={"ID":"1c2f879d-d624-4537-9ab4-f4076614d82e","Type":"ContainerStarted","Data":"3c8b3281a14fc08f10c40ea9ea4a1f68669481fb09f47334b9c08045f8422fe4"} Dec 07 16:18:12 crc kubenswrapper[4716]: I1207 16:18:12.196490 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-c4d5d" event={"ID":"1ce6b31a-2beb-4fc8-ba79-07cb3f4a7ec8","Type":"ContainerStarted","Data":"2dd4f0eccf36088a3dd7414d172b860996761216771cc27440379f05438017da"} Dec 07 16:18:12 crc kubenswrapper[4716]: I1207 16:18:12.196583 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-c4d5d" event={"ID":"1ce6b31a-2beb-4fc8-ba79-07cb3f4a7ec8","Type":"ContainerStarted","Data":"2a40ba0d85c8b58afd371de1ed1a9e6b93070ada34643f3e612ddae1a2245f91"} Dec 07 16:18:12 crc kubenswrapper[4716]: I1207 16:18:12.196939 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-dnnsf" podUID="9b870e60-acde-4d14-be71-bee41e0c53bd" containerName="registry-server" containerID="cri-o://4e61d65f1c66e3295fbc6e46d16e92fe2d66a11ff291bd2c1fff6fb4654b99d1" gracePeriod=2 Dec 07 16:18:12 crc kubenswrapper[4716]: I1207 16:18:12.209802 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-f5b6-account-create-update-b7hp7" podStartSLOduration=2.2097825110000002 podStartE2EDuration="2.209782511s" podCreationTimestamp="2025-12-07 16:18:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:18:12.20615401 +0000 UTC m=+954.896438922" watchObservedRunningTime="2025-12-07 16:18:12.209782511 +0000 UTC m=+954.900067423" Dec 07 16:18:12 crc kubenswrapper[4716]: I1207 16:18:12.229759 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-gq62d"] Dec 07 16:18:12 crc kubenswrapper[4716]: I1207 16:18:12.239674 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-create-c4d5d" podStartSLOduration=2.239646544 podStartE2EDuration="2.239646544s" podCreationTimestamp="2025-12-07 16:18:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:18:12.228439951 +0000 UTC m=+954.918724853" watchObservedRunningTime="2025-12-07 16:18:12.239646544 +0000 UTC m=+954.929931456" Dec 07 16:18:12 crc kubenswrapper[4716]: I1207 16:18:12.276539 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-create-7w7t8" podStartSLOduration=2.276513512 podStartE2EDuration="2.276513512s" podCreationTimestamp="2025-12-07 16:18:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:18:12.270565676 +0000 UTC m=+954.960850598" watchObservedRunningTime="2025-12-07 16:18:12.276513512 +0000 UTC m=+954.966798494" Dec 07 16:18:12 crc kubenswrapper[4716]: I1207 16:18:12.633066 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 07 16:18:12 crc kubenswrapper[4716]: I1207 16:18:12.695467 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-tsr96"] Dec 07 16:18:12 crc kubenswrapper[4716]: I1207 16:18:12.695808 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-86db49b7ff-tsr96" podUID="61f22dc6-45b1-4307-8eee-fd493fc96408" containerName="dnsmasq-dns" containerID="cri-o://8c8965356f476766caa898f00d666ba948d54ca797e2859835959799e1a859a0" gracePeriod=10 Dec 07 16:18:12 crc kubenswrapper[4716]: I1207 16:18:12.698381 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-86db49b7ff-tsr96" Dec 07 16:18:12 crc kubenswrapper[4716]: I1207 16:18:12.781622 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-698758b865-nj89r"] Dec 07 16:18:12 crc kubenswrapper[4716]: I1207 16:18:12.782892 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-nj89r" Dec 07 16:18:12 crc kubenswrapper[4716]: I1207 16:18:12.890147 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-nj89r"] Dec 07 16:18:12 crc kubenswrapper[4716]: I1207 16:18:12.937251 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8wkjm\" (UniqueName: \"kubernetes.io/projected/666f9fee-767e-40d6-ac44-5abe506e3266-kube-api-access-8wkjm\") pod \"dnsmasq-dns-698758b865-nj89r\" (UID: \"666f9fee-767e-40d6-ac44-5abe506e3266\") " pod="openstack/dnsmasq-dns-698758b865-nj89r" Dec 07 16:18:12 crc kubenswrapper[4716]: I1207 16:18:12.937309 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/666f9fee-767e-40d6-ac44-5abe506e3266-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-nj89r\" (UID: \"666f9fee-767e-40d6-ac44-5abe506e3266\") " pod="openstack/dnsmasq-dns-698758b865-nj89r" Dec 07 16:18:12 crc kubenswrapper[4716]: I1207 16:18:12.937349 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/666f9fee-767e-40d6-ac44-5abe506e3266-dns-svc\") pod \"dnsmasq-dns-698758b865-nj89r\" (UID: \"666f9fee-767e-40d6-ac44-5abe506e3266\") " pod="openstack/dnsmasq-dns-698758b865-nj89r" Dec 07 16:18:12 crc kubenswrapper[4716]: I1207 16:18:12.937393 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/666f9fee-767e-40d6-ac44-5abe506e3266-config\") pod \"dnsmasq-dns-698758b865-nj89r\" (UID: \"666f9fee-767e-40d6-ac44-5abe506e3266\") " pod="openstack/dnsmasq-dns-698758b865-nj89r" Dec 07 16:18:12 crc kubenswrapper[4716]: I1207 16:18:12.937426 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/666f9fee-767e-40d6-ac44-5abe506e3266-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-nj89r\" (UID: \"666f9fee-767e-40d6-ac44-5abe506e3266\") " pod="openstack/dnsmasq-dns-698758b865-nj89r" Dec 07 16:18:13 crc kubenswrapper[4716]: I1207 16:18:13.038411 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/666f9fee-767e-40d6-ac44-5abe506e3266-config\") pod \"dnsmasq-dns-698758b865-nj89r\" (UID: \"666f9fee-767e-40d6-ac44-5abe506e3266\") " pod="openstack/dnsmasq-dns-698758b865-nj89r" Dec 07 16:18:13 crc kubenswrapper[4716]: I1207 16:18:13.038472 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/666f9fee-767e-40d6-ac44-5abe506e3266-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-nj89r\" (UID: \"666f9fee-767e-40d6-ac44-5abe506e3266\") " pod="openstack/dnsmasq-dns-698758b865-nj89r" Dec 07 16:18:13 crc kubenswrapper[4716]: I1207 16:18:13.038513 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8wkjm\" (UniqueName: \"kubernetes.io/projected/666f9fee-767e-40d6-ac44-5abe506e3266-kube-api-access-8wkjm\") pod \"dnsmasq-dns-698758b865-nj89r\" (UID: \"666f9fee-767e-40d6-ac44-5abe506e3266\") " pod="openstack/dnsmasq-dns-698758b865-nj89r" Dec 07 16:18:13 crc kubenswrapper[4716]: I1207 16:18:13.038541 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/666f9fee-767e-40d6-ac44-5abe506e3266-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-nj89r\" (UID: \"666f9fee-767e-40d6-ac44-5abe506e3266\") " pod="openstack/dnsmasq-dns-698758b865-nj89r" Dec 07 16:18:13 crc kubenswrapper[4716]: I1207 16:18:13.038578 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/666f9fee-767e-40d6-ac44-5abe506e3266-dns-svc\") pod \"dnsmasq-dns-698758b865-nj89r\" (UID: \"666f9fee-767e-40d6-ac44-5abe506e3266\") " pod="openstack/dnsmasq-dns-698758b865-nj89r" Dec 07 16:18:13 crc kubenswrapper[4716]: I1207 16:18:13.039368 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/666f9fee-767e-40d6-ac44-5abe506e3266-dns-svc\") pod \"dnsmasq-dns-698758b865-nj89r\" (UID: \"666f9fee-767e-40d6-ac44-5abe506e3266\") " pod="openstack/dnsmasq-dns-698758b865-nj89r" Dec 07 16:18:13 crc kubenswrapper[4716]: I1207 16:18:13.039894 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/666f9fee-767e-40d6-ac44-5abe506e3266-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-nj89r\" (UID: \"666f9fee-767e-40d6-ac44-5abe506e3266\") " pod="openstack/dnsmasq-dns-698758b865-nj89r" Dec 07 16:18:13 crc kubenswrapper[4716]: I1207 16:18:13.039368 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/666f9fee-767e-40d6-ac44-5abe506e3266-config\") pod \"dnsmasq-dns-698758b865-nj89r\" (UID: \"666f9fee-767e-40d6-ac44-5abe506e3266\") " pod="openstack/dnsmasq-dns-698758b865-nj89r" Dec 07 16:18:13 crc kubenswrapper[4716]: I1207 16:18:13.040139 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/666f9fee-767e-40d6-ac44-5abe506e3266-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-nj89r\" (UID: \"666f9fee-767e-40d6-ac44-5abe506e3266\") " pod="openstack/dnsmasq-dns-698758b865-nj89r" Dec 07 16:18:13 crc kubenswrapper[4716]: I1207 16:18:13.067441 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8wkjm\" (UniqueName: \"kubernetes.io/projected/666f9fee-767e-40d6-ac44-5abe506e3266-kube-api-access-8wkjm\") pod \"dnsmasq-dns-698758b865-nj89r\" (UID: \"666f9fee-767e-40d6-ac44-5abe506e3266\") " pod="openstack/dnsmasq-dns-698758b865-nj89r" Dec 07 16:18:13 crc kubenswrapper[4716]: I1207 16:18:13.142832 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-nj89r" Dec 07 16:18:13 crc kubenswrapper[4716]: I1207 16:18:13.209636 4716 generic.go:334] "Generic (PLEG): container finished" podID="61f22dc6-45b1-4307-8eee-fd493fc96408" containerID="8c8965356f476766caa898f00d666ba948d54ca797e2859835959799e1a859a0" exitCode=0 Dec 07 16:18:13 crc kubenswrapper[4716]: I1207 16:18:13.209685 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-tsr96" event={"ID":"61f22dc6-45b1-4307-8eee-fd493fc96408","Type":"ContainerDied","Data":"8c8965356f476766caa898f00d666ba948d54ca797e2859835959799e1a859a0"} Dec 07 16:18:13 crc kubenswrapper[4716]: I1207 16:18:13.220847 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-gq62d" event={"ID":"a17893a0-50ef-4d05-8aaf-c6a9e3f8d20d","Type":"ContainerStarted","Data":"3c57a7abf3687f77612f42313bd678454e1b3ea2a3f21ab4e09349fb5536b2bb"} Dec 07 16:18:13 crc kubenswrapper[4716]: I1207 16:18:13.231303 4716 generic.go:334] "Generic (PLEG): container finished" podID="9b870e60-acde-4d14-be71-bee41e0c53bd" containerID="4e61d65f1c66e3295fbc6e46d16e92fe2d66a11ff291bd2c1fff6fb4654b99d1" exitCode=0 Dec 07 16:18:13 crc kubenswrapper[4716]: I1207 16:18:13.231992 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dnnsf" event={"ID":"9b870e60-acde-4d14-be71-bee41e0c53bd","Type":"ContainerDied","Data":"4e61d65f1c66e3295fbc6e46d16e92fe2d66a11ff291bd2c1fff6fb4654b99d1"} Dec 07 16:18:13 crc kubenswrapper[4716]: I1207 16:18:13.253208 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-ce39-account-create-update-5f2ml" podStartSLOduration=3.2531872330000002 podStartE2EDuration="3.253187233s" podCreationTimestamp="2025-12-07 16:18:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:18:13.250094807 +0000 UTC m=+955.940379719" watchObservedRunningTime="2025-12-07 16:18:13.253187233 +0000 UTC m=+955.943472145" Dec 07 16:18:13 crc kubenswrapper[4716]: I1207 16:18:13.640328 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-nj89r"] Dec 07 16:18:13 crc kubenswrapper[4716]: I1207 16:18:13.930863 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Dec 07 16:18:13 crc kubenswrapper[4716]: I1207 16:18:13.939861 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 07 16:18:13 crc kubenswrapper[4716]: I1207 16:18:13.945739 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Dec 07 16:18:13 crc kubenswrapper[4716]: I1207 16:18:13.946248 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Dec 07 16:18:13 crc kubenswrapper[4716]: I1207 16:18:13.946789 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Dec 07 16:18:13 crc kubenswrapper[4716]: I1207 16:18:13.947074 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-64jz6" Dec 07 16:18:13 crc kubenswrapper[4716]: I1207 16:18:13.952615 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 07 16:18:14 crc kubenswrapper[4716]: I1207 16:18:14.057338 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"swift-storage-0\" (UID: \"a714bbaa-9626-4b49-92ae-2abb029408d3\") " pod="openstack/swift-storage-0" Dec 07 16:18:14 crc kubenswrapper[4716]: I1207 16:18:14.057417 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hmfwp\" (UniqueName: \"kubernetes.io/projected/a714bbaa-9626-4b49-92ae-2abb029408d3-kube-api-access-hmfwp\") pod \"swift-storage-0\" (UID: \"a714bbaa-9626-4b49-92ae-2abb029408d3\") " pod="openstack/swift-storage-0" Dec 07 16:18:14 crc kubenswrapper[4716]: I1207 16:18:14.057444 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a714bbaa-9626-4b49-92ae-2abb029408d3-etc-swift\") pod \"swift-storage-0\" (UID: \"a714bbaa-9626-4b49-92ae-2abb029408d3\") " pod="openstack/swift-storage-0" Dec 07 16:18:14 crc kubenswrapper[4716]: I1207 16:18:14.057512 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/a714bbaa-9626-4b49-92ae-2abb029408d3-cache\") pod \"swift-storage-0\" (UID: \"a714bbaa-9626-4b49-92ae-2abb029408d3\") " pod="openstack/swift-storage-0" Dec 07 16:18:14 crc kubenswrapper[4716]: I1207 16:18:14.057669 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/a714bbaa-9626-4b49-92ae-2abb029408d3-lock\") pod \"swift-storage-0\" (UID: \"a714bbaa-9626-4b49-92ae-2abb029408d3\") " pod="openstack/swift-storage-0" Dec 07 16:18:14 crc kubenswrapper[4716]: I1207 16:18:14.158678 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/a714bbaa-9626-4b49-92ae-2abb029408d3-lock\") pod \"swift-storage-0\" (UID: \"a714bbaa-9626-4b49-92ae-2abb029408d3\") " pod="openstack/swift-storage-0" Dec 07 16:18:14 crc kubenswrapper[4716]: I1207 16:18:14.158725 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"swift-storage-0\" (UID: \"a714bbaa-9626-4b49-92ae-2abb029408d3\") " pod="openstack/swift-storage-0" Dec 07 16:18:14 crc kubenswrapper[4716]: I1207 16:18:14.158752 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hmfwp\" (UniqueName: \"kubernetes.io/projected/a714bbaa-9626-4b49-92ae-2abb029408d3-kube-api-access-hmfwp\") pod \"swift-storage-0\" (UID: \"a714bbaa-9626-4b49-92ae-2abb029408d3\") " pod="openstack/swift-storage-0" Dec 07 16:18:14 crc kubenswrapper[4716]: I1207 16:18:14.158773 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a714bbaa-9626-4b49-92ae-2abb029408d3-etc-swift\") pod \"swift-storage-0\" (UID: \"a714bbaa-9626-4b49-92ae-2abb029408d3\") " pod="openstack/swift-storage-0" Dec 07 16:18:14 crc kubenswrapper[4716]: I1207 16:18:14.158796 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/a714bbaa-9626-4b49-92ae-2abb029408d3-cache\") pod \"swift-storage-0\" (UID: \"a714bbaa-9626-4b49-92ae-2abb029408d3\") " pod="openstack/swift-storage-0" Dec 07 16:18:14 crc kubenswrapper[4716]: I1207 16:18:14.159385 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/a714bbaa-9626-4b49-92ae-2abb029408d3-cache\") pod \"swift-storage-0\" (UID: \"a714bbaa-9626-4b49-92ae-2abb029408d3\") " pod="openstack/swift-storage-0" Dec 07 16:18:14 crc kubenswrapper[4716]: I1207 16:18:14.159595 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/a714bbaa-9626-4b49-92ae-2abb029408d3-lock\") pod \"swift-storage-0\" (UID: \"a714bbaa-9626-4b49-92ae-2abb029408d3\") " pod="openstack/swift-storage-0" Dec 07 16:18:14 crc kubenswrapper[4716]: I1207 16:18:14.159879 4716 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"swift-storage-0\" (UID: \"a714bbaa-9626-4b49-92ae-2abb029408d3\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/swift-storage-0" Dec 07 16:18:14 crc kubenswrapper[4716]: E1207 16:18:14.160263 4716 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 07 16:18:14 crc kubenswrapper[4716]: E1207 16:18:14.160300 4716 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 07 16:18:14 crc kubenswrapper[4716]: E1207 16:18:14.160364 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a714bbaa-9626-4b49-92ae-2abb029408d3-etc-swift podName:a714bbaa-9626-4b49-92ae-2abb029408d3 nodeName:}" failed. No retries permitted until 2025-12-07 16:18:14.660337767 +0000 UTC m=+957.350622679 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/a714bbaa-9626-4b49-92ae-2abb029408d3-etc-swift") pod "swift-storage-0" (UID: "a714bbaa-9626-4b49-92ae-2abb029408d3") : configmap "swift-ring-files" not found Dec 07 16:18:14 crc kubenswrapper[4716]: I1207 16:18:14.161508 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dnnsf" Dec 07 16:18:14 crc kubenswrapper[4716]: I1207 16:18:14.183275 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hmfwp\" (UniqueName: \"kubernetes.io/projected/a714bbaa-9626-4b49-92ae-2abb029408d3-kube-api-access-hmfwp\") pod \"swift-storage-0\" (UID: \"a714bbaa-9626-4b49-92ae-2abb029408d3\") " pod="openstack/swift-storage-0" Dec 07 16:18:14 crc kubenswrapper[4716]: I1207 16:18:14.188816 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"swift-storage-0\" (UID: \"a714bbaa-9626-4b49-92ae-2abb029408d3\") " pod="openstack/swift-storage-0" Dec 07 16:18:14 crc kubenswrapper[4716]: I1207 16:18:14.240844 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-nj89r" event={"ID":"666f9fee-767e-40d6-ac44-5abe506e3266","Type":"ContainerStarted","Data":"79480ac09df21f06646f2b716c79bb5b771610a22a6eb43478fbb6cb63124dad"} Dec 07 16:18:14 crc kubenswrapper[4716]: I1207 16:18:14.243343 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dnnsf" event={"ID":"9b870e60-acde-4d14-be71-bee41e0c53bd","Type":"ContainerDied","Data":"d8b43898810fc04e8bd86c6076b7dc4ab287555ff3642b84fa38d231c67fd6cb"} Dec 07 16:18:14 crc kubenswrapper[4716]: I1207 16:18:14.243382 4716 scope.go:117] "RemoveContainer" containerID="4e61d65f1c66e3295fbc6e46d16e92fe2d66a11ff291bd2c1fff6fb4654b99d1" Dec 07 16:18:14 crc kubenswrapper[4716]: I1207 16:18:14.243490 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dnnsf" Dec 07 16:18:14 crc kubenswrapper[4716]: I1207 16:18:14.259721 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b870e60-acde-4d14-be71-bee41e0c53bd-catalog-content\") pod \"9b870e60-acde-4d14-be71-bee41e0c53bd\" (UID: \"9b870e60-acde-4d14-be71-bee41e0c53bd\") " Dec 07 16:18:14 crc kubenswrapper[4716]: I1207 16:18:14.260134 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b870e60-acde-4d14-be71-bee41e0c53bd-utilities\") pod \"9b870e60-acde-4d14-be71-bee41e0c53bd\" (UID: \"9b870e60-acde-4d14-be71-bee41e0c53bd\") " Dec 07 16:18:14 crc kubenswrapper[4716]: I1207 16:18:14.260261 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zx7cc\" (UniqueName: \"kubernetes.io/projected/9b870e60-acde-4d14-be71-bee41e0c53bd-kube-api-access-zx7cc\") pod \"9b870e60-acde-4d14-be71-bee41e0c53bd\" (UID: \"9b870e60-acde-4d14-be71-bee41e0c53bd\") " Dec 07 16:18:14 crc kubenswrapper[4716]: I1207 16:18:14.260933 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9b870e60-acde-4d14-be71-bee41e0c53bd-utilities" (OuterVolumeSpecName: "utilities") pod "9b870e60-acde-4d14-be71-bee41e0c53bd" (UID: "9b870e60-acde-4d14-be71-bee41e0c53bd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:18:14 crc kubenswrapper[4716]: I1207 16:18:14.264471 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b870e60-acde-4d14-be71-bee41e0c53bd-kube-api-access-zx7cc" (OuterVolumeSpecName: "kube-api-access-zx7cc") pod "9b870e60-acde-4d14-be71-bee41e0c53bd" (UID: "9b870e60-acde-4d14-be71-bee41e0c53bd"). InnerVolumeSpecName "kube-api-access-zx7cc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:18:14 crc kubenswrapper[4716]: I1207 16:18:14.307429 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9b870e60-acde-4d14-be71-bee41e0c53bd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9b870e60-acde-4d14-be71-bee41e0c53bd" (UID: "9b870e60-acde-4d14-be71-bee41e0c53bd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:18:14 crc kubenswrapper[4716]: I1207 16:18:14.362344 4716 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b870e60-acde-4d14-be71-bee41e0c53bd-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:14 crc kubenswrapper[4716]: I1207 16:18:14.362380 4716 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b870e60-acde-4d14-be71-bee41e0c53bd-utilities\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:14 crc kubenswrapper[4716]: I1207 16:18:14.362390 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zx7cc\" (UniqueName: \"kubernetes.io/projected/9b870e60-acde-4d14-be71-bee41e0c53bd-kube-api-access-zx7cc\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:14 crc kubenswrapper[4716]: I1207 16:18:14.386262 4716 scope.go:117] "RemoveContainer" containerID="a388ac549a0d8efe723c4e06f780c553909ba14958e61473baa94c584fd2b5ba" Dec 07 16:18:14 crc kubenswrapper[4716]: I1207 16:18:14.418203 4716 scope.go:117] "RemoveContainer" containerID="aedfaa33234d36f11ded958b11e6191ff13fb623d0c2d3262e693feca0e1a0a9" Dec 07 16:18:14 crc kubenswrapper[4716]: I1207 16:18:14.566161 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-jgrbf"] Dec 07 16:18:14 crc kubenswrapper[4716]: E1207 16:18:14.567035 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b870e60-acde-4d14-be71-bee41e0c53bd" containerName="extract-utilities" Dec 07 16:18:14 crc kubenswrapper[4716]: I1207 16:18:14.567061 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b870e60-acde-4d14-be71-bee41e0c53bd" containerName="extract-utilities" Dec 07 16:18:14 crc kubenswrapper[4716]: E1207 16:18:14.567116 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b870e60-acde-4d14-be71-bee41e0c53bd" containerName="registry-server" Dec 07 16:18:14 crc kubenswrapper[4716]: I1207 16:18:14.567125 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b870e60-acde-4d14-be71-bee41e0c53bd" containerName="registry-server" Dec 07 16:18:14 crc kubenswrapper[4716]: E1207 16:18:14.567203 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b870e60-acde-4d14-be71-bee41e0c53bd" containerName="extract-content" Dec 07 16:18:14 crc kubenswrapper[4716]: I1207 16:18:14.567214 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b870e60-acde-4d14-be71-bee41e0c53bd" containerName="extract-content" Dec 07 16:18:14 crc kubenswrapper[4716]: I1207 16:18:14.567581 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b870e60-acde-4d14-be71-bee41e0c53bd" containerName="registry-server" Dec 07 16:18:14 crc kubenswrapper[4716]: I1207 16:18:14.568803 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-jgrbf" Dec 07 16:18:14 crc kubenswrapper[4716]: I1207 16:18:14.582119 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-jgrbf"] Dec 07 16:18:14 crc kubenswrapper[4716]: I1207 16:18:14.584915 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 07 16:18:14 crc kubenswrapper[4716]: I1207 16:18:14.585297 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Dec 07 16:18:14 crc kubenswrapper[4716]: I1207 16:18:14.585438 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Dec 07 16:18:14 crc kubenswrapper[4716]: I1207 16:18:14.628728 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dnnsf"] Dec 07 16:18:14 crc kubenswrapper[4716]: I1207 16:18:14.636104 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-dnnsf"] Dec 07 16:18:14 crc kubenswrapper[4716]: I1207 16:18:14.670673 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2c371669-443a-492e-ad04-ab79ae978e5f-dispersionconf\") pod \"swift-ring-rebalance-jgrbf\" (UID: \"2c371669-443a-492e-ad04-ab79ae978e5f\") " pod="openstack/swift-ring-rebalance-jgrbf" Dec 07 16:18:14 crc kubenswrapper[4716]: I1207 16:18:14.670746 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2c371669-443a-492e-ad04-ab79ae978e5f-scripts\") pod \"swift-ring-rebalance-jgrbf\" (UID: \"2c371669-443a-492e-ad04-ab79ae978e5f\") " pod="openstack/swift-ring-rebalance-jgrbf" Dec 07 16:18:14 crc kubenswrapper[4716]: I1207 16:18:14.670797 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9khz\" (UniqueName: \"kubernetes.io/projected/2c371669-443a-492e-ad04-ab79ae978e5f-kube-api-access-t9khz\") pod \"swift-ring-rebalance-jgrbf\" (UID: \"2c371669-443a-492e-ad04-ab79ae978e5f\") " pod="openstack/swift-ring-rebalance-jgrbf" Dec 07 16:18:14 crc kubenswrapper[4716]: I1207 16:18:14.670821 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2c371669-443a-492e-ad04-ab79ae978e5f-ring-data-devices\") pod \"swift-ring-rebalance-jgrbf\" (UID: \"2c371669-443a-492e-ad04-ab79ae978e5f\") " pod="openstack/swift-ring-rebalance-jgrbf" Dec 07 16:18:14 crc kubenswrapper[4716]: I1207 16:18:14.670849 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2c371669-443a-492e-ad04-ab79ae978e5f-etc-swift\") pod \"swift-ring-rebalance-jgrbf\" (UID: \"2c371669-443a-492e-ad04-ab79ae978e5f\") " pod="openstack/swift-ring-rebalance-jgrbf" Dec 07 16:18:14 crc kubenswrapper[4716]: I1207 16:18:14.670869 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c371669-443a-492e-ad04-ab79ae978e5f-combined-ca-bundle\") pod \"swift-ring-rebalance-jgrbf\" (UID: \"2c371669-443a-492e-ad04-ab79ae978e5f\") " pod="openstack/swift-ring-rebalance-jgrbf" Dec 07 16:18:14 crc kubenswrapper[4716]: I1207 16:18:14.670919 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a714bbaa-9626-4b49-92ae-2abb029408d3-etc-swift\") pod \"swift-storage-0\" (UID: \"a714bbaa-9626-4b49-92ae-2abb029408d3\") " pod="openstack/swift-storage-0" Dec 07 16:18:14 crc kubenswrapper[4716]: I1207 16:18:14.670957 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2c371669-443a-492e-ad04-ab79ae978e5f-swiftconf\") pod \"swift-ring-rebalance-jgrbf\" (UID: \"2c371669-443a-492e-ad04-ab79ae978e5f\") " pod="openstack/swift-ring-rebalance-jgrbf" Dec 07 16:18:14 crc kubenswrapper[4716]: E1207 16:18:14.671179 4716 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 07 16:18:14 crc kubenswrapper[4716]: E1207 16:18:14.671193 4716 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 07 16:18:14 crc kubenswrapper[4716]: E1207 16:18:14.671233 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a714bbaa-9626-4b49-92ae-2abb029408d3-etc-swift podName:a714bbaa-9626-4b49-92ae-2abb029408d3 nodeName:}" failed. No retries permitted until 2025-12-07 16:18:15.671219812 +0000 UTC m=+958.361504714 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/a714bbaa-9626-4b49-92ae-2abb029408d3-etc-swift") pod "swift-storage-0" (UID: "a714bbaa-9626-4b49-92ae-2abb029408d3") : configmap "swift-ring-files" not found Dec 07 16:18:14 crc kubenswrapper[4716]: I1207 16:18:14.771958 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2c371669-443a-492e-ad04-ab79ae978e5f-etc-swift\") pod \"swift-ring-rebalance-jgrbf\" (UID: \"2c371669-443a-492e-ad04-ab79ae978e5f\") " pod="openstack/swift-ring-rebalance-jgrbf" Dec 07 16:18:14 crc kubenswrapper[4716]: I1207 16:18:14.772017 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c371669-443a-492e-ad04-ab79ae978e5f-combined-ca-bundle\") pod \"swift-ring-rebalance-jgrbf\" (UID: \"2c371669-443a-492e-ad04-ab79ae978e5f\") " pod="openstack/swift-ring-rebalance-jgrbf" Dec 07 16:18:14 crc kubenswrapper[4716]: I1207 16:18:14.772106 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2c371669-443a-492e-ad04-ab79ae978e5f-swiftconf\") pod \"swift-ring-rebalance-jgrbf\" (UID: \"2c371669-443a-492e-ad04-ab79ae978e5f\") " pod="openstack/swift-ring-rebalance-jgrbf" Dec 07 16:18:14 crc kubenswrapper[4716]: I1207 16:18:14.772128 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2c371669-443a-492e-ad04-ab79ae978e5f-dispersionconf\") pod \"swift-ring-rebalance-jgrbf\" (UID: \"2c371669-443a-492e-ad04-ab79ae978e5f\") " pod="openstack/swift-ring-rebalance-jgrbf" Dec 07 16:18:14 crc kubenswrapper[4716]: I1207 16:18:14.772165 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2c371669-443a-492e-ad04-ab79ae978e5f-scripts\") pod \"swift-ring-rebalance-jgrbf\" (UID: \"2c371669-443a-492e-ad04-ab79ae978e5f\") " pod="openstack/swift-ring-rebalance-jgrbf" Dec 07 16:18:14 crc kubenswrapper[4716]: I1207 16:18:14.772212 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9khz\" (UniqueName: \"kubernetes.io/projected/2c371669-443a-492e-ad04-ab79ae978e5f-kube-api-access-t9khz\") pod \"swift-ring-rebalance-jgrbf\" (UID: \"2c371669-443a-492e-ad04-ab79ae978e5f\") " pod="openstack/swift-ring-rebalance-jgrbf" Dec 07 16:18:14 crc kubenswrapper[4716]: I1207 16:18:14.772235 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2c371669-443a-492e-ad04-ab79ae978e5f-ring-data-devices\") pod \"swift-ring-rebalance-jgrbf\" (UID: \"2c371669-443a-492e-ad04-ab79ae978e5f\") " pod="openstack/swift-ring-rebalance-jgrbf" Dec 07 16:18:14 crc kubenswrapper[4716]: I1207 16:18:14.772907 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2c371669-443a-492e-ad04-ab79ae978e5f-ring-data-devices\") pod \"swift-ring-rebalance-jgrbf\" (UID: \"2c371669-443a-492e-ad04-ab79ae978e5f\") " pod="openstack/swift-ring-rebalance-jgrbf" Dec 07 16:18:14 crc kubenswrapper[4716]: I1207 16:18:14.773791 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2c371669-443a-492e-ad04-ab79ae978e5f-etc-swift\") pod \"swift-ring-rebalance-jgrbf\" (UID: \"2c371669-443a-492e-ad04-ab79ae978e5f\") " pod="openstack/swift-ring-rebalance-jgrbf" Dec 07 16:18:14 crc kubenswrapper[4716]: I1207 16:18:14.775758 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2c371669-443a-492e-ad04-ab79ae978e5f-scripts\") pod \"swift-ring-rebalance-jgrbf\" (UID: \"2c371669-443a-492e-ad04-ab79ae978e5f\") " pod="openstack/swift-ring-rebalance-jgrbf" Dec 07 16:18:14 crc kubenswrapper[4716]: I1207 16:18:14.777136 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c371669-443a-492e-ad04-ab79ae978e5f-combined-ca-bundle\") pod \"swift-ring-rebalance-jgrbf\" (UID: \"2c371669-443a-492e-ad04-ab79ae978e5f\") " pod="openstack/swift-ring-rebalance-jgrbf" Dec 07 16:18:14 crc kubenswrapper[4716]: I1207 16:18:14.777591 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2c371669-443a-492e-ad04-ab79ae978e5f-swiftconf\") pod \"swift-ring-rebalance-jgrbf\" (UID: \"2c371669-443a-492e-ad04-ab79ae978e5f\") " pod="openstack/swift-ring-rebalance-jgrbf" Dec 07 16:18:14 crc kubenswrapper[4716]: I1207 16:18:14.778201 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2c371669-443a-492e-ad04-ab79ae978e5f-dispersionconf\") pod \"swift-ring-rebalance-jgrbf\" (UID: \"2c371669-443a-492e-ad04-ab79ae978e5f\") " pod="openstack/swift-ring-rebalance-jgrbf" Dec 07 16:18:14 crc kubenswrapper[4716]: I1207 16:18:14.789465 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9khz\" (UniqueName: \"kubernetes.io/projected/2c371669-443a-492e-ad04-ab79ae978e5f-kube-api-access-t9khz\") pod \"swift-ring-rebalance-jgrbf\" (UID: \"2c371669-443a-492e-ad04-ab79ae978e5f\") " pod="openstack/swift-ring-rebalance-jgrbf" Dec 07 16:18:14 crc kubenswrapper[4716]: I1207 16:18:14.894069 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-jgrbf" Dec 07 16:18:15 crc kubenswrapper[4716]: I1207 16:18:15.138486 4716 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-86db49b7ff-tsr96" podUID="61f22dc6-45b1-4307-8eee-fd493fc96408" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.111:5353: connect: connection refused" Dec 07 16:18:15 crc kubenswrapper[4716]: I1207 16:18:15.404096 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-jgrbf"] Dec 07 16:18:15 crc kubenswrapper[4716]: I1207 16:18:15.675222 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9b870e60-acde-4d14-be71-bee41e0c53bd" path="/var/lib/kubelet/pods/9b870e60-acde-4d14-be71-bee41e0c53bd/volumes" Dec 07 16:18:15 crc kubenswrapper[4716]: I1207 16:18:15.686232 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a714bbaa-9626-4b49-92ae-2abb029408d3-etc-swift\") pod \"swift-storage-0\" (UID: \"a714bbaa-9626-4b49-92ae-2abb029408d3\") " pod="openstack/swift-storage-0" Dec 07 16:18:15 crc kubenswrapper[4716]: E1207 16:18:15.686486 4716 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 07 16:18:15 crc kubenswrapper[4716]: E1207 16:18:15.686505 4716 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 07 16:18:15 crc kubenswrapper[4716]: E1207 16:18:15.686547 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a714bbaa-9626-4b49-92ae-2abb029408d3-etc-swift podName:a714bbaa-9626-4b49-92ae-2abb029408d3 nodeName:}" failed. No retries permitted until 2025-12-07 16:18:17.686531661 +0000 UTC m=+960.376816573 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/a714bbaa-9626-4b49-92ae-2abb029408d3-etc-swift") pod "swift-storage-0" (UID: "a714bbaa-9626-4b49-92ae-2abb029408d3") : configmap "swift-ring-files" not found Dec 07 16:18:16 crc kubenswrapper[4716]: I1207 16:18:16.232384 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-tsr96" Dec 07 16:18:16 crc kubenswrapper[4716]: I1207 16:18:16.268463 4716 generic.go:334] "Generic (PLEG): container finished" podID="1ce6b31a-2beb-4fc8-ba79-07cb3f4a7ec8" containerID="2dd4f0eccf36088a3dd7414d172b860996761216771cc27440379f05438017da" exitCode=0 Dec 07 16:18:16 crc kubenswrapper[4716]: I1207 16:18:16.268534 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-c4d5d" event={"ID":"1ce6b31a-2beb-4fc8-ba79-07cb3f4a7ec8","Type":"ContainerDied","Data":"2dd4f0eccf36088a3dd7414d172b860996761216771cc27440379f05438017da"} Dec 07 16:18:16 crc kubenswrapper[4716]: I1207 16:18:16.271752 4716 generic.go:334] "Generic (PLEG): container finished" podID="666f9fee-767e-40d6-ac44-5abe506e3266" containerID="f5ea6b68fe720e1798d5e831a08cd1ed5dbe813349bb3eb0aad5c6346dd997cf" exitCode=0 Dec 07 16:18:16 crc kubenswrapper[4716]: I1207 16:18:16.271830 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-nj89r" event={"ID":"666f9fee-767e-40d6-ac44-5abe506e3266","Type":"ContainerDied","Data":"f5ea6b68fe720e1798d5e831a08cd1ed5dbe813349bb3eb0aad5c6346dd997cf"} Dec 07 16:18:16 crc kubenswrapper[4716]: I1207 16:18:16.273506 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-jgrbf" event={"ID":"2c371669-443a-492e-ad04-ab79ae978e5f","Type":"ContainerStarted","Data":"07d85826576be1e43cfa0fb3bbaedbd81b85e1e60da9e98f7702be82cf998308"} Dec 07 16:18:16 crc kubenswrapper[4716]: I1207 16:18:16.277610 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-tsr96" event={"ID":"61f22dc6-45b1-4307-8eee-fd493fc96408","Type":"ContainerDied","Data":"86af27eb39f5d5e4e2abd2ef9fc47d0c36e0f61f95a94e4e99f102b13a975d29"} Dec 07 16:18:16 crc kubenswrapper[4716]: I1207 16:18:16.277649 4716 scope.go:117] "RemoveContainer" containerID="8c8965356f476766caa898f00d666ba948d54ca797e2859835959799e1a859a0" Dec 07 16:18:16 crc kubenswrapper[4716]: I1207 16:18:16.277758 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-tsr96" Dec 07 16:18:16 crc kubenswrapper[4716]: I1207 16:18:16.286289 4716 generic.go:334] "Generic (PLEG): container finished" podID="db19870c-0848-4366-8e07-f6028a151c86" containerID="510e1c2305c4e0f7b4156bbad7d847d8690ba86d7fd7e2cd035b8d2aaa5a1421" exitCode=0 Dec 07 16:18:16 crc kubenswrapper[4716]: I1207 16:18:16.286416 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-f5b6-account-create-update-b7hp7" event={"ID":"db19870c-0848-4366-8e07-f6028a151c86","Type":"ContainerDied","Data":"510e1c2305c4e0f7b4156bbad7d847d8690ba86d7fd7e2cd035b8d2aaa5a1421"} Dec 07 16:18:16 crc kubenswrapper[4716]: I1207 16:18:16.310844 4716 generic.go:334] "Generic (PLEG): container finished" podID="9f2a2677-eb1a-40c1-a947-c3347f6c8c5a" containerID="97495ae9bc006e7a09396ed08e1ff1a2212153c8be2a793c8ea9ff8c6182ddf1" exitCode=0 Dec 07 16:18:16 crc kubenswrapper[4716]: I1207 16:18:16.311007 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-7w7t8" event={"ID":"9f2a2677-eb1a-40c1-a947-c3347f6c8c5a","Type":"ContainerDied","Data":"97495ae9bc006e7a09396ed08e1ff1a2212153c8be2a793c8ea9ff8c6182ddf1"} Dec 07 16:18:16 crc kubenswrapper[4716]: I1207 16:18:16.326722 4716 generic.go:334] "Generic (PLEG): container finished" podID="1c2f879d-d624-4537-9ab4-f4076614d82e" containerID="7ee46a0bc54b0c8ca6cfa63abfeb8f724ab05bbea382bf6025caf8c59b98af4e" exitCode=0 Dec 07 16:18:16 crc kubenswrapper[4716]: I1207 16:18:16.326785 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-ce39-account-create-update-5f2ml" event={"ID":"1c2f879d-d624-4537-9ab4-f4076614d82e","Type":"ContainerDied","Data":"7ee46a0bc54b0c8ca6cfa63abfeb8f724ab05bbea382bf6025caf8c59b98af4e"} Dec 07 16:18:16 crc kubenswrapper[4716]: I1207 16:18:16.334512 4716 scope.go:117] "RemoveContainer" containerID="426ee91ecd278e5884292d74ef2d1931368271bc3d735ed09ca634e3000eb534" Dec 07 16:18:16 crc kubenswrapper[4716]: I1207 16:18:16.406283 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/61f22dc6-45b1-4307-8eee-fd493fc96408-dns-svc\") pod \"61f22dc6-45b1-4307-8eee-fd493fc96408\" (UID: \"61f22dc6-45b1-4307-8eee-fd493fc96408\") " Dec 07 16:18:16 crc kubenswrapper[4716]: I1207 16:18:16.406554 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/61f22dc6-45b1-4307-8eee-fd493fc96408-ovsdbserver-nb\") pod \"61f22dc6-45b1-4307-8eee-fd493fc96408\" (UID: \"61f22dc6-45b1-4307-8eee-fd493fc96408\") " Dec 07 16:18:16 crc kubenswrapper[4716]: I1207 16:18:16.406612 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c9wjv\" (UniqueName: \"kubernetes.io/projected/61f22dc6-45b1-4307-8eee-fd493fc96408-kube-api-access-c9wjv\") pod \"61f22dc6-45b1-4307-8eee-fd493fc96408\" (UID: \"61f22dc6-45b1-4307-8eee-fd493fc96408\") " Dec 07 16:18:16 crc kubenswrapper[4716]: I1207 16:18:16.406758 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61f22dc6-45b1-4307-8eee-fd493fc96408-config\") pod \"61f22dc6-45b1-4307-8eee-fd493fc96408\" (UID: \"61f22dc6-45b1-4307-8eee-fd493fc96408\") " Dec 07 16:18:16 crc kubenswrapper[4716]: I1207 16:18:16.406787 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/61f22dc6-45b1-4307-8eee-fd493fc96408-ovsdbserver-sb\") pod \"61f22dc6-45b1-4307-8eee-fd493fc96408\" (UID: \"61f22dc6-45b1-4307-8eee-fd493fc96408\") " Dec 07 16:18:16 crc kubenswrapper[4716]: I1207 16:18:16.412792 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61f22dc6-45b1-4307-8eee-fd493fc96408-kube-api-access-c9wjv" (OuterVolumeSpecName: "kube-api-access-c9wjv") pod "61f22dc6-45b1-4307-8eee-fd493fc96408" (UID: "61f22dc6-45b1-4307-8eee-fd493fc96408"). InnerVolumeSpecName "kube-api-access-c9wjv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:18:16 crc kubenswrapper[4716]: I1207 16:18:16.452092 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/61f22dc6-45b1-4307-8eee-fd493fc96408-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "61f22dc6-45b1-4307-8eee-fd493fc96408" (UID: "61f22dc6-45b1-4307-8eee-fd493fc96408"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:18:16 crc kubenswrapper[4716]: I1207 16:18:16.452232 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/61f22dc6-45b1-4307-8eee-fd493fc96408-config" (OuterVolumeSpecName: "config") pod "61f22dc6-45b1-4307-8eee-fd493fc96408" (UID: "61f22dc6-45b1-4307-8eee-fd493fc96408"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:18:16 crc kubenswrapper[4716]: I1207 16:18:16.476201 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/61f22dc6-45b1-4307-8eee-fd493fc96408-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "61f22dc6-45b1-4307-8eee-fd493fc96408" (UID: "61f22dc6-45b1-4307-8eee-fd493fc96408"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:18:16 crc kubenswrapper[4716]: I1207 16:18:16.479020 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/61f22dc6-45b1-4307-8eee-fd493fc96408-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "61f22dc6-45b1-4307-8eee-fd493fc96408" (UID: "61f22dc6-45b1-4307-8eee-fd493fc96408"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:18:16 crc kubenswrapper[4716]: I1207 16:18:16.508693 4716 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61f22dc6-45b1-4307-8eee-fd493fc96408-config\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:16 crc kubenswrapper[4716]: I1207 16:18:16.508727 4716 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/61f22dc6-45b1-4307-8eee-fd493fc96408-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:16 crc kubenswrapper[4716]: I1207 16:18:16.508737 4716 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/61f22dc6-45b1-4307-8eee-fd493fc96408-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:16 crc kubenswrapper[4716]: I1207 16:18:16.508745 4716 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/61f22dc6-45b1-4307-8eee-fd493fc96408-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:16 crc kubenswrapper[4716]: I1207 16:18:16.508755 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c9wjv\" (UniqueName: \"kubernetes.io/projected/61f22dc6-45b1-4307-8eee-fd493fc96408-kube-api-access-c9wjv\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:16 crc kubenswrapper[4716]: I1207 16:18:16.618533 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-tsr96"] Dec 07 16:18:16 crc kubenswrapper[4716]: I1207 16:18:16.626309 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-tsr96"] Dec 07 16:18:17 crc kubenswrapper[4716]: I1207 16:18:17.369769 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-nj89r" event={"ID":"666f9fee-767e-40d6-ac44-5abe506e3266","Type":"ContainerStarted","Data":"ebe51569d6f1b0c4d1019e4f87d89008f07452249943e222bf04cf1af5375d29"} Dec 07 16:18:17 crc kubenswrapper[4716]: I1207 16:18:17.370439 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-nj89r" Dec 07 16:18:17 crc kubenswrapper[4716]: I1207 16:18:17.386397 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-698758b865-nj89r" podStartSLOduration=5.386377176 podStartE2EDuration="5.386377176s" podCreationTimestamp="2025-12-07 16:18:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:18:17.386359966 +0000 UTC m=+960.076644878" watchObservedRunningTime="2025-12-07 16:18:17.386377176 +0000 UTC m=+960.076662098" Dec 07 16:18:17 crc kubenswrapper[4716]: I1207 16:18:17.676057 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="61f22dc6-45b1-4307-8eee-fd493fc96408" path="/var/lib/kubelet/pods/61f22dc6-45b1-4307-8eee-fd493fc96408/volumes" Dec 07 16:18:17 crc kubenswrapper[4716]: I1207 16:18:17.734634 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a714bbaa-9626-4b49-92ae-2abb029408d3-etc-swift\") pod \"swift-storage-0\" (UID: \"a714bbaa-9626-4b49-92ae-2abb029408d3\") " pod="openstack/swift-storage-0" Dec 07 16:18:17 crc kubenswrapper[4716]: E1207 16:18:17.734882 4716 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 07 16:18:17 crc kubenswrapper[4716]: E1207 16:18:17.734907 4716 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 07 16:18:17 crc kubenswrapper[4716]: E1207 16:18:17.734954 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a714bbaa-9626-4b49-92ae-2abb029408d3-etc-swift podName:a714bbaa-9626-4b49-92ae-2abb029408d3 nodeName:}" failed. No retries permitted until 2025-12-07 16:18:21.734937905 +0000 UTC m=+964.425222817 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/a714bbaa-9626-4b49-92ae-2abb029408d3-etc-swift") pod "swift-storage-0" (UID: "a714bbaa-9626-4b49-92ae-2abb029408d3") : configmap "swift-ring-files" not found Dec 07 16:18:19 crc kubenswrapper[4716]: I1207 16:18:19.229406 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-f5b6-account-create-update-b7hp7" Dec 07 16:18:19 crc kubenswrapper[4716]: I1207 16:18:19.235185 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-7w7t8" Dec 07 16:18:19 crc kubenswrapper[4716]: I1207 16:18:19.244668 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-ce39-account-create-update-5f2ml" Dec 07 16:18:19 crc kubenswrapper[4716]: I1207 16:18:19.254779 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-c4d5d" Dec 07 16:18:19 crc kubenswrapper[4716]: I1207 16:18:19.361869 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1c2f879d-d624-4537-9ab4-f4076614d82e-operator-scripts\") pod \"1c2f879d-d624-4537-9ab4-f4076614d82e\" (UID: \"1c2f879d-d624-4537-9ab4-f4076614d82e\") " Dec 07 16:18:19 crc kubenswrapper[4716]: I1207 16:18:19.362206 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4dxzz\" (UniqueName: \"kubernetes.io/projected/9f2a2677-eb1a-40c1-a947-c3347f6c8c5a-kube-api-access-4dxzz\") pod \"9f2a2677-eb1a-40c1-a947-c3347f6c8c5a\" (UID: \"9f2a2677-eb1a-40c1-a947-c3347f6c8c5a\") " Dec 07 16:18:19 crc kubenswrapper[4716]: I1207 16:18:19.362354 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1ce6b31a-2beb-4fc8-ba79-07cb3f4a7ec8-operator-scripts\") pod \"1ce6b31a-2beb-4fc8-ba79-07cb3f4a7ec8\" (UID: \"1ce6b31a-2beb-4fc8-ba79-07cb3f4a7ec8\") " Dec 07 16:18:19 crc kubenswrapper[4716]: I1207 16:18:19.362414 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/db19870c-0848-4366-8e07-f6028a151c86-operator-scripts\") pod \"db19870c-0848-4366-8e07-f6028a151c86\" (UID: \"db19870c-0848-4366-8e07-f6028a151c86\") " Dec 07 16:18:19 crc kubenswrapper[4716]: I1207 16:18:19.362430 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mc9m8\" (UniqueName: \"kubernetes.io/projected/db19870c-0848-4366-8e07-f6028a151c86-kube-api-access-mc9m8\") pod \"db19870c-0848-4366-8e07-f6028a151c86\" (UID: \"db19870c-0848-4366-8e07-f6028a151c86\") " Dec 07 16:18:19 crc kubenswrapper[4716]: I1207 16:18:19.362525 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-swsbh\" (UniqueName: \"kubernetes.io/projected/1c2f879d-d624-4537-9ab4-f4076614d82e-kube-api-access-swsbh\") pod \"1c2f879d-d624-4537-9ab4-f4076614d82e\" (UID: \"1c2f879d-d624-4537-9ab4-f4076614d82e\") " Dec 07 16:18:19 crc kubenswrapper[4716]: I1207 16:18:19.362548 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zbj\" (UniqueName: \"kubernetes.io/projected/1ce6b31a-2beb-4fc8-ba79-07cb3f4a7ec8-kube-api-access-x7zbj\") pod \"1ce6b31a-2beb-4fc8-ba79-07cb3f4a7ec8\" (UID: \"1ce6b31a-2beb-4fc8-ba79-07cb3f4a7ec8\") " Dec 07 16:18:19 crc kubenswrapper[4716]: I1207 16:18:19.362582 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9f2a2677-eb1a-40c1-a947-c3347f6c8c5a-operator-scripts\") pod \"9f2a2677-eb1a-40c1-a947-c3347f6c8c5a\" (UID: \"9f2a2677-eb1a-40c1-a947-c3347f6c8c5a\") " Dec 07 16:18:19 crc kubenswrapper[4716]: I1207 16:18:19.363213 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/db19870c-0848-4366-8e07-f6028a151c86-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "db19870c-0848-4366-8e07-f6028a151c86" (UID: "db19870c-0848-4366-8e07-f6028a151c86"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:18:19 crc kubenswrapper[4716]: I1207 16:18:19.363461 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1c2f879d-d624-4537-9ab4-f4076614d82e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1c2f879d-d624-4537-9ab4-f4076614d82e" (UID: "1c2f879d-d624-4537-9ab4-f4076614d82e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:18:19 crc kubenswrapper[4716]: I1207 16:18:19.364221 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ce6b31a-2beb-4fc8-ba79-07cb3f4a7ec8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1ce6b31a-2beb-4fc8-ba79-07cb3f4a7ec8" (UID: "1ce6b31a-2beb-4fc8-ba79-07cb3f4a7ec8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:18:19 crc kubenswrapper[4716]: I1207 16:18:19.367534 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9f2a2677-eb1a-40c1-a947-c3347f6c8c5a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9f2a2677-eb1a-40c1-a947-c3347f6c8c5a" (UID: "9f2a2677-eb1a-40c1-a947-c3347f6c8c5a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:18:19 crc kubenswrapper[4716]: I1207 16:18:19.375453 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f2a2677-eb1a-40c1-a947-c3347f6c8c5a-kube-api-access-4dxzz" (OuterVolumeSpecName: "kube-api-access-4dxzz") pod "9f2a2677-eb1a-40c1-a947-c3347f6c8c5a" (UID: "9f2a2677-eb1a-40c1-a947-c3347f6c8c5a"). InnerVolumeSpecName "kube-api-access-4dxzz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:18:19 crc kubenswrapper[4716]: I1207 16:18:19.377291 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db19870c-0848-4366-8e07-f6028a151c86-kube-api-access-mc9m8" (OuterVolumeSpecName: "kube-api-access-mc9m8") pod "db19870c-0848-4366-8e07-f6028a151c86" (UID: "db19870c-0848-4366-8e07-f6028a151c86"). InnerVolumeSpecName "kube-api-access-mc9m8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:18:19 crc kubenswrapper[4716]: I1207 16:18:19.380262 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c2f879d-d624-4537-9ab4-f4076614d82e-kube-api-access-swsbh" (OuterVolumeSpecName: "kube-api-access-swsbh") pod "1c2f879d-d624-4537-9ab4-f4076614d82e" (UID: "1c2f879d-d624-4537-9ab4-f4076614d82e"). InnerVolumeSpecName "kube-api-access-swsbh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:18:19 crc kubenswrapper[4716]: I1207 16:18:19.380326 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ce6b31a-2beb-4fc8-ba79-07cb3f4a7ec8-kube-api-access-x7zbj" (OuterVolumeSpecName: "kube-api-access-x7zbj") pod "1ce6b31a-2beb-4fc8-ba79-07cb3f4a7ec8" (UID: "1ce6b31a-2beb-4fc8-ba79-07cb3f4a7ec8"). InnerVolumeSpecName "kube-api-access-x7zbj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:18:19 crc kubenswrapper[4716]: I1207 16:18:19.432257 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-7w7t8" Dec 07 16:18:19 crc kubenswrapper[4716]: I1207 16:18:19.432993 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-7w7t8" event={"ID":"9f2a2677-eb1a-40c1-a947-c3347f6c8c5a","Type":"ContainerDied","Data":"616cb9cb90f8b90ac477f676661e63812b79055fa3dda227c7a9f37a024a8465"} Dec 07 16:18:19 crc kubenswrapper[4716]: I1207 16:18:19.447771 4716 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="616cb9cb90f8b90ac477f676661e63812b79055fa3dda227c7a9f37a024a8465" Dec 07 16:18:19 crc kubenswrapper[4716]: I1207 16:18:19.449375 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-ce39-account-create-update-5f2ml" Dec 07 16:18:19 crc kubenswrapper[4716]: I1207 16:18:19.449495 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-ce39-account-create-update-5f2ml" event={"ID":"1c2f879d-d624-4537-9ab4-f4076614d82e","Type":"ContainerDied","Data":"3c8b3281a14fc08f10c40ea9ea4a1f68669481fb09f47334b9c08045f8422fe4"} Dec 07 16:18:19 crc kubenswrapper[4716]: I1207 16:18:19.449535 4716 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3c8b3281a14fc08f10c40ea9ea4a1f68669481fb09f47334b9c08045f8422fe4" Dec 07 16:18:19 crc kubenswrapper[4716]: I1207 16:18:19.454968 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-c4d5d" event={"ID":"1ce6b31a-2beb-4fc8-ba79-07cb3f4a7ec8","Type":"ContainerDied","Data":"2a40ba0d85c8b58afd371de1ed1a9e6b93070ada34643f3e612ddae1a2245f91"} Dec 07 16:18:19 crc kubenswrapper[4716]: I1207 16:18:19.455008 4716 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2a40ba0d85c8b58afd371de1ed1a9e6b93070ada34643f3e612ddae1a2245f91" Dec 07 16:18:19 crc kubenswrapper[4716]: I1207 16:18:19.456613 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-c4d5d" Dec 07 16:18:19 crc kubenswrapper[4716]: I1207 16:18:19.463802 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-f5b6-account-create-update-b7hp7" event={"ID":"db19870c-0848-4366-8e07-f6028a151c86","Type":"ContainerDied","Data":"e1c5b50c560d2a9661cfe8b19ac4030ef8ea65998a8d6695aa5335ae7fed5d38"} Dec 07 16:18:19 crc kubenswrapper[4716]: I1207 16:18:19.463847 4716 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e1c5b50c560d2a9661cfe8b19ac4030ef8ea65998a8d6695aa5335ae7fed5d38" Dec 07 16:18:19 crc kubenswrapper[4716]: I1207 16:18:19.465900 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-f5b6-account-create-update-b7hp7" Dec 07 16:18:19 crc kubenswrapper[4716]: I1207 16:18:19.470092 4716 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1ce6b31a-2beb-4fc8-ba79-07cb3f4a7ec8-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:19 crc kubenswrapper[4716]: I1207 16:18:19.470118 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mc9m8\" (UniqueName: \"kubernetes.io/projected/db19870c-0848-4366-8e07-f6028a151c86-kube-api-access-mc9m8\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:19 crc kubenswrapper[4716]: I1207 16:18:19.470130 4716 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/db19870c-0848-4366-8e07-f6028a151c86-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:19 crc kubenswrapper[4716]: I1207 16:18:19.470141 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-swsbh\" (UniqueName: \"kubernetes.io/projected/1c2f879d-d624-4537-9ab4-f4076614d82e-kube-api-access-swsbh\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:19 crc kubenswrapper[4716]: I1207 16:18:19.470156 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zbj\" (UniqueName: \"kubernetes.io/projected/1ce6b31a-2beb-4fc8-ba79-07cb3f4a7ec8-kube-api-access-x7zbj\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:19 crc kubenswrapper[4716]: I1207 16:18:19.470166 4716 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9f2a2677-eb1a-40c1-a947-c3347f6c8c5a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:19 crc kubenswrapper[4716]: I1207 16:18:19.470174 4716 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1c2f879d-d624-4537-9ab4-f4076614d82e-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:19 crc kubenswrapper[4716]: I1207 16:18:19.470184 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4dxzz\" (UniqueName: \"kubernetes.io/projected/9f2a2677-eb1a-40c1-a947-c3347f6c8c5a-kube-api-access-4dxzz\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:20 crc kubenswrapper[4716]: I1207 16:18:20.471565 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-jgrbf" event={"ID":"2c371669-443a-492e-ad04-ab79ae978e5f","Type":"ContainerStarted","Data":"cebfc09ad50e86c5b45ce0d1d9b6d4d9bca9d6bdb89d414e96b79dbb5034e9b2"} Dec 07 16:18:20 crc kubenswrapper[4716]: I1207 16:18:20.493502 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-jgrbf" podStartSLOduration=2.746031702 podStartE2EDuration="6.493487269s" podCreationTimestamp="2025-12-07 16:18:14 +0000 UTC" firstStartedPulling="2025-12-07 16:18:15.405547126 +0000 UTC m=+958.095832038" lastFinishedPulling="2025-12-07 16:18:19.153002673 +0000 UTC m=+961.843287605" observedRunningTime="2025-12-07 16:18:20.486393281 +0000 UTC m=+963.176678213" watchObservedRunningTime="2025-12-07 16:18:20.493487269 +0000 UTC m=+963.183772171" Dec 07 16:18:20 crc kubenswrapper[4716]: I1207 16:18:20.690119 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Dec 07 16:18:21 crc kubenswrapper[4716]: I1207 16:18:21.807283 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a714bbaa-9626-4b49-92ae-2abb029408d3-etc-swift\") pod \"swift-storage-0\" (UID: \"a714bbaa-9626-4b49-92ae-2abb029408d3\") " pod="openstack/swift-storage-0" Dec 07 16:18:21 crc kubenswrapper[4716]: E1207 16:18:21.807467 4716 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 07 16:18:21 crc kubenswrapper[4716]: E1207 16:18:21.807671 4716 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 07 16:18:21 crc kubenswrapper[4716]: E1207 16:18:21.807719 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a714bbaa-9626-4b49-92ae-2abb029408d3-etc-swift podName:a714bbaa-9626-4b49-92ae-2abb029408d3 nodeName:}" failed. No retries permitted until 2025-12-07 16:18:29.807704163 +0000 UTC m=+972.497989075 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/a714bbaa-9626-4b49-92ae-2abb029408d3-etc-swift") pod "swift-storage-0" (UID: "a714bbaa-9626-4b49-92ae-2abb029408d3") : configmap "swift-ring-files" not found Dec 07 16:18:22 crc kubenswrapper[4716]: I1207 16:18:22.488935 4716 generic.go:334] "Generic (PLEG): container finished" podID="9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe" containerID="dec9f36ef119ff579ca865b84eed4ecbc4e8e9b8f0a6ad017672916770251362" exitCode=0 Dec 07 16:18:22 crc kubenswrapper[4716]: I1207 16:18:22.488981 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe","Type":"ContainerDied","Data":"dec9f36ef119ff579ca865b84eed4ecbc4e8e9b8f0a6ad017672916770251362"} Dec 07 16:18:23 crc kubenswrapper[4716]: I1207 16:18:23.145949 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-698758b865-nj89r" Dec 07 16:18:23 crc kubenswrapper[4716]: I1207 16:18:23.197832 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-ktd8n"] Dec 07 16:18:23 crc kubenswrapper[4716]: I1207 16:18:23.198316 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-ktd8n" podUID="240ebe8a-eb59-4726-a66f-4ce1e2900e4c" containerName="dnsmasq-dns" containerID="cri-o://7bd18f80e3166b1c08642a65ba11421da6ac8e3311a8c296181cfabdf597a7cd" gracePeriod=10 Dec 07 16:18:23 crc kubenswrapper[4716]: I1207 16:18:23.501572 4716 generic.go:334] "Generic (PLEG): container finished" podID="240ebe8a-eb59-4726-a66f-4ce1e2900e4c" containerID="7bd18f80e3166b1c08642a65ba11421da6ac8e3311a8c296181cfabdf597a7cd" exitCode=0 Dec 07 16:18:23 crc kubenswrapper[4716]: I1207 16:18:23.501612 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-ktd8n" event={"ID":"240ebe8a-eb59-4726-a66f-4ce1e2900e4c","Type":"ContainerDied","Data":"7bd18f80e3166b1c08642a65ba11421da6ac8e3311a8c296181cfabdf597a7cd"} Dec 07 16:18:25 crc kubenswrapper[4716]: I1207 16:18:25.545392 4716 generic.go:334] "Generic (PLEG): container finished" podID="f653df87-c234-4cb6-8ee8-6b16b4fe0044" containerID="c4250e9cb4d0c6bd1c2d3e3bc3ca8c43b9a53e5cd186b5bd154837621d4a55f1" exitCode=0 Dec 07 16:18:25 crc kubenswrapper[4716]: I1207 16:18:25.545725 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f653df87-c234-4cb6-8ee8-6b16b4fe0044","Type":"ContainerDied","Data":"c4250e9cb4d0c6bd1c2d3e3bc3ca8c43b9a53e5cd186b5bd154837621d4a55f1"} Dec 07 16:18:26 crc kubenswrapper[4716]: I1207 16:18:26.563287 4716 generic.go:334] "Generic (PLEG): container finished" podID="2c371669-443a-492e-ad04-ab79ae978e5f" containerID="cebfc09ad50e86c5b45ce0d1d9b6d4d9bca9d6bdb89d414e96b79dbb5034e9b2" exitCode=0 Dec 07 16:18:26 crc kubenswrapper[4716]: I1207 16:18:26.563365 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-jgrbf" event={"ID":"2c371669-443a-492e-ad04-ab79ae978e5f","Type":"ContainerDied","Data":"cebfc09ad50e86c5b45ce0d1d9b6d4d9bca9d6bdb89d414e96b79dbb5034e9b2"} Dec 07 16:18:26 crc kubenswrapper[4716]: I1207 16:18:26.834381 4716 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-57d769cc4f-ktd8n" podUID="240ebe8a-eb59-4726-a66f-4ce1e2900e4c" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.96:5353: connect: connection refused" Dec 07 16:18:28 crc kubenswrapper[4716]: I1207 16:18:28.435008 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-jgrbf" Dec 07 16:18:28 crc kubenswrapper[4716]: I1207 16:18:28.580838 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-jgrbf" event={"ID":"2c371669-443a-492e-ad04-ab79ae978e5f","Type":"ContainerDied","Data":"07d85826576be1e43cfa0fb3bbaedbd81b85e1e60da9e98f7702be82cf998308"} Dec 07 16:18:28 crc kubenswrapper[4716]: I1207 16:18:28.581184 4716 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="07d85826576be1e43cfa0fb3bbaedbd81b85e1e60da9e98f7702be82cf998308" Dec 07 16:18:28 crc kubenswrapper[4716]: I1207 16:18:28.580912 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-jgrbf" Dec 07 16:18:28 crc kubenswrapper[4716]: I1207 16:18:28.599200 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c371669-443a-492e-ad04-ab79ae978e5f-combined-ca-bundle\") pod \"2c371669-443a-492e-ad04-ab79ae978e5f\" (UID: \"2c371669-443a-492e-ad04-ab79ae978e5f\") " Dec 07 16:18:28 crc kubenswrapper[4716]: I1207 16:18:28.599287 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2c371669-443a-492e-ad04-ab79ae978e5f-ring-data-devices\") pod \"2c371669-443a-492e-ad04-ab79ae978e5f\" (UID: \"2c371669-443a-492e-ad04-ab79ae978e5f\") " Dec 07 16:18:28 crc kubenswrapper[4716]: I1207 16:18:28.599374 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2c371669-443a-492e-ad04-ab79ae978e5f-dispersionconf\") pod \"2c371669-443a-492e-ad04-ab79ae978e5f\" (UID: \"2c371669-443a-492e-ad04-ab79ae978e5f\") " Dec 07 16:18:28 crc kubenswrapper[4716]: I1207 16:18:28.599427 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t9khz\" (UniqueName: \"kubernetes.io/projected/2c371669-443a-492e-ad04-ab79ae978e5f-kube-api-access-t9khz\") pod \"2c371669-443a-492e-ad04-ab79ae978e5f\" (UID: \"2c371669-443a-492e-ad04-ab79ae978e5f\") " Dec 07 16:18:28 crc kubenswrapper[4716]: I1207 16:18:28.599514 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2c371669-443a-492e-ad04-ab79ae978e5f-swiftconf\") pod \"2c371669-443a-492e-ad04-ab79ae978e5f\" (UID: \"2c371669-443a-492e-ad04-ab79ae978e5f\") " Dec 07 16:18:28 crc kubenswrapper[4716]: I1207 16:18:28.599535 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2c371669-443a-492e-ad04-ab79ae978e5f-etc-swift\") pod \"2c371669-443a-492e-ad04-ab79ae978e5f\" (UID: \"2c371669-443a-492e-ad04-ab79ae978e5f\") " Dec 07 16:18:28 crc kubenswrapper[4716]: I1207 16:18:28.599551 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2c371669-443a-492e-ad04-ab79ae978e5f-scripts\") pod \"2c371669-443a-492e-ad04-ab79ae978e5f\" (UID: \"2c371669-443a-492e-ad04-ab79ae978e5f\") " Dec 07 16:18:28 crc kubenswrapper[4716]: I1207 16:18:28.606442 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2c371669-443a-492e-ad04-ab79ae978e5f-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "2c371669-443a-492e-ad04-ab79ae978e5f" (UID: "2c371669-443a-492e-ad04-ab79ae978e5f"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:18:28 crc kubenswrapper[4716]: I1207 16:18:28.607560 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2c371669-443a-492e-ad04-ab79ae978e5f-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "2c371669-443a-492e-ad04-ab79ae978e5f" (UID: "2c371669-443a-492e-ad04-ab79ae978e5f"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:18:28 crc kubenswrapper[4716]: I1207 16:18:28.607831 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c371669-443a-492e-ad04-ab79ae978e5f-kube-api-access-t9khz" (OuterVolumeSpecName: "kube-api-access-t9khz") pod "2c371669-443a-492e-ad04-ab79ae978e5f" (UID: "2c371669-443a-492e-ad04-ab79ae978e5f"). InnerVolumeSpecName "kube-api-access-t9khz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:18:28 crc kubenswrapper[4716]: I1207 16:18:28.610464 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c371669-443a-492e-ad04-ab79ae978e5f-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "2c371669-443a-492e-ad04-ab79ae978e5f" (UID: "2c371669-443a-492e-ad04-ab79ae978e5f"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:18:28 crc kubenswrapper[4716]: I1207 16:18:28.633254 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-ktd8n" Dec 07 16:18:28 crc kubenswrapper[4716]: I1207 16:18:28.644694 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2c371669-443a-492e-ad04-ab79ae978e5f-scripts" (OuterVolumeSpecName: "scripts") pod "2c371669-443a-492e-ad04-ab79ae978e5f" (UID: "2c371669-443a-492e-ad04-ab79ae978e5f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:18:28 crc kubenswrapper[4716]: I1207 16:18:28.671245 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c371669-443a-492e-ad04-ab79ae978e5f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2c371669-443a-492e-ad04-ab79ae978e5f" (UID: "2c371669-443a-492e-ad04-ab79ae978e5f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:18:28 crc kubenswrapper[4716]: I1207 16:18:28.680568 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c371669-443a-492e-ad04-ab79ae978e5f-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "2c371669-443a-492e-ad04-ab79ae978e5f" (UID: "2c371669-443a-492e-ad04-ab79ae978e5f"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:18:28 crc kubenswrapper[4716]: I1207 16:18:28.704170 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t9khz\" (UniqueName: \"kubernetes.io/projected/2c371669-443a-492e-ad04-ab79ae978e5f-kube-api-access-t9khz\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:28 crc kubenswrapper[4716]: I1207 16:18:28.704468 4716 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2c371669-443a-492e-ad04-ab79ae978e5f-swiftconf\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:28 crc kubenswrapper[4716]: I1207 16:18:28.704619 4716 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2c371669-443a-492e-ad04-ab79ae978e5f-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:28 crc kubenswrapper[4716]: I1207 16:18:28.704692 4716 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2c371669-443a-492e-ad04-ab79ae978e5f-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:28 crc kubenswrapper[4716]: I1207 16:18:28.704759 4716 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c371669-443a-492e-ad04-ab79ae978e5f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:28 crc kubenswrapper[4716]: I1207 16:18:28.704834 4716 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2c371669-443a-492e-ad04-ab79ae978e5f-ring-data-devices\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:28 crc kubenswrapper[4716]: I1207 16:18:28.704899 4716 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2c371669-443a-492e-ad04-ab79ae978e5f-dispersionconf\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:28 crc kubenswrapper[4716]: I1207 16:18:28.806542 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/240ebe8a-eb59-4726-a66f-4ce1e2900e4c-config\") pod \"240ebe8a-eb59-4726-a66f-4ce1e2900e4c\" (UID: \"240ebe8a-eb59-4726-a66f-4ce1e2900e4c\") " Dec 07 16:18:28 crc kubenswrapper[4716]: I1207 16:18:28.806595 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/240ebe8a-eb59-4726-a66f-4ce1e2900e4c-dns-svc\") pod \"240ebe8a-eb59-4726-a66f-4ce1e2900e4c\" (UID: \"240ebe8a-eb59-4726-a66f-4ce1e2900e4c\") " Dec 07 16:18:28 crc kubenswrapper[4716]: I1207 16:18:28.806640 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cx6ns\" (UniqueName: \"kubernetes.io/projected/240ebe8a-eb59-4726-a66f-4ce1e2900e4c-kube-api-access-cx6ns\") pod \"240ebe8a-eb59-4726-a66f-4ce1e2900e4c\" (UID: \"240ebe8a-eb59-4726-a66f-4ce1e2900e4c\") " Dec 07 16:18:28 crc kubenswrapper[4716]: I1207 16:18:28.810022 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/240ebe8a-eb59-4726-a66f-4ce1e2900e4c-kube-api-access-cx6ns" (OuterVolumeSpecName: "kube-api-access-cx6ns") pod "240ebe8a-eb59-4726-a66f-4ce1e2900e4c" (UID: "240ebe8a-eb59-4726-a66f-4ce1e2900e4c"). InnerVolumeSpecName "kube-api-access-cx6ns". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:18:28 crc kubenswrapper[4716]: E1207 16:18:28.862254 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/240ebe8a-eb59-4726-a66f-4ce1e2900e4c-dns-svc podName:240ebe8a-eb59-4726-a66f-4ce1e2900e4c nodeName:}" failed. No retries permitted until 2025-12-07 16:18:29.362220599 +0000 UTC m=+972.052505521 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "dns-svc" (UniqueName: "kubernetes.io/configmap/240ebe8a-eb59-4726-a66f-4ce1e2900e4c-dns-svc") pod "240ebe8a-eb59-4726-a66f-4ce1e2900e4c" (UID: "240ebe8a-eb59-4726-a66f-4ce1e2900e4c") : error deleting /var/lib/kubelet/pods/240ebe8a-eb59-4726-a66f-4ce1e2900e4c/volume-subpaths: remove /var/lib/kubelet/pods/240ebe8a-eb59-4726-a66f-4ce1e2900e4c/volume-subpaths: no such file or directory Dec 07 16:18:28 crc kubenswrapper[4716]: I1207 16:18:28.862445 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/240ebe8a-eb59-4726-a66f-4ce1e2900e4c-config" (OuterVolumeSpecName: "config") pod "240ebe8a-eb59-4726-a66f-4ce1e2900e4c" (UID: "240ebe8a-eb59-4726-a66f-4ce1e2900e4c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:18:28 crc kubenswrapper[4716]: I1207 16:18:28.908580 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cx6ns\" (UniqueName: \"kubernetes.io/projected/240ebe8a-eb59-4726-a66f-4ce1e2900e4c-kube-api-access-cx6ns\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:28 crc kubenswrapper[4716]: I1207 16:18:28.908613 4716 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/240ebe8a-eb59-4726-a66f-4ce1e2900e4c-config\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:29 crc kubenswrapper[4716]: I1207 16:18:29.417035 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/240ebe8a-eb59-4726-a66f-4ce1e2900e4c-dns-svc\") pod \"240ebe8a-eb59-4726-a66f-4ce1e2900e4c\" (UID: \"240ebe8a-eb59-4726-a66f-4ce1e2900e4c\") " Dec 07 16:18:29 crc kubenswrapper[4716]: I1207 16:18:29.417495 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/240ebe8a-eb59-4726-a66f-4ce1e2900e4c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "240ebe8a-eb59-4726-a66f-4ce1e2900e4c" (UID: "240ebe8a-eb59-4726-a66f-4ce1e2900e4c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:18:29 crc kubenswrapper[4716]: I1207 16:18:29.521634 4716 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/240ebe8a-eb59-4726-a66f-4ce1e2900e4c-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:29 crc kubenswrapper[4716]: I1207 16:18:29.588495 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f653df87-c234-4cb6-8ee8-6b16b4fe0044","Type":"ContainerStarted","Data":"e24ca70ffe380dd6ae0b3ce56869f1dfe6b1c7a31169cf168b558517d0a1aa65"} Dec 07 16:18:29 crc kubenswrapper[4716]: I1207 16:18:29.589570 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 07 16:18:29 crc kubenswrapper[4716]: I1207 16:18:29.591267 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-ktd8n" event={"ID":"240ebe8a-eb59-4726-a66f-4ce1e2900e4c","Type":"ContainerDied","Data":"e9b4fd94e94a98f504eeef3860b2d6161de289fdcc23e3137c8f71f653b02250"} Dec 07 16:18:29 crc kubenswrapper[4716]: I1207 16:18:29.591334 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-ktd8n" Dec 07 16:18:29 crc kubenswrapper[4716]: I1207 16:18:29.591357 4716 scope.go:117] "RemoveContainer" containerID="7bd18f80e3166b1c08642a65ba11421da6ac8e3311a8c296181cfabdf597a7cd" Dec 07 16:18:29 crc kubenswrapper[4716]: I1207 16:18:29.592527 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-gq62d" event={"ID":"a17893a0-50ef-4d05-8aaf-c6a9e3f8d20d","Type":"ContainerStarted","Data":"ffa4933c139800dce7b364c70ad2bd22dc6cb74649ec6f2f4c0be145c6b55016"} Dec 07 16:18:29 crc kubenswrapper[4716]: I1207 16:18:29.594364 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe","Type":"ContainerStarted","Data":"9c52f289d4d35a7ab538719f9c3c5eb4d453bb8387f9e697d5cf26704c98cd04"} Dec 07 16:18:29 crc kubenswrapper[4716]: I1207 16:18:29.595154 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:18:29 crc kubenswrapper[4716]: I1207 16:18:29.612603 4716 scope.go:117] "RemoveContainer" containerID="ed81504e8f9d8db9e2c7703a8dc46d342a761904bb7ad96a3c70efde8df007f8" Dec 07 16:18:29 crc kubenswrapper[4716]: I1207 16:18:29.620122 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=42.811318268 podStartE2EDuration="1m3.620105211s" podCreationTimestamp="2025-12-07 16:17:26 +0000 UTC" firstStartedPulling="2025-12-07 16:17:27.893444443 +0000 UTC m=+910.583729355" lastFinishedPulling="2025-12-07 16:17:48.702231376 +0000 UTC m=+931.392516298" observedRunningTime="2025-12-07 16:18:29.616006666 +0000 UTC m=+972.306291578" watchObservedRunningTime="2025-12-07 16:18:29.620105211 +0000 UTC m=+972.310390123" Dec 07 16:18:29 crc kubenswrapper[4716]: I1207 16:18:29.674049 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=43.15487945 podStartE2EDuration="1m3.674022944s" podCreationTimestamp="2025-12-07 16:17:26 +0000 UTC" firstStartedPulling="2025-12-07 16:17:28.217071809 +0000 UTC m=+910.907356711" lastFinishedPulling="2025-12-07 16:17:48.736215293 +0000 UTC m=+931.426500205" observedRunningTime="2025-12-07 16:18:29.665047734 +0000 UTC m=+972.355332646" watchObservedRunningTime="2025-12-07 16:18:29.674022944 +0000 UTC m=+972.364307856" Dec 07 16:18:29 crc kubenswrapper[4716]: I1207 16:18:29.690823 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-gq62d" podStartSLOduration=2.484621007 podStartE2EDuration="18.690792842s" podCreationTimestamp="2025-12-07 16:18:11 +0000 UTC" firstStartedPulling="2025-12-07 16:18:12.24705397 +0000 UTC m=+954.937338882" lastFinishedPulling="2025-12-07 16:18:28.453225805 +0000 UTC m=+971.143510717" observedRunningTime="2025-12-07 16:18:29.684248629 +0000 UTC m=+972.374533541" watchObservedRunningTime="2025-12-07 16:18:29.690792842 +0000 UTC m=+972.381077754" Dec 07 16:18:29 crc kubenswrapper[4716]: I1207 16:18:29.708483 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-ktd8n"] Dec 07 16:18:29 crc kubenswrapper[4716]: I1207 16:18:29.711638 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-ktd8n"] Dec 07 16:18:29 crc kubenswrapper[4716]: I1207 16:18:29.829267 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a714bbaa-9626-4b49-92ae-2abb029408d3-etc-swift\") pod \"swift-storage-0\" (UID: \"a714bbaa-9626-4b49-92ae-2abb029408d3\") " pod="openstack/swift-storage-0" Dec 07 16:18:29 crc kubenswrapper[4716]: I1207 16:18:29.837450 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a714bbaa-9626-4b49-92ae-2abb029408d3-etc-swift\") pod \"swift-storage-0\" (UID: \"a714bbaa-9626-4b49-92ae-2abb029408d3\") " pod="openstack/swift-storage-0" Dec 07 16:18:29 crc kubenswrapper[4716]: I1207 16:18:29.861743 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 07 16:18:30 crc kubenswrapper[4716]: I1207 16:18:30.468426 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 07 16:18:30 crc kubenswrapper[4716]: I1207 16:18:30.605070 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a714bbaa-9626-4b49-92ae-2abb029408d3","Type":"ContainerStarted","Data":"c76f762331fb2c985b946dd304ad88023901dfbf87715ffd13a750d93d2aa8d9"} Dec 07 16:18:31 crc kubenswrapper[4716]: I1207 16:18:31.280151 4716 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-tgl27" podUID="399984bf-94af-4630-a7c3-9375a388159a" containerName="ovn-controller" probeResult="failure" output=< Dec 07 16:18:31 crc kubenswrapper[4716]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 07 16:18:31 crc kubenswrapper[4716]: > Dec 07 16:18:31 crc kubenswrapper[4716]: I1207 16:18:31.291797 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-g6nwl" Dec 07 16:18:31 crc kubenswrapper[4716]: I1207 16:18:31.295641 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-g6nwl" Dec 07 16:18:31 crc kubenswrapper[4716]: I1207 16:18:31.531288 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-tgl27-config-8c6m2"] Dec 07 16:18:31 crc kubenswrapper[4716]: E1207 16:18:31.532001 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="240ebe8a-eb59-4726-a66f-4ce1e2900e4c" containerName="dnsmasq-dns" Dec 07 16:18:31 crc kubenswrapper[4716]: I1207 16:18:31.532023 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="240ebe8a-eb59-4726-a66f-4ce1e2900e4c" containerName="dnsmasq-dns" Dec 07 16:18:31 crc kubenswrapper[4716]: E1207 16:18:31.532036 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db19870c-0848-4366-8e07-f6028a151c86" containerName="mariadb-account-create-update" Dec 07 16:18:31 crc kubenswrapper[4716]: I1207 16:18:31.532044 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="db19870c-0848-4366-8e07-f6028a151c86" containerName="mariadb-account-create-update" Dec 07 16:18:31 crc kubenswrapper[4716]: E1207 16:18:31.532059 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ce6b31a-2beb-4fc8-ba79-07cb3f4a7ec8" containerName="mariadb-database-create" Dec 07 16:18:31 crc kubenswrapper[4716]: I1207 16:18:31.532067 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ce6b31a-2beb-4fc8-ba79-07cb3f4a7ec8" containerName="mariadb-database-create" Dec 07 16:18:31 crc kubenswrapper[4716]: E1207 16:18:31.532102 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61f22dc6-45b1-4307-8eee-fd493fc96408" containerName="dnsmasq-dns" Dec 07 16:18:31 crc kubenswrapper[4716]: I1207 16:18:31.532111 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="61f22dc6-45b1-4307-8eee-fd493fc96408" containerName="dnsmasq-dns" Dec 07 16:18:31 crc kubenswrapper[4716]: E1207 16:18:31.532122 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f2a2677-eb1a-40c1-a947-c3347f6c8c5a" containerName="mariadb-database-create" Dec 07 16:18:31 crc kubenswrapper[4716]: I1207 16:18:31.532130 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f2a2677-eb1a-40c1-a947-c3347f6c8c5a" containerName="mariadb-database-create" Dec 07 16:18:31 crc kubenswrapper[4716]: E1207 16:18:31.532139 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c371669-443a-492e-ad04-ab79ae978e5f" containerName="swift-ring-rebalance" Dec 07 16:18:31 crc kubenswrapper[4716]: I1207 16:18:31.532146 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c371669-443a-492e-ad04-ab79ae978e5f" containerName="swift-ring-rebalance" Dec 07 16:18:31 crc kubenswrapper[4716]: E1207 16:18:31.532172 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="240ebe8a-eb59-4726-a66f-4ce1e2900e4c" containerName="init" Dec 07 16:18:31 crc kubenswrapper[4716]: I1207 16:18:31.532181 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="240ebe8a-eb59-4726-a66f-4ce1e2900e4c" containerName="init" Dec 07 16:18:31 crc kubenswrapper[4716]: E1207 16:18:31.532196 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c2f879d-d624-4537-9ab4-f4076614d82e" containerName="mariadb-account-create-update" Dec 07 16:18:31 crc kubenswrapper[4716]: I1207 16:18:31.532205 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c2f879d-d624-4537-9ab4-f4076614d82e" containerName="mariadb-account-create-update" Dec 07 16:18:31 crc kubenswrapper[4716]: E1207 16:18:31.532244 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61f22dc6-45b1-4307-8eee-fd493fc96408" containerName="init" Dec 07 16:18:31 crc kubenswrapper[4716]: I1207 16:18:31.532254 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="61f22dc6-45b1-4307-8eee-fd493fc96408" containerName="init" Dec 07 16:18:31 crc kubenswrapper[4716]: I1207 16:18:31.532433 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c2f879d-d624-4537-9ab4-f4076614d82e" containerName="mariadb-account-create-update" Dec 07 16:18:31 crc kubenswrapper[4716]: I1207 16:18:31.532452 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="61f22dc6-45b1-4307-8eee-fd493fc96408" containerName="dnsmasq-dns" Dec 07 16:18:31 crc kubenswrapper[4716]: I1207 16:18:31.532469 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ce6b31a-2beb-4fc8-ba79-07cb3f4a7ec8" containerName="mariadb-database-create" Dec 07 16:18:31 crc kubenswrapper[4716]: I1207 16:18:31.532480 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="db19870c-0848-4366-8e07-f6028a151c86" containerName="mariadb-account-create-update" Dec 07 16:18:31 crc kubenswrapper[4716]: I1207 16:18:31.532495 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f2a2677-eb1a-40c1-a947-c3347f6c8c5a" containerName="mariadb-database-create" Dec 07 16:18:31 crc kubenswrapper[4716]: I1207 16:18:31.532507 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="240ebe8a-eb59-4726-a66f-4ce1e2900e4c" containerName="dnsmasq-dns" Dec 07 16:18:31 crc kubenswrapper[4716]: I1207 16:18:31.532519 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c371669-443a-492e-ad04-ab79ae978e5f" containerName="swift-ring-rebalance" Dec 07 16:18:31 crc kubenswrapper[4716]: I1207 16:18:31.533023 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-tgl27-config-8c6m2" Dec 07 16:18:31 crc kubenswrapper[4716]: I1207 16:18:31.542354 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 07 16:18:31 crc kubenswrapper[4716]: I1207 16:18:31.563811 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-tgl27-config-8c6m2"] Dec 07 16:18:31 crc kubenswrapper[4716]: I1207 16:18:31.659451 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a45ca231-bf99-455d-99de-df3459f6e8c1-var-run\") pod \"ovn-controller-tgl27-config-8c6m2\" (UID: \"a45ca231-bf99-455d-99de-df3459f6e8c1\") " pod="openstack/ovn-controller-tgl27-config-8c6m2" Dec 07 16:18:31 crc kubenswrapper[4716]: I1207 16:18:31.659510 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a45ca231-bf99-455d-99de-df3459f6e8c1-scripts\") pod \"ovn-controller-tgl27-config-8c6m2\" (UID: \"a45ca231-bf99-455d-99de-df3459f6e8c1\") " pod="openstack/ovn-controller-tgl27-config-8c6m2" Dec 07 16:18:31 crc kubenswrapper[4716]: I1207 16:18:31.659577 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/a45ca231-bf99-455d-99de-df3459f6e8c1-var-run-ovn\") pod \"ovn-controller-tgl27-config-8c6m2\" (UID: \"a45ca231-bf99-455d-99de-df3459f6e8c1\") " pod="openstack/ovn-controller-tgl27-config-8c6m2" Dec 07 16:18:31 crc kubenswrapper[4716]: I1207 16:18:31.659633 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/a45ca231-bf99-455d-99de-df3459f6e8c1-additional-scripts\") pod \"ovn-controller-tgl27-config-8c6m2\" (UID: \"a45ca231-bf99-455d-99de-df3459f6e8c1\") " pod="openstack/ovn-controller-tgl27-config-8c6m2" Dec 07 16:18:31 crc kubenswrapper[4716]: I1207 16:18:31.659670 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/a45ca231-bf99-455d-99de-df3459f6e8c1-var-log-ovn\") pod \"ovn-controller-tgl27-config-8c6m2\" (UID: \"a45ca231-bf99-455d-99de-df3459f6e8c1\") " pod="openstack/ovn-controller-tgl27-config-8c6m2" Dec 07 16:18:31 crc kubenswrapper[4716]: I1207 16:18:31.659714 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8c8sw\" (UniqueName: \"kubernetes.io/projected/a45ca231-bf99-455d-99de-df3459f6e8c1-kube-api-access-8c8sw\") pod \"ovn-controller-tgl27-config-8c6m2\" (UID: \"a45ca231-bf99-455d-99de-df3459f6e8c1\") " pod="openstack/ovn-controller-tgl27-config-8c6m2" Dec 07 16:18:31 crc kubenswrapper[4716]: I1207 16:18:31.671484 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="240ebe8a-eb59-4726-a66f-4ce1e2900e4c" path="/var/lib/kubelet/pods/240ebe8a-eb59-4726-a66f-4ce1e2900e4c/volumes" Dec 07 16:18:31 crc kubenswrapper[4716]: I1207 16:18:31.761209 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8c8sw\" (UniqueName: \"kubernetes.io/projected/a45ca231-bf99-455d-99de-df3459f6e8c1-kube-api-access-8c8sw\") pod \"ovn-controller-tgl27-config-8c6m2\" (UID: \"a45ca231-bf99-455d-99de-df3459f6e8c1\") " pod="openstack/ovn-controller-tgl27-config-8c6m2" Dec 07 16:18:31 crc kubenswrapper[4716]: I1207 16:18:31.761302 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a45ca231-bf99-455d-99de-df3459f6e8c1-var-run\") pod \"ovn-controller-tgl27-config-8c6m2\" (UID: \"a45ca231-bf99-455d-99de-df3459f6e8c1\") " pod="openstack/ovn-controller-tgl27-config-8c6m2" Dec 07 16:18:31 crc kubenswrapper[4716]: I1207 16:18:31.761365 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a45ca231-bf99-455d-99de-df3459f6e8c1-scripts\") pod \"ovn-controller-tgl27-config-8c6m2\" (UID: \"a45ca231-bf99-455d-99de-df3459f6e8c1\") " pod="openstack/ovn-controller-tgl27-config-8c6m2" Dec 07 16:18:31 crc kubenswrapper[4716]: I1207 16:18:31.761475 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/a45ca231-bf99-455d-99de-df3459f6e8c1-var-run-ovn\") pod \"ovn-controller-tgl27-config-8c6m2\" (UID: \"a45ca231-bf99-455d-99de-df3459f6e8c1\") " pod="openstack/ovn-controller-tgl27-config-8c6m2" Dec 07 16:18:31 crc kubenswrapper[4716]: I1207 16:18:31.761562 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/a45ca231-bf99-455d-99de-df3459f6e8c1-additional-scripts\") pod \"ovn-controller-tgl27-config-8c6m2\" (UID: \"a45ca231-bf99-455d-99de-df3459f6e8c1\") " pod="openstack/ovn-controller-tgl27-config-8c6m2" Dec 07 16:18:31 crc kubenswrapper[4716]: I1207 16:18:31.761648 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/a45ca231-bf99-455d-99de-df3459f6e8c1-var-log-ovn\") pod \"ovn-controller-tgl27-config-8c6m2\" (UID: \"a45ca231-bf99-455d-99de-df3459f6e8c1\") " pod="openstack/ovn-controller-tgl27-config-8c6m2" Dec 07 16:18:31 crc kubenswrapper[4716]: I1207 16:18:31.763270 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/a45ca231-bf99-455d-99de-df3459f6e8c1-additional-scripts\") pod \"ovn-controller-tgl27-config-8c6m2\" (UID: \"a45ca231-bf99-455d-99de-df3459f6e8c1\") " pod="openstack/ovn-controller-tgl27-config-8c6m2" Dec 07 16:18:31 crc kubenswrapper[4716]: I1207 16:18:31.763458 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/a45ca231-bf99-455d-99de-df3459f6e8c1-var-run-ovn\") pod \"ovn-controller-tgl27-config-8c6m2\" (UID: \"a45ca231-bf99-455d-99de-df3459f6e8c1\") " pod="openstack/ovn-controller-tgl27-config-8c6m2" Dec 07 16:18:31 crc kubenswrapper[4716]: I1207 16:18:31.763802 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/a45ca231-bf99-455d-99de-df3459f6e8c1-var-log-ovn\") pod \"ovn-controller-tgl27-config-8c6m2\" (UID: \"a45ca231-bf99-455d-99de-df3459f6e8c1\") " pod="openstack/ovn-controller-tgl27-config-8c6m2" Dec 07 16:18:31 crc kubenswrapper[4716]: I1207 16:18:31.763959 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a45ca231-bf99-455d-99de-df3459f6e8c1-var-run\") pod \"ovn-controller-tgl27-config-8c6m2\" (UID: \"a45ca231-bf99-455d-99de-df3459f6e8c1\") " pod="openstack/ovn-controller-tgl27-config-8c6m2" Dec 07 16:18:31 crc kubenswrapper[4716]: I1207 16:18:31.765995 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a45ca231-bf99-455d-99de-df3459f6e8c1-scripts\") pod \"ovn-controller-tgl27-config-8c6m2\" (UID: \"a45ca231-bf99-455d-99de-df3459f6e8c1\") " pod="openstack/ovn-controller-tgl27-config-8c6m2" Dec 07 16:18:31 crc kubenswrapper[4716]: I1207 16:18:31.788740 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8c8sw\" (UniqueName: \"kubernetes.io/projected/a45ca231-bf99-455d-99de-df3459f6e8c1-kube-api-access-8c8sw\") pod \"ovn-controller-tgl27-config-8c6m2\" (UID: \"a45ca231-bf99-455d-99de-df3459f6e8c1\") " pod="openstack/ovn-controller-tgl27-config-8c6m2" Dec 07 16:18:31 crc kubenswrapper[4716]: I1207 16:18:31.856821 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-tgl27-config-8c6m2" Dec 07 16:18:32 crc kubenswrapper[4716]: I1207 16:18:32.384898 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-tgl27-config-8c6m2"] Dec 07 16:18:32 crc kubenswrapper[4716]: I1207 16:18:32.667924 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a714bbaa-9626-4b49-92ae-2abb029408d3","Type":"ContainerStarted","Data":"4c09b9776fa5617b941ceb9a114fe6ab7c97ca8bb1b86cff6ffea88a236b004c"} Dec 07 16:18:32 crc kubenswrapper[4716]: I1207 16:18:32.667969 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a714bbaa-9626-4b49-92ae-2abb029408d3","Type":"ContainerStarted","Data":"f29adbd0a6ca14f0d0f76517e1ce93412cc3e0679de6c770fab170e1a11cefce"} Dec 07 16:18:32 crc kubenswrapper[4716]: I1207 16:18:32.667980 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a714bbaa-9626-4b49-92ae-2abb029408d3","Type":"ContainerStarted","Data":"fef24e5b6c1c286f510a84995b3da67f66463b49941b49f65115662d911fb90d"} Dec 07 16:18:32 crc kubenswrapper[4716]: I1207 16:18:32.669480 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-tgl27-config-8c6m2" event={"ID":"a45ca231-bf99-455d-99de-df3459f6e8c1","Type":"ContainerStarted","Data":"5940c7306c7511e0dce7253051e78164edc449a4ce94b67bc2f7dedf299072a9"} Dec 07 16:18:33 crc kubenswrapper[4716]: I1207 16:18:33.679753 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a714bbaa-9626-4b49-92ae-2abb029408d3","Type":"ContainerStarted","Data":"3b539e672a16f3ef680d42f0f4d9ab39a9e964ee12d2defadcf3de2e73724c3d"} Dec 07 16:18:33 crc kubenswrapper[4716]: I1207 16:18:33.682548 4716 generic.go:334] "Generic (PLEG): container finished" podID="a45ca231-bf99-455d-99de-df3459f6e8c1" containerID="a97b565989cc9fd8373a3674aaa3245f870237bc8cf96066c747220dd37285f9" exitCode=0 Dec 07 16:18:33 crc kubenswrapper[4716]: I1207 16:18:33.682618 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-tgl27-config-8c6m2" event={"ID":"a45ca231-bf99-455d-99de-df3459f6e8c1","Type":"ContainerDied","Data":"a97b565989cc9fd8373a3674aaa3245f870237bc8cf96066c747220dd37285f9"} Dec 07 16:18:34 crc kubenswrapper[4716]: I1207 16:18:34.704811 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a714bbaa-9626-4b49-92ae-2abb029408d3","Type":"ContainerStarted","Data":"65be1dce03bda7a3a0ac6bef1fc8740b241d5cbc15f282fbf1061ded4c9ced2f"} Dec 07 16:18:35 crc kubenswrapper[4716]: I1207 16:18:35.044905 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-tgl27-config-8c6m2" Dec 07 16:18:35 crc kubenswrapper[4716]: I1207 16:18:35.107965 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a45ca231-bf99-455d-99de-df3459f6e8c1-var-run\") pod \"a45ca231-bf99-455d-99de-df3459f6e8c1\" (UID: \"a45ca231-bf99-455d-99de-df3459f6e8c1\") " Dec 07 16:18:35 crc kubenswrapper[4716]: I1207 16:18:35.108030 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/a45ca231-bf99-455d-99de-df3459f6e8c1-additional-scripts\") pod \"a45ca231-bf99-455d-99de-df3459f6e8c1\" (UID: \"a45ca231-bf99-455d-99de-df3459f6e8c1\") " Dec 07 16:18:35 crc kubenswrapper[4716]: I1207 16:18:35.108066 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/a45ca231-bf99-455d-99de-df3459f6e8c1-var-log-ovn\") pod \"a45ca231-bf99-455d-99de-df3459f6e8c1\" (UID: \"a45ca231-bf99-455d-99de-df3459f6e8c1\") " Dec 07 16:18:35 crc kubenswrapper[4716]: I1207 16:18:35.108106 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a45ca231-bf99-455d-99de-df3459f6e8c1-var-run" (OuterVolumeSpecName: "var-run") pod "a45ca231-bf99-455d-99de-df3459f6e8c1" (UID: "a45ca231-bf99-455d-99de-df3459f6e8c1"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 16:18:35 crc kubenswrapper[4716]: I1207 16:18:35.108160 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/a45ca231-bf99-455d-99de-df3459f6e8c1-var-run-ovn\") pod \"a45ca231-bf99-455d-99de-df3459f6e8c1\" (UID: \"a45ca231-bf99-455d-99de-df3459f6e8c1\") " Dec 07 16:18:35 crc kubenswrapper[4716]: I1207 16:18:35.108214 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8c8sw\" (UniqueName: \"kubernetes.io/projected/a45ca231-bf99-455d-99de-df3459f6e8c1-kube-api-access-8c8sw\") pod \"a45ca231-bf99-455d-99de-df3459f6e8c1\" (UID: \"a45ca231-bf99-455d-99de-df3459f6e8c1\") " Dec 07 16:18:35 crc kubenswrapper[4716]: I1207 16:18:35.108406 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a45ca231-bf99-455d-99de-df3459f6e8c1-scripts\") pod \"a45ca231-bf99-455d-99de-df3459f6e8c1\" (UID: \"a45ca231-bf99-455d-99de-df3459f6e8c1\") " Dec 07 16:18:35 crc kubenswrapper[4716]: I1207 16:18:35.108730 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a45ca231-bf99-455d-99de-df3459f6e8c1-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "a45ca231-bf99-455d-99de-df3459f6e8c1" (UID: "a45ca231-bf99-455d-99de-df3459f6e8c1"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 16:18:35 crc kubenswrapper[4716]: I1207 16:18:35.109014 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a45ca231-bf99-455d-99de-df3459f6e8c1-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "a45ca231-bf99-455d-99de-df3459f6e8c1" (UID: "a45ca231-bf99-455d-99de-df3459f6e8c1"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:18:35 crc kubenswrapper[4716]: I1207 16:18:35.109444 4716 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a45ca231-bf99-455d-99de-df3459f6e8c1-var-run\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:35 crc kubenswrapper[4716]: I1207 16:18:35.109472 4716 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/a45ca231-bf99-455d-99de-df3459f6e8c1-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:35 crc kubenswrapper[4716]: I1207 16:18:35.109485 4716 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/a45ca231-bf99-455d-99de-df3459f6e8c1-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:35 crc kubenswrapper[4716]: I1207 16:18:35.109775 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a45ca231-bf99-455d-99de-df3459f6e8c1-scripts" (OuterVolumeSpecName: "scripts") pod "a45ca231-bf99-455d-99de-df3459f6e8c1" (UID: "a45ca231-bf99-455d-99de-df3459f6e8c1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:18:35 crc kubenswrapper[4716]: I1207 16:18:35.109828 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a45ca231-bf99-455d-99de-df3459f6e8c1-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "a45ca231-bf99-455d-99de-df3459f6e8c1" (UID: "a45ca231-bf99-455d-99de-df3459f6e8c1"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 16:18:35 crc kubenswrapper[4716]: I1207 16:18:35.114842 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a45ca231-bf99-455d-99de-df3459f6e8c1-kube-api-access-8c8sw" (OuterVolumeSpecName: "kube-api-access-8c8sw") pod "a45ca231-bf99-455d-99de-df3459f6e8c1" (UID: "a45ca231-bf99-455d-99de-df3459f6e8c1"). InnerVolumeSpecName "kube-api-access-8c8sw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:18:35 crc kubenswrapper[4716]: I1207 16:18:35.211416 4716 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/a45ca231-bf99-455d-99de-df3459f6e8c1-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:35 crc kubenswrapper[4716]: I1207 16:18:35.211775 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8c8sw\" (UniqueName: \"kubernetes.io/projected/a45ca231-bf99-455d-99de-df3459f6e8c1-kube-api-access-8c8sw\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:35 crc kubenswrapper[4716]: I1207 16:18:35.211791 4716 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a45ca231-bf99-455d-99de-df3459f6e8c1-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:35 crc kubenswrapper[4716]: I1207 16:18:35.719685 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a714bbaa-9626-4b49-92ae-2abb029408d3","Type":"ContainerStarted","Data":"c993bdd24acae82bce459eae68683138a353d856ad3be59f8a30db03125538b6"} Dec 07 16:18:35 crc kubenswrapper[4716]: I1207 16:18:35.719735 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a714bbaa-9626-4b49-92ae-2abb029408d3","Type":"ContainerStarted","Data":"63ad4f8fbf5c4ff5bae0da56bcc01da72c0d4e7b35460a61cbd4ed8045c09f69"} Dec 07 16:18:35 crc kubenswrapper[4716]: I1207 16:18:35.719747 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a714bbaa-9626-4b49-92ae-2abb029408d3","Type":"ContainerStarted","Data":"6a887503f5e5edca650f8a18c504dfa18015d609ebf4a0fe10bfb1fe6afff394"} Dec 07 16:18:35 crc kubenswrapper[4716]: I1207 16:18:35.721790 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-tgl27-config-8c6m2" event={"ID":"a45ca231-bf99-455d-99de-df3459f6e8c1","Type":"ContainerDied","Data":"5940c7306c7511e0dce7253051e78164edc449a4ce94b67bc2f7dedf299072a9"} Dec 07 16:18:35 crc kubenswrapper[4716]: I1207 16:18:35.722006 4716 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5940c7306c7511e0dce7253051e78164edc449a4ce94b67bc2f7dedf299072a9" Dec 07 16:18:35 crc kubenswrapper[4716]: I1207 16:18:35.722017 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-tgl27-config-8c6m2" Dec 07 16:18:36 crc kubenswrapper[4716]: I1207 16:18:36.150988 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-tgl27-config-8c6m2"] Dec 07 16:18:36 crc kubenswrapper[4716]: I1207 16:18:36.159752 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-tgl27-config-8c6m2"] Dec 07 16:18:36 crc kubenswrapper[4716]: I1207 16:18:36.252009 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-tgl27-config-fxwks"] Dec 07 16:18:36 crc kubenswrapper[4716]: E1207 16:18:36.253920 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a45ca231-bf99-455d-99de-df3459f6e8c1" containerName="ovn-config" Dec 07 16:18:36 crc kubenswrapper[4716]: I1207 16:18:36.254008 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="a45ca231-bf99-455d-99de-df3459f6e8c1" containerName="ovn-config" Dec 07 16:18:36 crc kubenswrapper[4716]: I1207 16:18:36.254313 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="a45ca231-bf99-455d-99de-df3459f6e8c1" containerName="ovn-config" Dec 07 16:18:36 crc kubenswrapper[4716]: I1207 16:18:36.254993 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-tgl27-config-fxwks" Dec 07 16:18:36 crc kubenswrapper[4716]: I1207 16:18:36.257921 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 07 16:18:36 crc kubenswrapper[4716]: I1207 16:18:36.265567 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-tgl27-config-fxwks"] Dec 07 16:18:36 crc kubenswrapper[4716]: I1207 16:18:36.310491 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-tgl27" Dec 07 16:18:36 crc kubenswrapper[4716]: I1207 16:18:36.329988 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/3db80f80-2634-47c4-ad4f-7df066de39d8-var-run-ovn\") pod \"ovn-controller-tgl27-config-fxwks\" (UID: \"3db80f80-2634-47c4-ad4f-7df066de39d8\") " pod="openstack/ovn-controller-tgl27-config-fxwks" Dec 07 16:18:36 crc kubenswrapper[4716]: I1207 16:18:36.330029 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/3db80f80-2634-47c4-ad4f-7df066de39d8-var-run\") pod \"ovn-controller-tgl27-config-fxwks\" (UID: \"3db80f80-2634-47c4-ad4f-7df066de39d8\") " pod="openstack/ovn-controller-tgl27-config-fxwks" Dec 07 16:18:36 crc kubenswrapper[4716]: I1207 16:18:36.330058 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/3db80f80-2634-47c4-ad4f-7df066de39d8-var-log-ovn\") pod \"ovn-controller-tgl27-config-fxwks\" (UID: \"3db80f80-2634-47c4-ad4f-7df066de39d8\") " pod="openstack/ovn-controller-tgl27-config-fxwks" Dec 07 16:18:36 crc kubenswrapper[4716]: I1207 16:18:36.330182 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqsf7\" (UniqueName: \"kubernetes.io/projected/3db80f80-2634-47c4-ad4f-7df066de39d8-kube-api-access-dqsf7\") pod \"ovn-controller-tgl27-config-fxwks\" (UID: \"3db80f80-2634-47c4-ad4f-7df066de39d8\") " pod="openstack/ovn-controller-tgl27-config-fxwks" Dec 07 16:18:36 crc kubenswrapper[4716]: I1207 16:18:36.330260 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3db80f80-2634-47c4-ad4f-7df066de39d8-scripts\") pod \"ovn-controller-tgl27-config-fxwks\" (UID: \"3db80f80-2634-47c4-ad4f-7df066de39d8\") " pod="openstack/ovn-controller-tgl27-config-fxwks" Dec 07 16:18:36 crc kubenswrapper[4716]: I1207 16:18:36.330276 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/3db80f80-2634-47c4-ad4f-7df066de39d8-additional-scripts\") pod \"ovn-controller-tgl27-config-fxwks\" (UID: \"3db80f80-2634-47c4-ad4f-7df066de39d8\") " pod="openstack/ovn-controller-tgl27-config-fxwks" Dec 07 16:18:36 crc kubenswrapper[4716]: I1207 16:18:36.431774 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/3db80f80-2634-47c4-ad4f-7df066de39d8-var-log-ovn\") pod \"ovn-controller-tgl27-config-fxwks\" (UID: \"3db80f80-2634-47c4-ad4f-7df066de39d8\") " pod="openstack/ovn-controller-tgl27-config-fxwks" Dec 07 16:18:36 crc kubenswrapper[4716]: I1207 16:18:36.431872 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqsf7\" (UniqueName: \"kubernetes.io/projected/3db80f80-2634-47c4-ad4f-7df066de39d8-kube-api-access-dqsf7\") pod \"ovn-controller-tgl27-config-fxwks\" (UID: \"3db80f80-2634-47c4-ad4f-7df066de39d8\") " pod="openstack/ovn-controller-tgl27-config-fxwks" Dec 07 16:18:36 crc kubenswrapper[4716]: I1207 16:18:36.431932 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3db80f80-2634-47c4-ad4f-7df066de39d8-scripts\") pod \"ovn-controller-tgl27-config-fxwks\" (UID: \"3db80f80-2634-47c4-ad4f-7df066de39d8\") " pod="openstack/ovn-controller-tgl27-config-fxwks" Dec 07 16:18:36 crc kubenswrapper[4716]: I1207 16:18:36.431951 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/3db80f80-2634-47c4-ad4f-7df066de39d8-additional-scripts\") pod \"ovn-controller-tgl27-config-fxwks\" (UID: \"3db80f80-2634-47c4-ad4f-7df066de39d8\") " pod="openstack/ovn-controller-tgl27-config-fxwks" Dec 07 16:18:36 crc kubenswrapper[4716]: I1207 16:18:36.432013 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/3db80f80-2634-47c4-ad4f-7df066de39d8-var-run-ovn\") pod \"ovn-controller-tgl27-config-fxwks\" (UID: \"3db80f80-2634-47c4-ad4f-7df066de39d8\") " pod="openstack/ovn-controller-tgl27-config-fxwks" Dec 07 16:18:36 crc kubenswrapper[4716]: I1207 16:18:36.432029 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/3db80f80-2634-47c4-ad4f-7df066de39d8-var-run\") pod \"ovn-controller-tgl27-config-fxwks\" (UID: \"3db80f80-2634-47c4-ad4f-7df066de39d8\") " pod="openstack/ovn-controller-tgl27-config-fxwks" Dec 07 16:18:36 crc kubenswrapper[4716]: I1207 16:18:36.432179 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/3db80f80-2634-47c4-ad4f-7df066de39d8-var-log-ovn\") pod \"ovn-controller-tgl27-config-fxwks\" (UID: \"3db80f80-2634-47c4-ad4f-7df066de39d8\") " pod="openstack/ovn-controller-tgl27-config-fxwks" Dec 07 16:18:36 crc kubenswrapper[4716]: I1207 16:18:36.432259 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/3db80f80-2634-47c4-ad4f-7df066de39d8-var-run\") pod \"ovn-controller-tgl27-config-fxwks\" (UID: \"3db80f80-2634-47c4-ad4f-7df066de39d8\") " pod="openstack/ovn-controller-tgl27-config-fxwks" Dec 07 16:18:36 crc kubenswrapper[4716]: I1207 16:18:36.432885 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/3db80f80-2634-47c4-ad4f-7df066de39d8-var-run-ovn\") pod \"ovn-controller-tgl27-config-fxwks\" (UID: \"3db80f80-2634-47c4-ad4f-7df066de39d8\") " pod="openstack/ovn-controller-tgl27-config-fxwks" Dec 07 16:18:36 crc kubenswrapper[4716]: I1207 16:18:36.433018 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/3db80f80-2634-47c4-ad4f-7df066de39d8-additional-scripts\") pod \"ovn-controller-tgl27-config-fxwks\" (UID: \"3db80f80-2634-47c4-ad4f-7df066de39d8\") " pod="openstack/ovn-controller-tgl27-config-fxwks" Dec 07 16:18:36 crc kubenswrapper[4716]: I1207 16:18:36.434022 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3db80f80-2634-47c4-ad4f-7df066de39d8-scripts\") pod \"ovn-controller-tgl27-config-fxwks\" (UID: \"3db80f80-2634-47c4-ad4f-7df066de39d8\") " pod="openstack/ovn-controller-tgl27-config-fxwks" Dec 07 16:18:36 crc kubenswrapper[4716]: I1207 16:18:36.461549 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqsf7\" (UniqueName: \"kubernetes.io/projected/3db80f80-2634-47c4-ad4f-7df066de39d8-kube-api-access-dqsf7\") pod \"ovn-controller-tgl27-config-fxwks\" (UID: \"3db80f80-2634-47c4-ad4f-7df066de39d8\") " pod="openstack/ovn-controller-tgl27-config-fxwks" Dec 07 16:18:36 crc kubenswrapper[4716]: I1207 16:18:36.597398 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-tgl27-config-fxwks" Dec 07 16:18:36 crc kubenswrapper[4716]: I1207 16:18:36.733787 4716 generic.go:334] "Generic (PLEG): container finished" podID="a17893a0-50ef-4d05-8aaf-c6a9e3f8d20d" containerID="ffa4933c139800dce7b364c70ad2bd22dc6cb74649ec6f2f4c0be145c6b55016" exitCode=0 Dec 07 16:18:36 crc kubenswrapper[4716]: I1207 16:18:36.733823 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-gq62d" event={"ID":"a17893a0-50ef-4d05-8aaf-c6a9e3f8d20d","Type":"ContainerDied","Data":"ffa4933c139800dce7b364c70ad2bd22dc6cb74649ec6f2f4c0be145c6b55016"} Dec 07 16:18:37 crc kubenswrapper[4716]: I1207 16:18:37.077965 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-tgl27-config-fxwks"] Dec 07 16:18:37 crc kubenswrapper[4716]: W1207 16:18:37.088041 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3db80f80_2634_47c4_ad4f_7df066de39d8.slice/crio-c95e6e3fd98615f462378b5f8f1f006f3ff783898923abb6e29ed364b0adcd9e WatchSource:0}: Error finding container c95e6e3fd98615f462378b5f8f1f006f3ff783898923abb6e29ed364b0adcd9e: Status 404 returned error can't find the container with id c95e6e3fd98615f462378b5f8f1f006f3ff783898923abb6e29ed364b0adcd9e Dec 07 16:18:37 crc kubenswrapper[4716]: I1207 16:18:37.668430 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a45ca231-bf99-455d-99de-df3459f6e8c1" path="/var/lib/kubelet/pods/a45ca231-bf99-455d-99de-df3459f6e8c1/volumes" Dec 07 16:18:37 crc kubenswrapper[4716]: I1207 16:18:37.741840 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-tgl27-config-fxwks" event={"ID":"3db80f80-2634-47c4-ad4f-7df066de39d8","Type":"ContainerStarted","Data":"f4bac783db6117885371eea1a50ac5ca886c3831c5a98dd5a59c5aaaf4c1c5cd"} Dec 07 16:18:37 crc kubenswrapper[4716]: I1207 16:18:37.741899 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-tgl27-config-fxwks" event={"ID":"3db80f80-2634-47c4-ad4f-7df066de39d8","Type":"ContainerStarted","Data":"c95e6e3fd98615f462378b5f8f1f006f3ff783898923abb6e29ed364b0adcd9e"} Dec 07 16:18:37 crc kubenswrapper[4716]: I1207 16:18:37.746597 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a714bbaa-9626-4b49-92ae-2abb029408d3","Type":"ContainerStarted","Data":"77cd0116fdb20942ce185af0ad9867f78c66a054293f487c25cffcd494f3ab7c"} Dec 07 16:18:37 crc kubenswrapper[4716]: I1207 16:18:37.746639 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a714bbaa-9626-4b49-92ae-2abb029408d3","Type":"ContainerStarted","Data":"1ff83588ebbc4b59e265e3dc7fc72f8cc11292a51f6a45785fbaa19465b5c21b"} Dec 07 16:18:37 crc kubenswrapper[4716]: I1207 16:18:37.746667 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a714bbaa-9626-4b49-92ae-2abb029408d3","Type":"ContainerStarted","Data":"1f72673f9f847a35b8b61ef60b60be555d896e83ed0749d727a37fb92dbac114"} Dec 07 16:18:37 crc kubenswrapper[4716]: I1207 16:18:37.746681 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a714bbaa-9626-4b49-92ae-2abb029408d3","Type":"ContainerStarted","Data":"e1622e4fc843038aaa3cc795d9ca172521ea30450ae9fccc480ad0287ea19bd0"} Dec 07 16:18:37 crc kubenswrapper[4716]: I1207 16:18:37.746690 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a714bbaa-9626-4b49-92ae-2abb029408d3","Type":"ContainerStarted","Data":"51758b3a67e007000eca4d8284c0f27ebedd9ca8514eb650212b10632520bd88"} Dec 07 16:18:37 crc kubenswrapper[4716]: I1207 16:18:37.777780 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-tgl27-config-fxwks" podStartSLOduration=1.777756171 podStartE2EDuration="1.777756171s" podCreationTimestamp="2025-12-07 16:18:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:18:37.768456859 +0000 UTC m=+980.458741771" watchObservedRunningTime="2025-12-07 16:18:37.777756171 +0000 UTC m=+980.468041083" Dec 07 16:18:38 crc kubenswrapper[4716]: I1207 16:18:38.208980 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-gq62d" Dec 07 16:18:38 crc kubenswrapper[4716]: I1207 16:18:38.272047 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sp8kc\" (UniqueName: \"kubernetes.io/projected/a17893a0-50ef-4d05-8aaf-c6a9e3f8d20d-kube-api-access-sp8kc\") pod \"a17893a0-50ef-4d05-8aaf-c6a9e3f8d20d\" (UID: \"a17893a0-50ef-4d05-8aaf-c6a9e3f8d20d\") " Dec 07 16:18:38 crc kubenswrapper[4716]: I1207 16:18:38.272247 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a17893a0-50ef-4d05-8aaf-c6a9e3f8d20d-combined-ca-bundle\") pod \"a17893a0-50ef-4d05-8aaf-c6a9e3f8d20d\" (UID: \"a17893a0-50ef-4d05-8aaf-c6a9e3f8d20d\") " Dec 07 16:18:38 crc kubenswrapper[4716]: I1207 16:18:38.272282 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a17893a0-50ef-4d05-8aaf-c6a9e3f8d20d-db-sync-config-data\") pod \"a17893a0-50ef-4d05-8aaf-c6a9e3f8d20d\" (UID: \"a17893a0-50ef-4d05-8aaf-c6a9e3f8d20d\") " Dec 07 16:18:38 crc kubenswrapper[4716]: I1207 16:18:38.272359 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a17893a0-50ef-4d05-8aaf-c6a9e3f8d20d-config-data\") pod \"a17893a0-50ef-4d05-8aaf-c6a9e3f8d20d\" (UID: \"a17893a0-50ef-4d05-8aaf-c6a9e3f8d20d\") " Dec 07 16:18:38 crc kubenswrapper[4716]: I1207 16:18:38.278936 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a17893a0-50ef-4d05-8aaf-c6a9e3f8d20d-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "a17893a0-50ef-4d05-8aaf-c6a9e3f8d20d" (UID: "a17893a0-50ef-4d05-8aaf-c6a9e3f8d20d"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:18:38 crc kubenswrapper[4716]: I1207 16:18:38.279094 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a17893a0-50ef-4d05-8aaf-c6a9e3f8d20d-kube-api-access-sp8kc" (OuterVolumeSpecName: "kube-api-access-sp8kc") pod "a17893a0-50ef-4d05-8aaf-c6a9e3f8d20d" (UID: "a17893a0-50ef-4d05-8aaf-c6a9e3f8d20d"). InnerVolumeSpecName "kube-api-access-sp8kc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:18:38 crc kubenswrapper[4716]: I1207 16:18:38.301025 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a17893a0-50ef-4d05-8aaf-c6a9e3f8d20d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a17893a0-50ef-4d05-8aaf-c6a9e3f8d20d" (UID: "a17893a0-50ef-4d05-8aaf-c6a9e3f8d20d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:18:38 crc kubenswrapper[4716]: I1207 16:18:38.322972 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a17893a0-50ef-4d05-8aaf-c6a9e3f8d20d-config-data" (OuterVolumeSpecName: "config-data") pod "a17893a0-50ef-4d05-8aaf-c6a9e3f8d20d" (UID: "a17893a0-50ef-4d05-8aaf-c6a9e3f8d20d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:18:38 crc kubenswrapper[4716]: I1207 16:18:38.374764 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sp8kc\" (UniqueName: \"kubernetes.io/projected/a17893a0-50ef-4d05-8aaf-c6a9e3f8d20d-kube-api-access-sp8kc\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:38 crc kubenswrapper[4716]: I1207 16:18:38.374817 4716 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a17893a0-50ef-4d05-8aaf-c6a9e3f8d20d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:38 crc kubenswrapper[4716]: I1207 16:18:38.374826 4716 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a17893a0-50ef-4d05-8aaf-c6a9e3f8d20d-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:38 crc kubenswrapper[4716]: I1207 16:18:38.374835 4716 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a17893a0-50ef-4d05-8aaf-c6a9e3f8d20d-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:38 crc kubenswrapper[4716]: I1207 16:18:38.766751 4716 generic.go:334] "Generic (PLEG): container finished" podID="3db80f80-2634-47c4-ad4f-7df066de39d8" containerID="f4bac783db6117885371eea1a50ac5ca886c3831c5a98dd5a59c5aaaf4c1c5cd" exitCode=0 Dec 07 16:18:38 crc kubenswrapper[4716]: I1207 16:18:38.766819 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-tgl27-config-fxwks" event={"ID":"3db80f80-2634-47c4-ad4f-7df066de39d8","Type":"ContainerDied","Data":"f4bac783db6117885371eea1a50ac5ca886c3831c5a98dd5a59c5aaaf4c1c5cd"} Dec 07 16:18:38 crc kubenswrapper[4716]: I1207 16:18:38.769895 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-gq62d" event={"ID":"a17893a0-50ef-4d05-8aaf-c6a9e3f8d20d","Type":"ContainerDied","Data":"3c57a7abf3687f77612f42313bd678454e1b3ea2a3f21ab4e09349fb5536b2bb"} Dec 07 16:18:38 crc kubenswrapper[4716]: I1207 16:18:38.769935 4716 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3c57a7abf3687f77612f42313bd678454e1b3ea2a3f21ab4e09349fb5536b2bb" Dec 07 16:18:38 crc kubenswrapper[4716]: I1207 16:18:38.769980 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-gq62d" Dec 07 16:18:39 crc kubenswrapper[4716]: I1207 16:18:39.235804 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-qrmc9"] Dec 07 16:18:39 crc kubenswrapper[4716]: E1207 16:18:39.238963 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a17893a0-50ef-4d05-8aaf-c6a9e3f8d20d" containerName="glance-db-sync" Dec 07 16:18:39 crc kubenswrapper[4716]: I1207 16:18:39.238987 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="a17893a0-50ef-4d05-8aaf-c6a9e3f8d20d" containerName="glance-db-sync" Dec 07 16:18:39 crc kubenswrapper[4716]: I1207 16:18:39.239194 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="a17893a0-50ef-4d05-8aaf-c6a9e3f8d20d" containerName="glance-db-sync" Dec 07 16:18:39 crc kubenswrapper[4716]: I1207 16:18:39.240021 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b946c75cc-qrmc9" Dec 07 16:18:39 crc kubenswrapper[4716]: I1207 16:18:39.291141 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-qrmc9"] Dec 07 16:18:39 crc kubenswrapper[4716]: I1207 16:18:39.390357 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ce44eee2-5f6e-48f2-abfe-260aedab8b69-ovsdbserver-nb\") pod \"dnsmasq-dns-5b946c75cc-qrmc9\" (UID: \"ce44eee2-5f6e-48f2-abfe-260aedab8b69\") " pod="openstack/dnsmasq-dns-5b946c75cc-qrmc9" Dec 07 16:18:39 crc kubenswrapper[4716]: I1207 16:18:39.390445 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ce44eee2-5f6e-48f2-abfe-260aedab8b69-dns-svc\") pod \"dnsmasq-dns-5b946c75cc-qrmc9\" (UID: \"ce44eee2-5f6e-48f2-abfe-260aedab8b69\") " pod="openstack/dnsmasq-dns-5b946c75cc-qrmc9" Dec 07 16:18:39 crc kubenswrapper[4716]: I1207 16:18:39.390470 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ce44eee2-5f6e-48f2-abfe-260aedab8b69-ovsdbserver-sb\") pod \"dnsmasq-dns-5b946c75cc-qrmc9\" (UID: \"ce44eee2-5f6e-48f2-abfe-260aedab8b69\") " pod="openstack/dnsmasq-dns-5b946c75cc-qrmc9" Dec 07 16:18:39 crc kubenswrapper[4716]: I1207 16:18:39.390493 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce44eee2-5f6e-48f2-abfe-260aedab8b69-config\") pod \"dnsmasq-dns-5b946c75cc-qrmc9\" (UID: \"ce44eee2-5f6e-48f2-abfe-260aedab8b69\") " pod="openstack/dnsmasq-dns-5b946c75cc-qrmc9" Dec 07 16:18:39 crc kubenswrapper[4716]: I1207 16:18:39.390530 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27m9d\" (UniqueName: \"kubernetes.io/projected/ce44eee2-5f6e-48f2-abfe-260aedab8b69-kube-api-access-27m9d\") pod \"dnsmasq-dns-5b946c75cc-qrmc9\" (UID: \"ce44eee2-5f6e-48f2-abfe-260aedab8b69\") " pod="openstack/dnsmasq-dns-5b946c75cc-qrmc9" Dec 07 16:18:39 crc kubenswrapper[4716]: I1207 16:18:39.491881 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ce44eee2-5f6e-48f2-abfe-260aedab8b69-ovsdbserver-nb\") pod \"dnsmasq-dns-5b946c75cc-qrmc9\" (UID: \"ce44eee2-5f6e-48f2-abfe-260aedab8b69\") " pod="openstack/dnsmasq-dns-5b946c75cc-qrmc9" Dec 07 16:18:39 crc kubenswrapper[4716]: I1207 16:18:39.491966 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ce44eee2-5f6e-48f2-abfe-260aedab8b69-dns-svc\") pod \"dnsmasq-dns-5b946c75cc-qrmc9\" (UID: \"ce44eee2-5f6e-48f2-abfe-260aedab8b69\") " pod="openstack/dnsmasq-dns-5b946c75cc-qrmc9" Dec 07 16:18:39 crc kubenswrapper[4716]: I1207 16:18:39.491990 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ce44eee2-5f6e-48f2-abfe-260aedab8b69-ovsdbserver-sb\") pod \"dnsmasq-dns-5b946c75cc-qrmc9\" (UID: \"ce44eee2-5f6e-48f2-abfe-260aedab8b69\") " pod="openstack/dnsmasq-dns-5b946c75cc-qrmc9" Dec 07 16:18:39 crc kubenswrapper[4716]: I1207 16:18:39.492013 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27m9d\" (UniqueName: \"kubernetes.io/projected/ce44eee2-5f6e-48f2-abfe-260aedab8b69-kube-api-access-27m9d\") pod \"dnsmasq-dns-5b946c75cc-qrmc9\" (UID: \"ce44eee2-5f6e-48f2-abfe-260aedab8b69\") " pod="openstack/dnsmasq-dns-5b946c75cc-qrmc9" Dec 07 16:18:39 crc kubenswrapper[4716]: I1207 16:18:39.492031 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce44eee2-5f6e-48f2-abfe-260aedab8b69-config\") pod \"dnsmasq-dns-5b946c75cc-qrmc9\" (UID: \"ce44eee2-5f6e-48f2-abfe-260aedab8b69\") " pod="openstack/dnsmasq-dns-5b946c75cc-qrmc9" Dec 07 16:18:39 crc kubenswrapper[4716]: I1207 16:18:39.493002 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce44eee2-5f6e-48f2-abfe-260aedab8b69-config\") pod \"dnsmasq-dns-5b946c75cc-qrmc9\" (UID: \"ce44eee2-5f6e-48f2-abfe-260aedab8b69\") " pod="openstack/dnsmasq-dns-5b946c75cc-qrmc9" Dec 07 16:18:39 crc kubenswrapper[4716]: I1207 16:18:39.493157 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ce44eee2-5f6e-48f2-abfe-260aedab8b69-dns-svc\") pod \"dnsmasq-dns-5b946c75cc-qrmc9\" (UID: \"ce44eee2-5f6e-48f2-abfe-260aedab8b69\") " pod="openstack/dnsmasq-dns-5b946c75cc-qrmc9" Dec 07 16:18:39 crc kubenswrapper[4716]: I1207 16:18:39.493544 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ce44eee2-5f6e-48f2-abfe-260aedab8b69-ovsdbserver-sb\") pod \"dnsmasq-dns-5b946c75cc-qrmc9\" (UID: \"ce44eee2-5f6e-48f2-abfe-260aedab8b69\") " pod="openstack/dnsmasq-dns-5b946c75cc-qrmc9" Dec 07 16:18:39 crc kubenswrapper[4716]: I1207 16:18:39.493790 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ce44eee2-5f6e-48f2-abfe-260aedab8b69-ovsdbserver-nb\") pod \"dnsmasq-dns-5b946c75cc-qrmc9\" (UID: \"ce44eee2-5f6e-48f2-abfe-260aedab8b69\") " pod="openstack/dnsmasq-dns-5b946c75cc-qrmc9" Dec 07 16:18:39 crc kubenswrapper[4716]: I1207 16:18:39.537217 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27m9d\" (UniqueName: \"kubernetes.io/projected/ce44eee2-5f6e-48f2-abfe-260aedab8b69-kube-api-access-27m9d\") pod \"dnsmasq-dns-5b946c75cc-qrmc9\" (UID: \"ce44eee2-5f6e-48f2-abfe-260aedab8b69\") " pod="openstack/dnsmasq-dns-5b946c75cc-qrmc9" Dec 07 16:18:39 crc kubenswrapper[4716]: I1207 16:18:39.624658 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b946c75cc-qrmc9" Dec 07 16:18:39 crc kubenswrapper[4716]: I1207 16:18:39.793150 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a714bbaa-9626-4b49-92ae-2abb029408d3","Type":"ContainerStarted","Data":"86f93e452e530b95bc3b1fc46b06d12fab16a003943cb93f881f0c9f141040f6"} Dec 07 16:18:39 crc kubenswrapper[4716]: I1207 16:18:39.793200 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a714bbaa-9626-4b49-92ae-2abb029408d3","Type":"ContainerStarted","Data":"cac7c7a62023412b6b18c55cb372170d6e6dfd362e27679a459a6ca339ddc766"} Dec 07 16:18:39 crc kubenswrapper[4716]: I1207 16:18:39.846612 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=21.750205388 podStartE2EDuration="27.846585154s" podCreationTimestamp="2025-12-07 16:18:12 +0000 UTC" firstStartedPulling="2025-12-07 16:18:30.481394165 +0000 UTC m=+973.171679077" lastFinishedPulling="2025-12-07 16:18:36.577773931 +0000 UTC m=+979.268058843" observedRunningTime="2025-12-07 16:18:39.834173307 +0000 UTC m=+982.524458219" watchObservedRunningTime="2025-12-07 16:18:39.846585154 +0000 UTC m=+982.536870066" Dec 07 16:18:40 crc kubenswrapper[4716]: I1207 16:18:40.133891 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-qrmc9"] Dec 07 16:18:40 crc kubenswrapper[4716]: I1207 16:18:40.140241 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-tgl27-config-fxwks" Dec 07 16:18:40 crc kubenswrapper[4716]: I1207 16:18:40.157461 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-2cv95"] Dec 07 16:18:40 crc kubenswrapper[4716]: E1207 16:18:40.157846 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3db80f80-2634-47c4-ad4f-7df066de39d8" containerName="ovn-config" Dec 07 16:18:40 crc kubenswrapper[4716]: I1207 16:18:40.157861 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="3db80f80-2634-47c4-ad4f-7df066de39d8" containerName="ovn-config" Dec 07 16:18:40 crc kubenswrapper[4716]: I1207 16:18:40.158048 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="3db80f80-2634-47c4-ad4f-7df066de39d8" containerName="ovn-config" Dec 07 16:18:40 crc kubenswrapper[4716]: I1207 16:18:40.159161 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-2cv95" Dec 07 16:18:40 crc kubenswrapper[4716]: I1207 16:18:40.161990 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Dec 07 16:18:40 crc kubenswrapper[4716]: I1207 16:18:40.176231 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-2cv95"] Dec 07 16:18:40 crc kubenswrapper[4716]: I1207 16:18:40.194184 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-qrmc9"] Dec 07 16:18:40 crc kubenswrapper[4716]: I1207 16:18:40.213663 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/3db80f80-2634-47c4-ad4f-7df066de39d8-var-log-ovn\") pod \"3db80f80-2634-47c4-ad4f-7df066de39d8\" (UID: \"3db80f80-2634-47c4-ad4f-7df066de39d8\") " Dec 07 16:18:40 crc kubenswrapper[4716]: I1207 16:18:40.213754 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dqsf7\" (UniqueName: \"kubernetes.io/projected/3db80f80-2634-47c4-ad4f-7df066de39d8-kube-api-access-dqsf7\") pod \"3db80f80-2634-47c4-ad4f-7df066de39d8\" (UID: \"3db80f80-2634-47c4-ad4f-7df066de39d8\") " Dec 07 16:18:40 crc kubenswrapper[4716]: I1207 16:18:40.213809 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/3db80f80-2634-47c4-ad4f-7df066de39d8-var-run\") pod \"3db80f80-2634-47c4-ad4f-7df066de39d8\" (UID: \"3db80f80-2634-47c4-ad4f-7df066de39d8\") " Dec 07 16:18:40 crc kubenswrapper[4716]: I1207 16:18:40.213873 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/3db80f80-2634-47c4-ad4f-7df066de39d8-additional-scripts\") pod \"3db80f80-2634-47c4-ad4f-7df066de39d8\" (UID: \"3db80f80-2634-47c4-ad4f-7df066de39d8\") " Dec 07 16:18:40 crc kubenswrapper[4716]: I1207 16:18:40.213939 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/3db80f80-2634-47c4-ad4f-7df066de39d8-var-run-ovn\") pod \"3db80f80-2634-47c4-ad4f-7df066de39d8\" (UID: \"3db80f80-2634-47c4-ad4f-7df066de39d8\") " Dec 07 16:18:40 crc kubenswrapper[4716]: I1207 16:18:40.213974 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3db80f80-2634-47c4-ad4f-7df066de39d8-scripts\") pod \"3db80f80-2634-47c4-ad4f-7df066de39d8\" (UID: \"3db80f80-2634-47c4-ad4f-7df066de39d8\") " Dec 07 16:18:40 crc kubenswrapper[4716]: I1207 16:18:40.214130 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqqlg\" (UniqueName: \"kubernetes.io/projected/7a57dd80-9afb-4d22-bd71-7104bc574ab0-kube-api-access-dqqlg\") pod \"dnsmasq-dns-74f6bcbc87-2cv95\" (UID: \"7a57dd80-9afb-4d22-bd71-7104bc574ab0\") " pod="openstack/dnsmasq-dns-74f6bcbc87-2cv95" Dec 07 16:18:40 crc kubenswrapper[4716]: I1207 16:18:40.214218 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7a57dd80-9afb-4d22-bd71-7104bc574ab0-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-2cv95\" (UID: \"7a57dd80-9afb-4d22-bd71-7104bc574ab0\") " pod="openstack/dnsmasq-dns-74f6bcbc87-2cv95" Dec 07 16:18:40 crc kubenswrapper[4716]: I1207 16:18:40.214272 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a57dd80-9afb-4d22-bd71-7104bc574ab0-config\") pod \"dnsmasq-dns-74f6bcbc87-2cv95\" (UID: \"7a57dd80-9afb-4d22-bd71-7104bc574ab0\") " pod="openstack/dnsmasq-dns-74f6bcbc87-2cv95" Dec 07 16:18:40 crc kubenswrapper[4716]: I1207 16:18:40.214295 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7a57dd80-9afb-4d22-bd71-7104bc574ab0-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-2cv95\" (UID: \"7a57dd80-9afb-4d22-bd71-7104bc574ab0\") " pod="openstack/dnsmasq-dns-74f6bcbc87-2cv95" Dec 07 16:18:40 crc kubenswrapper[4716]: I1207 16:18:40.214313 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7a57dd80-9afb-4d22-bd71-7104bc574ab0-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-2cv95\" (UID: \"7a57dd80-9afb-4d22-bd71-7104bc574ab0\") " pod="openstack/dnsmasq-dns-74f6bcbc87-2cv95" Dec 07 16:18:40 crc kubenswrapper[4716]: I1207 16:18:40.214330 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7a57dd80-9afb-4d22-bd71-7104bc574ab0-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-2cv95\" (UID: \"7a57dd80-9afb-4d22-bd71-7104bc574ab0\") " pod="openstack/dnsmasq-dns-74f6bcbc87-2cv95" Dec 07 16:18:40 crc kubenswrapper[4716]: I1207 16:18:40.214412 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3db80f80-2634-47c4-ad4f-7df066de39d8-var-run" (OuterVolumeSpecName: "var-run") pod "3db80f80-2634-47c4-ad4f-7df066de39d8" (UID: "3db80f80-2634-47c4-ad4f-7df066de39d8"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 16:18:40 crc kubenswrapper[4716]: I1207 16:18:40.214547 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3db80f80-2634-47c4-ad4f-7df066de39d8-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "3db80f80-2634-47c4-ad4f-7df066de39d8" (UID: "3db80f80-2634-47c4-ad4f-7df066de39d8"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 16:18:40 crc kubenswrapper[4716]: I1207 16:18:40.214624 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3db80f80-2634-47c4-ad4f-7df066de39d8-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "3db80f80-2634-47c4-ad4f-7df066de39d8" (UID: "3db80f80-2634-47c4-ad4f-7df066de39d8"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 16:18:40 crc kubenswrapper[4716]: I1207 16:18:40.215036 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3db80f80-2634-47c4-ad4f-7df066de39d8-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "3db80f80-2634-47c4-ad4f-7df066de39d8" (UID: "3db80f80-2634-47c4-ad4f-7df066de39d8"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:18:40 crc kubenswrapper[4716]: I1207 16:18:40.215334 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3db80f80-2634-47c4-ad4f-7df066de39d8-scripts" (OuterVolumeSpecName: "scripts") pod "3db80f80-2634-47c4-ad4f-7df066de39d8" (UID: "3db80f80-2634-47c4-ad4f-7df066de39d8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:18:40 crc kubenswrapper[4716]: I1207 16:18:40.217336 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3db80f80-2634-47c4-ad4f-7df066de39d8-kube-api-access-dqsf7" (OuterVolumeSpecName: "kube-api-access-dqsf7") pod "3db80f80-2634-47c4-ad4f-7df066de39d8" (UID: "3db80f80-2634-47c4-ad4f-7df066de39d8"). InnerVolumeSpecName "kube-api-access-dqsf7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:18:40 crc kubenswrapper[4716]: I1207 16:18:40.315422 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a57dd80-9afb-4d22-bd71-7104bc574ab0-config\") pod \"dnsmasq-dns-74f6bcbc87-2cv95\" (UID: \"7a57dd80-9afb-4d22-bd71-7104bc574ab0\") " pod="openstack/dnsmasq-dns-74f6bcbc87-2cv95" Dec 07 16:18:40 crc kubenswrapper[4716]: I1207 16:18:40.315473 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7a57dd80-9afb-4d22-bd71-7104bc574ab0-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-2cv95\" (UID: \"7a57dd80-9afb-4d22-bd71-7104bc574ab0\") " pod="openstack/dnsmasq-dns-74f6bcbc87-2cv95" Dec 07 16:18:40 crc kubenswrapper[4716]: I1207 16:18:40.315492 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7a57dd80-9afb-4d22-bd71-7104bc574ab0-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-2cv95\" (UID: \"7a57dd80-9afb-4d22-bd71-7104bc574ab0\") " pod="openstack/dnsmasq-dns-74f6bcbc87-2cv95" Dec 07 16:18:40 crc kubenswrapper[4716]: I1207 16:18:40.315574 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7a57dd80-9afb-4d22-bd71-7104bc574ab0-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-2cv95\" (UID: \"7a57dd80-9afb-4d22-bd71-7104bc574ab0\") " pod="openstack/dnsmasq-dns-74f6bcbc87-2cv95" Dec 07 16:18:40 crc kubenswrapper[4716]: I1207 16:18:40.315601 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqqlg\" (UniqueName: \"kubernetes.io/projected/7a57dd80-9afb-4d22-bd71-7104bc574ab0-kube-api-access-dqqlg\") pod \"dnsmasq-dns-74f6bcbc87-2cv95\" (UID: \"7a57dd80-9afb-4d22-bd71-7104bc574ab0\") " pod="openstack/dnsmasq-dns-74f6bcbc87-2cv95" Dec 07 16:18:40 crc kubenswrapper[4716]: I1207 16:18:40.315668 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7a57dd80-9afb-4d22-bd71-7104bc574ab0-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-2cv95\" (UID: \"7a57dd80-9afb-4d22-bd71-7104bc574ab0\") " pod="openstack/dnsmasq-dns-74f6bcbc87-2cv95" Dec 07 16:18:40 crc kubenswrapper[4716]: I1207 16:18:40.315715 4716 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/3db80f80-2634-47c4-ad4f-7df066de39d8-var-run\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:40 crc kubenswrapper[4716]: I1207 16:18:40.315730 4716 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/3db80f80-2634-47c4-ad4f-7df066de39d8-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:40 crc kubenswrapper[4716]: I1207 16:18:40.315742 4716 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/3db80f80-2634-47c4-ad4f-7df066de39d8-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:40 crc kubenswrapper[4716]: I1207 16:18:40.315753 4716 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3db80f80-2634-47c4-ad4f-7df066de39d8-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:40 crc kubenswrapper[4716]: I1207 16:18:40.315825 4716 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/3db80f80-2634-47c4-ad4f-7df066de39d8-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:40 crc kubenswrapper[4716]: I1207 16:18:40.315859 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dqsf7\" (UniqueName: \"kubernetes.io/projected/3db80f80-2634-47c4-ad4f-7df066de39d8-kube-api-access-dqsf7\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:40 crc kubenswrapper[4716]: I1207 16:18:40.316941 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7a57dd80-9afb-4d22-bd71-7104bc574ab0-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-2cv95\" (UID: \"7a57dd80-9afb-4d22-bd71-7104bc574ab0\") " pod="openstack/dnsmasq-dns-74f6bcbc87-2cv95" Dec 07 16:18:40 crc kubenswrapper[4716]: I1207 16:18:40.317468 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a57dd80-9afb-4d22-bd71-7104bc574ab0-config\") pod \"dnsmasq-dns-74f6bcbc87-2cv95\" (UID: \"7a57dd80-9afb-4d22-bd71-7104bc574ab0\") " pod="openstack/dnsmasq-dns-74f6bcbc87-2cv95" Dec 07 16:18:40 crc kubenswrapper[4716]: I1207 16:18:40.318019 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7a57dd80-9afb-4d22-bd71-7104bc574ab0-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-2cv95\" (UID: \"7a57dd80-9afb-4d22-bd71-7104bc574ab0\") " pod="openstack/dnsmasq-dns-74f6bcbc87-2cv95" Dec 07 16:18:40 crc kubenswrapper[4716]: I1207 16:18:40.318593 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7a57dd80-9afb-4d22-bd71-7104bc574ab0-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-2cv95\" (UID: \"7a57dd80-9afb-4d22-bd71-7104bc574ab0\") " pod="openstack/dnsmasq-dns-74f6bcbc87-2cv95" Dec 07 16:18:40 crc kubenswrapper[4716]: I1207 16:18:40.320127 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7a57dd80-9afb-4d22-bd71-7104bc574ab0-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-2cv95\" (UID: \"7a57dd80-9afb-4d22-bd71-7104bc574ab0\") " pod="openstack/dnsmasq-dns-74f6bcbc87-2cv95" Dec 07 16:18:40 crc kubenswrapper[4716]: I1207 16:18:40.335751 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqqlg\" (UniqueName: \"kubernetes.io/projected/7a57dd80-9afb-4d22-bd71-7104bc574ab0-kube-api-access-dqqlg\") pod \"dnsmasq-dns-74f6bcbc87-2cv95\" (UID: \"7a57dd80-9afb-4d22-bd71-7104bc574ab0\") " pod="openstack/dnsmasq-dns-74f6bcbc87-2cv95" Dec 07 16:18:40 crc kubenswrapper[4716]: I1207 16:18:40.478146 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-2cv95" Dec 07 16:18:40 crc kubenswrapper[4716]: I1207 16:18:40.804204 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-tgl27-config-fxwks" event={"ID":"3db80f80-2634-47c4-ad4f-7df066de39d8","Type":"ContainerDied","Data":"c95e6e3fd98615f462378b5f8f1f006f3ff783898923abb6e29ed364b0adcd9e"} Dec 07 16:18:40 crc kubenswrapper[4716]: I1207 16:18:40.804691 4716 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c95e6e3fd98615f462378b5f8f1f006f3ff783898923abb6e29ed364b0adcd9e" Dec 07 16:18:40 crc kubenswrapper[4716]: I1207 16:18:40.804242 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-tgl27-config-fxwks" Dec 07 16:18:40 crc kubenswrapper[4716]: I1207 16:18:40.805856 4716 generic.go:334] "Generic (PLEG): container finished" podID="ce44eee2-5f6e-48f2-abfe-260aedab8b69" containerID="a0eb4c2ef996ab07ddec84d7f53c00e1bc2b98dfbeae664d1d7696f1b35dda1c" exitCode=0 Dec 07 16:18:40 crc kubenswrapper[4716]: I1207 16:18:40.806008 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b946c75cc-qrmc9" event={"ID":"ce44eee2-5f6e-48f2-abfe-260aedab8b69","Type":"ContainerDied","Data":"a0eb4c2ef996ab07ddec84d7f53c00e1bc2b98dfbeae664d1d7696f1b35dda1c"} Dec 07 16:18:40 crc kubenswrapper[4716]: I1207 16:18:40.806093 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b946c75cc-qrmc9" event={"ID":"ce44eee2-5f6e-48f2-abfe-260aedab8b69","Type":"ContainerStarted","Data":"2d1204bfa66c4ba5ed75fa8610848558d9c8d92b96a61030b8e09fe9a7ec63fa"} Dec 07 16:18:40 crc kubenswrapper[4716]: I1207 16:18:40.991379 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-2cv95"] Dec 07 16:18:41 crc kubenswrapper[4716]: I1207 16:18:41.159171 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b946c75cc-qrmc9" Dec 07 16:18:41 crc kubenswrapper[4716]: I1207 16:18:41.239902 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-tgl27-config-fxwks"] Dec 07 16:18:41 crc kubenswrapper[4716]: I1207 16:18:41.240937 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ce44eee2-5f6e-48f2-abfe-260aedab8b69-ovsdbserver-nb\") pod \"ce44eee2-5f6e-48f2-abfe-260aedab8b69\" (UID: \"ce44eee2-5f6e-48f2-abfe-260aedab8b69\") " Dec 07 16:18:41 crc kubenswrapper[4716]: I1207 16:18:41.241035 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ce44eee2-5f6e-48f2-abfe-260aedab8b69-dns-svc\") pod \"ce44eee2-5f6e-48f2-abfe-260aedab8b69\" (UID: \"ce44eee2-5f6e-48f2-abfe-260aedab8b69\") " Dec 07 16:18:41 crc kubenswrapper[4716]: I1207 16:18:41.241117 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce44eee2-5f6e-48f2-abfe-260aedab8b69-config\") pod \"ce44eee2-5f6e-48f2-abfe-260aedab8b69\" (UID: \"ce44eee2-5f6e-48f2-abfe-260aedab8b69\") " Dec 07 16:18:41 crc kubenswrapper[4716]: I1207 16:18:41.241194 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-27m9d\" (UniqueName: \"kubernetes.io/projected/ce44eee2-5f6e-48f2-abfe-260aedab8b69-kube-api-access-27m9d\") pod \"ce44eee2-5f6e-48f2-abfe-260aedab8b69\" (UID: \"ce44eee2-5f6e-48f2-abfe-260aedab8b69\") " Dec 07 16:18:41 crc kubenswrapper[4716]: I1207 16:18:41.241236 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ce44eee2-5f6e-48f2-abfe-260aedab8b69-ovsdbserver-sb\") pod \"ce44eee2-5f6e-48f2-abfe-260aedab8b69\" (UID: \"ce44eee2-5f6e-48f2-abfe-260aedab8b69\") " Dec 07 16:18:41 crc kubenswrapper[4716]: I1207 16:18:41.245436 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce44eee2-5f6e-48f2-abfe-260aedab8b69-kube-api-access-27m9d" (OuterVolumeSpecName: "kube-api-access-27m9d") pod "ce44eee2-5f6e-48f2-abfe-260aedab8b69" (UID: "ce44eee2-5f6e-48f2-abfe-260aedab8b69"). InnerVolumeSpecName "kube-api-access-27m9d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:18:41 crc kubenswrapper[4716]: I1207 16:18:41.248595 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-tgl27-config-fxwks"] Dec 07 16:18:41 crc kubenswrapper[4716]: I1207 16:18:41.274990 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce44eee2-5f6e-48f2-abfe-260aedab8b69-config" (OuterVolumeSpecName: "config") pod "ce44eee2-5f6e-48f2-abfe-260aedab8b69" (UID: "ce44eee2-5f6e-48f2-abfe-260aedab8b69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:18:41 crc kubenswrapper[4716]: I1207 16:18:41.281869 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce44eee2-5f6e-48f2-abfe-260aedab8b69-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ce44eee2-5f6e-48f2-abfe-260aedab8b69" (UID: "ce44eee2-5f6e-48f2-abfe-260aedab8b69"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:18:41 crc kubenswrapper[4716]: I1207 16:18:41.284819 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce44eee2-5f6e-48f2-abfe-260aedab8b69-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ce44eee2-5f6e-48f2-abfe-260aedab8b69" (UID: "ce44eee2-5f6e-48f2-abfe-260aedab8b69"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:18:41 crc kubenswrapper[4716]: I1207 16:18:41.291887 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce44eee2-5f6e-48f2-abfe-260aedab8b69-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ce44eee2-5f6e-48f2-abfe-260aedab8b69" (UID: "ce44eee2-5f6e-48f2-abfe-260aedab8b69"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:18:41 crc kubenswrapper[4716]: I1207 16:18:41.344002 4716 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ce44eee2-5f6e-48f2-abfe-260aedab8b69-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:41 crc kubenswrapper[4716]: I1207 16:18:41.344056 4716 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ce44eee2-5f6e-48f2-abfe-260aedab8b69-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:41 crc kubenswrapper[4716]: I1207 16:18:41.344066 4716 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce44eee2-5f6e-48f2-abfe-260aedab8b69-config\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:41 crc kubenswrapper[4716]: I1207 16:18:41.344147 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-27m9d\" (UniqueName: \"kubernetes.io/projected/ce44eee2-5f6e-48f2-abfe-260aedab8b69-kube-api-access-27m9d\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:41 crc kubenswrapper[4716]: I1207 16:18:41.344165 4716 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ce44eee2-5f6e-48f2-abfe-260aedab8b69-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:41 crc kubenswrapper[4716]: I1207 16:18:41.668899 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3db80f80-2634-47c4-ad4f-7df066de39d8" path="/var/lib/kubelet/pods/3db80f80-2634-47c4-ad4f-7df066de39d8/volumes" Dec 07 16:18:41 crc kubenswrapper[4716]: I1207 16:18:41.815682 4716 generic.go:334] "Generic (PLEG): container finished" podID="7a57dd80-9afb-4d22-bd71-7104bc574ab0" containerID="80cfd80e2499f2e974a5f55feaab69e7d3d1fc4484c0b6b5aa1f696d3561813c" exitCode=0 Dec 07 16:18:41 crc kubenswrapper[4716]: I1207 16:18:41.815776 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-2cv95" event={"ID":"7a57dd80-9afb-4d22-bd71-7104bc574ab0","Type":"ContainerDied","Data":"80cfd80e2499f2e974a5f55feaab69e7d3d1fc4484c0b6b5aa1f696d3561813c"} Dec 07 16:18:41 crc kubenswrapper[4716]: I1207 16:18:41.816147 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-2cv95" event={"ID":"7a57dd80-9afb-4d22-bd71-7104bc574ab0","Type":"ContainerStarted","Data":"57cf0bd73a383215f36e9929371ab9c687a55f290d3c8edd4083c3bed0d35241"} Dec 07 16:18:41 crc kubenswrapper[4716]: I1207 16:18:41.818486 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b946c75cc-qrmc9" event={"ID":"ce44eee2-5f6e-48f2-abfe-260aedab8b69","Type":"ContainerDied","Data":"2d1204bfa66c4ba5ed75fa8610848558d9c8d92b96a61030b8e09fe9a7ec63fa"} Dec 07 16:18:41 crc kubenswrapper[4716]: I1207 16:18:41.818546 4716 scope.go:117] "RemoveContainer" containerID="a0eb4c2ef996ab07ddec84d7f53c00e1bc2b98dfbeae664d1d7696f1b35dda1c" Dec 07 16:18:41 crc kubenswrapper[4716]: I1207 16:18:41.818595 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b946c75cc-qrmc9" Dec 07 16:18:41 crc kubenswrapper[4716]: I1207 16:18:41.926192 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-qrmc9"] Dec 07 16:18:41 crc kubenswrapper[4716]: I1207 16:18:41.934560 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-qrmc9"] Dec 07 16:18:42 crc kubenswrapper[4716]: I1207 16:18:42.827634 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-2cv95" event={"ID":"7a57dd80-9afb-4d22-bd71-7104bc574ab0","Type":"ContainerStarted","Data":"670d21ca07fde13d32e5cc95349a99ca26fd9a0580fa5a23a4f056dd9038b973"} Dec 07 16:18:42 crc kubenswrapper[4716]: I1207 16:18:42.827991 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-74f6bcbc87-2cv95" Dec 07 16:18:42 crc kubenswrapper[4716]: I1207 16:18:42.849040 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-74f6bcbc87-2cv95" podStartSLOduration=2.849018199 podStartE2EDuration="2.849018199s" podCreationTimestamp="2025-12-07 16:18:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:18:42.847485938 +0000 UTC m=+985.537770860" watchObservedRunningTime="2025-12-07 16:18:42.849018199 +0000 UTC m=+985.539303111" Dec 07 16:18:43 crc kubenswrapper[4716]: I1207 16:18:43.672799 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce44eee2-5f6e-48f2-abfe-260aedab8b69" path="/var/lib/kubelet/pods/ce44eee2-5f6e-48f2-abfe-260aedab8b69/volumes" Dec 07 16:18:47 crc kubenswrapper[4716]: I1207 16:18:47.389276 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 07 16:18:47 crc kubenswrapper[4716]: I1207 16:18:47.679263 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:18:47 crc kubenswrapper[4716]: I1207 16:18:47.745434 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-t7n4f"] Dec 07 16:18:47 crc kubenswrapper[4716]: E1207 16:18:47.746003 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce44eee2-5f6e-48f2-abfe-260aedab8b69" containerName="init" Dec 07 16:18:47 crc kubenswrapper[4716]: I1207 16:18:47.746022 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce44eee2-5f6e-48f2-abfe-260aedab8b69" containerName="init" Dec 07 16:18:47 crc kubenswrapper[4716]: I1207 16:18:47.746401 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce44eee2-5f6e-48f2-abfe-260aedab8b69" containerName="init" Dec 07 16:18:47 crc kubenswrapper[4716]: I1207 16:18:47.747046 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-t7n4f" Dec 07 16:18:47 crc kubenswrapper[4716]: I1207 16:18:47.793142 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-t7n4f"] Dec 07 16:18:47 crc kubenswrapper[4716]: I1207 16:18:47.861483 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-jct5c"] Dec 07 16:18:47 crc kubenswrapper[4716]: I1207 16:18:47.862680 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-jct5c" Dec 07 16:18:47 crc kubenswrapper[4716]: I1207 16:18:47.870634 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-9210-account-create-update-lvz68"] Dec 07 16:18:47 crc kubenswrapper[4716]: I1207 16:18:47.872201 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-9210-account-create-update-lvz68" Dec 07 16:18:47 crc kubenswrapper[4716]: I1207 16:18:47.880755 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-jct5c"] Dec 07 16:18:47 crc kubenswrapper[4716]: I1207 16:18:47.881523 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Dec 07 16:18:47 crc kubenswrapper[4716]: I1207 16:18:47.885632 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-9210-account-create-update-lvz68"] Dec 07 16:18:47 crc kubenswrapper[4716]: I1207 16:18:47.959907 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/50a30fe7-3fb4-4b9d-b186-139c2b48295b-operator-scripts\") pod \"cinder-db-create-t7n4f\" (UID: \"50a30fe7-3fb4-4b9d-b186-139c2b48295b\") " pod="openstack/cinder-db-create-t7n4f" Dec 07 16:18:47 crc kubenswrapper[4716]: I1207 16:18:47.959985 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9d1c048e-719c-4406-bfae-1e5a629534ed-operator-scripts\") pod \"barbican-db-create-jct5c\" (UID: \"9d1c048e-719c-4406-bfae-1e5a629534ed\") " pod="openstack/barbican-db-create-jct5c" Dec 07 16:18:47 crc kubenswrapper[4716]: I1207 16:18:47.960025 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4mzpj\" (UniqueName: \"kubernetes.io/projected/9d1c048e-719c-4406-bfae-1e5a629534ed-kube-api-access-4mzpj\") pod \"barbican-db-create-jct5c\" (UID: \"9d1c048e-719c-4406-bfae-1e5a629534ed\") " pod="openstack/barbican-db-create-jct5c" Dec 07 16:18:47 crc kubenswrapper[4716]: I1207 16:18:47.960043 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/34a0af5c-2914-4a67-8ede-f178f08d473f-operator-scripts\") pod \"barbican-9210-account-create-update-lvz68\" (UID: \"34a0af5c-2914-4a67-8ede-f178f08d473f\") " pod="openstack/barbican-9210-account-create-update-lvz68" Dec 07 16:18:47 crc kubenswrapper[4716]: I1207 16:18:47.960098 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5jjbx\" (UniqueName: \"kubernetes.io/projected/34a0af5c-2914-4a67-8ede-f178f08d473f-kube-api-access-5jjbx\") pod \"barbican-9210-account-create-update-lvz68\" (UID: \"34a0af5c-2914-4a67-8ede-f178f08d473f\") " pod="openstack/barbican-9210-account-create-update-lvz68" Dec 07 16:18:47 crc kubenswrapper[4716]: I1207 16:18:47.960124 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8xtlm\" (UniqueName: \"kubernetes.io/projected/50a30fe7-3fb4-4b9d-b186-139c2b48295b-kube-api-access-8xtlm\") pod \"cinder-db-create-t7n4f\" (UID: \"50a30fe7-3fb4-4b9d-b186-139c2b48295b\") " pod="openstack/cinder-db-create-t7n4f" Dec 07 16:18:47 crc kubenswrapper[4716]: I1207 16:18:47.961262 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-e8e1-account-create-update-frpnc"] Dec 07 16:18:47 crc kubenswrapper[4716]: I1207 16:18:47.962556 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-e8e1-account-create-update-frpnc" Dec 07 16:18:47 crc kubenswrapper[4716]: I1207 16:18:47.964634 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Dec 07 16:18:47 crc kubenswrapper[4716]: I1207 16:18:47.981246 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-e8e1-account-create-update-frpnc"] Dec 07 16:18:48 crc kubenswrapper[4716]: I1207 16:18:48.061901 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9d1c048e-719c-4406-bfae-1e5a629534ed-operator-scripts\") pod \"barbican-db-create-jct5c\" (UID: \"9d1c048e-719c-4406-bfae-1e5a629534ed\") " pod="openstack/barbican-db-create-jct5c" Dec 07 16:18:48 crc kubenswrapper[4716]: I1207 16:18:48.061949 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4mzpj\" (UniqueName: \"kubernetes.io/projected/9d1c048e-719c-4406-bfae-1e5a629534ed-kube-api-access-4mzpj\") pod \"barbican-db-create-jct5c\" (UID: \"9d1c048e-719c-4406-bfae-1e5a629534ed\") " pod="openstack/barbican-db-create-jct5c" Dec 07 16:18:48 crc kubenswrapper[4716]: I1207 16:18:48.061971 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/34a0af5c-2914-4a67-8ede-f178f08d473f-operator-scripts\") pod \"barbican-9210-account-create-update-lvz68\" (UID: \"34a0af5c-2914-4a67-8ede-f178f08d473f\") " pod="openstack/barbican-9210-account-create-update-lvz68" Dec 07 16:18:48 crc kubenswrapper[4716]: I1207 16:18:48.061993 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbfnz\" (UniqueName: \"kubernetes.io/projected/862acceb-bcf0-4ef4-9a53-df61d6084806-kube-api-access-rbfnz\") pod \"cinder-e8e1-account-create-update-frpnc\" (UID: \"862acceb-bcf0-4ef4-9a53-df61d6084806\") " pod="openstack/cinder-e8e1-account-create-update-frpnc" Dec 07 16:18:48 crc kubenswrapper[4716]: I1207 16:18:48.062032 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5jjbx\" (UniqueName: \"kubernetes.io/projected/34a0af5c-2914-4a67-8ede-f178f08d473f-kube-api-access-5jjbx\") pod \"barbican-9210-account-create-update-lvz68\" (UID: \"34a0af5c-2914-4a67-8ede-f178f08d473f\") " pod="openstack/barbican-9210-account-create-update-lvz68" Dec 07 16:18:48 crc kubenswrapper[4716]: I1207 16:18:48.062056 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8xtlm\" (UniqueName: \"kubernetes.io/projected/50a30fe7-3fb4-4b9d-b186-139c2b48295b-kube-api-access-8xtlm\") pod \"cinder-db-create-t7n4f\" (UID: \"50a30fe7-3fb4-4b9d-b186-139c2b48295b\") " pod="openstack/cinder-db-create-t7n4f" Dec 07 16:18:48 crc kubenswrapper[4716]: I1207 16:18:48.062104 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/862acceb-bcf0-4ef4-9a53-df61d6084806-operator-scripts\") pod \"cinder-e8e1-account-create-update-frpnc\" (UID: \"862acceb-bcf0-4ef4-9a53-df61d6084806\") " pod="openstack/cinder-e8e1-account-create-update-frpnc" Dec 07 16:18:48 crc kubenswrapper[4716]: I1207 16:18:48.062127 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/50a30fe7-3fb4-4b9d-b186-139c2b48295b-operator-scripts\") pod \"cinder-db-create-t7n4f\" (UID: \"50a30fe7-3fb4-4b9d-b186-139c2b48295b\") " pod="openstack/cinder-db-create-t7n4f" Dec 07 16:18:48 crc kubenswrapper[4716]: I1207 16:18:48.062667 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/50a30fe7-3fb4-4b9d-b186-139c2b48295b-operator-scripts\") pod \"cinder-db-create-t7n4f\" (UID: \"50a30fe7-3fb4-4b9d-b186-139c2b48295b\") " pod="openstack/cinder-db-create-t7n4f" Dec 07 16:18:48 crc kubenswrapper[4716]: I1207 16:18:48.063123 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9d1c048e-719c-4406-bfae-1e5a629534ed-operator-scripts\") pod \"barbican-db-create-jct5c\" (UID: \"9d1c048e-719c-4406-bfae-1e5a629534ed\") " pod="openstack/barbican-db-create-jct5c" Dec 07 16:18:48 crc kubenswrapper[4716]: I1207 16:18:48.063760 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/34a0af5c-2914-4a67-8ede-f178f08d473f-operator-scripts\") pod \"barbican-9210-account-create-update-lvz68\" (UID: \"34a0af5c-2914-4a67-8ede-f178f08d473f\") " pod="openstack/barbican-9210-account-create-update-lvz68" Dec 07 16:18:48 crc kubenswrapper[4716]: I1207 16:18:48.085324 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8xtlm\" (UniqueName: \"kubernetes.io/projected/50a30fe7-3fb4-4b9d-b186-139c2b48295b-kube-api-access-8xtlm\") pod \"cinder-db-create-t7n4f\" (UID: \"50a30fe7-3fb4-4b9d-b186-139c2b48295b\") " pod="openstack/cinder-db-create-t7n4f" Dec 07 16:18:48 crc kubenswrapper[4716]: I1207 16:18:48.085759 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4mzpj\" (UniqueName: \"kubernetes.io/projected/9d1c048e-719c-4406-bfae-1e5a629534ed-kube-api-access-4mzpj\") pod \"barbican-db-create-jct5c\" (UID: \"9d1c048e-719c-4406-bfae-1e5a629534ed\") " pod="openstack/barbican-db-create-jct5c" Dec 07 16:18:48 crc kubenswrapper[4716]: I1207 16:18:48.095536 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5jjbx\" (UniqueName: \"kubernetes.io/projected/34a0af5c-2914-4a67-8ede-f178f08d473f-kube-api-access-5jjbx\") pod \"barbican-9210-account-create-update-lvz68\" (UID: \"34a0af5c-2914-4a67-8ede-f178f08d473f\") " pod="openstack/barbican-9210-account-create-update-lvz68" Dec 07 16:18:48 crc kubenswrapper[4716]: I1207 16:18:48.104192 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-t7n4f" Dec 07 16:18:48 crc kubenswrapper[4716]: I1207 16:18:48.142319 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-jh7l6"] Dec 07 16:18:48 crc kubenswrapper[4716]: I1207 16:18:48.143293 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-jh7l6" Dec 07 16:18:48 crc kubenswrapper[4716]: I1207 16:18:48.163908 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/862acceb-bcf0-4ef4-9a53-df61d6084806-operator-scripts\") pod \"cinder-e8e1-account-create-update-frpnc\" (UID: \"862acceb-bcf0-4ef4-9a53-df61d6084806\") " pod="openstack/cinder-e8e1-account-create-update-frpnc" Dec 07 16:18:48 crc kubenswrapper[4716]: I1207 16:18:48.164275 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbfnz\" (UniqueName: \"kubernetes.io/projected/862acceb-bcf0-4ef4-9a53-df61d6084806-kube-api-access-rbfnz\") pod \"cinder-e8e1-account-create-update-frpnc\" (UID: \"862acceb-bcf0-4ef4-9a53-df61d6084806\") " pod="openstack/cinder-e8e1-account-create-update-frpnc" Dec 07 16:18:48 crc kubenswrapper[4716]: I1207 16:18:48.165181 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/862acceb-bcf0-4ef4-9a53-df61d6084806-operator-scripts\") pod \"cinder-e8e1-account-create-update-frpnc\" (UID: \"862acceb-bcf0-4ef4-9a53-df61d6084806\") " pod="openstack/cinder-e8e1-account-create-update-frpnc" Dec 07 16:18:48 crc kubenswrapper[4716]: I1207 16:18:48.189040 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-jct5c" Dec 07 16:18:48 crc kubenswrapper[4716]: I1207 16:18:48.196209 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-jh7l6"] Dec 07 16:18:48 crc kubenswrapper[4716]: I1207 16:18:48.197887 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-9210-account-create-update-lvz68" Dec 07 16:18:48 crc kubenswrapper[4716]: I1207 16:18:48.204220 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbfnz\" (UniqueName: \"kubernetes.io/projected/862acceb-bcf0-4ef4-9a53-df61d6084806-kube-api-access-rbfnz\") pod \"cinder-e8e1-account-create-update-frpnc\" (UID: \"862acceb-bcf0-4ef4-9a53-df61d6084806\") " pod="openstack/cinder-e8e1-account-create-update-frpnc" Dec 07 16:18:48 crc kubenswrapper[4716]: I1207 16:18:48.221925 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-9b37-account-create-update-rz5hs"] Dec 07 16:18:48 crc kubenswrapper[4716]: I1207 16:18:48.222979 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-9b37-account-create-update-rz5hs" Dec 07 16:18:48 crc kubenswrapper[4716]: I1207 16:18:48.230741 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Dec 07 16:18:48 crc kubenswrapper[4716]: I1207 16:18:48.250700 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-9b37-account-create-update-rz5hs"] Dec 07 16:18:48 crc kubenswrapper[4716]: I1207 16:18:48.267134 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0b362916-6a42-4fae-88be-f364dc650240-operator-scripts\") pod \"neutron-db-create-jh7l6\" (UID: \"0b362916-6a42-4fae-88be-f364dc650240\") " pod="openstack/neutron-db-create-jh7l6" Dec 07 16:18:48 crc kubenswrapper[4716]: I1207 16:18:48.267197 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mjgh9\" (UniqueName: \"kubernetes.io/projected/0b362916-6a42-4fae-88be-f364dc650240-kube-api-access-mjgh9\") pod \"neutron-db-create-jh7l6\" (UID: \"0b362916-6a42-4fae-88be-f364dc650240\") " pod="openstack/neutron-db-create-jh7l6" Dec 07 16:18:48 crc kubenswrapper[4716]: I1207 16:18:48.282682 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-e8e1-account-create-update-frpnc" Dec 07 16:18:48 crc kubenswrapper[4716]: I1207 16:18:48.326177 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-rxd6q"] Dec 07 16:18:48 crc kubenswrapper[4716]: I1207 16:18:48.327963 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-rxd6q" Dec 07 16:18:48 crc kubenswrapper[4716]: I1207 16:18:48.329315 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 07 16:18:48 crc kubenswrapper[4716]: I1207 16:18:48.330229 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 07 16:18:48 crc kubenswrapper[4716]: I1207 16:18:48.330517 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 07 16:18:48 crc kubenswrapper[4716]: I1207 16:18:48.330926 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-lhq7l" Dec 07 16:18:48 crc kubenswrapper[4716]: I1207 16:18:48.361019 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-rxd6q"] Dec 07 16:18:48 crc kubenswrapper[4716]: I1207 16:18:48.368792 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0b362916-6a42-4fae-88be-f364dc650240-operator-scripts\") pod \"neutron-db-create-jh7l6\" (UID: \"0b362916-6a42-4fae-88be-f364dc650240\") " pod="openstack/neutron-db-create-jh7l6" Dec 07 16:18:48 crc kubenswrapper[4716]: I1207 16:18:48.368866 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mjgh9\" (UniqueName: \"kubernetes.io/projected/0b362916-6a42-4fae-88be-f364dc650240-kube-api-access-mjgh9\") pod \"neutron-db-create-jh7l6\" (UID: \"0b362916-6a42-4fae-88be-f364dc650240\") " pod="openstack/neutron-db-create-jh7l6" Dec 07 16:18:48 crc kubenswrapper[4716]: I1207 16:18:48.368913 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qjjg\" (UniqueName: \"kubernetes.io/projected/bd43dd05-4a94-4e12-9e2a-788135a9ee76-kube-api-access-9qjjg\") pod \"neutron-9b37-account-create-update-rz5hs\" (UID: \"bd43dd05-4a94-4e12-9e2a-788135a9ee76\") " pod="openstack/neutron-9b37-account-create-update-rz5hs" Dec 07 16:18:48 crc kubenswrapper[4716]: I1207 16:18:48.369022 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bd43dd05-4a94-4e12-9e2a-788135a9ee76-operator-scripts\") pod \"neutron-9b37-account-create-update-rz5hs\" (UID: \"bd43dd05-4a94-4e12-9e2a-788135a9ee76\") " pod="openstack/neutron-9b37-account-create-update-rz5hs" Dec 07 16:18:48 crc kubenswrapper[4716]: I1207 16:18:48.369871 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0b362916-6a42-4fae-88be-f364dc650240-operator-scripts\") pod \"neutron-db-create-jh7l6\" (UID: \"0b362916-6a42-4fae-88be-f364dc650240\") " pod="openstack/neutron-db-create-jh7l6" Dec 07 16:18:48 crc kubenswrapper[4716]: I1207 16:18:48.390660 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mjgh9\" (UniqueName: \"kubernetes.io/projected/0b362916-6a42-4fae-88be-f364dc650240-kube-api-access-mjgh9\") pod \"neutron-db-create-jh7l6\" (UID: \"0b362916-6a42-4fae-88be-f364dc650240\") " pod="openstack/neutron-db-create-jh7l6" Dec 07 16:18:48 crc kubenswrapper[4716]: I1207 16:18:48.470814 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qjjg\" (UniqueName: \"kubernetes.io/projected/bd43dd05-4a94-4e12-9e2a-788135a9ee76-kube-api-access-9qjjg\") pod \"neutron-9b37-account-create-update-rz5hs\" (UID: \"bd43dd05-4a94-4e12-9e2a-788135a9ee76\") " pod="openstack/neutron-9b37-account-create-update-rz5hs" Dec 07 16:18:48 crc kubenswrapper[4716]: I1207 16:18:48.470878 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f94a6cb-d8ed-47f6-a90a-fa705985ecaf-combined-ca-bundle\") pod \"keystone-db-sync-rxd6q\" (UID: \"6f94a6cb-d8ed-47f6-a90a-fa705985ecaf\") " pod="openstack/keystone-db-sync-rxd6q" Dec 07 16:18:48 crc kubenswrapper[4716]: I1207 16:18:48.470965 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmm2z\" (UniqueName: \"kubernetes.io/projected/6f94a6cb-d8ed-47f6-a90a-fa705985ecaf-kube-api-access-mmm2z\") pod \"keystone-db-sync-rxd6q\" (UID: \"6f94a6cb-d8ed-47f6-a90a-fa705985ecaf\") " pod="openstack/keystone-db-sync-rxd6q" Dec 07 16:18:48 crc kubenswrapper[4716]: I1207 16:18:48.471025 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bd43dd05-4a94-4e12-9e2a-788135a9ee76-operator-scripts\") pod \"neutron-9b37-account-create-update-rz5hs\" (UID: \"bd43dd05-4a94-4e12-9e2a-788135a9ee76\") " pod="openstack/neutron-9b37-account-create-update-rz5hs" Dec 07 16:18:48 crc kubenswrapper[4716]: I1207 16:18:48.471070 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f94a6cb-d8ed-47f6-a90a-fa705985ecaf-config-data\") pod \"keystone-db-sync-rxd6q\" (UID: \"6f94a6cb-d8ed-47f6-a90a-fa705985ecaf\") " pod="openstack/keystone-db-sync-rxd6q" Dec 07 16:18:48 crc kubenswrapper[4716]: I1207 16:18:48.472221 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bd43dd05-4a94-4e12-9e2a-788135a9ee76-operator-scripts\") pod \"neutron-9b37-account-create-update-rz5hs\" (UID: \"bd43dd05-4a94-4e12-9e2a-788135a9ee76\") " pod="openstack/neutron-9b37-account-create-update-rz5hs" Dec 07 16:18:48 crc kubenswrapper[4716]: I1207 16:18:48.486726 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qjjg\" (UniqueName: \"kubernetes.io/projected/bd43dd05-4a94-4e12-9e2a-788135a9ee76-kube-api-access-9qjjg\") pod \"neutron-9b37-account-create-update-rz5hs\" (UID: \"bd43dd05-4a94-4e12-9e2a-788135a9ee76\") " pod="openstack/neutron-9b37-account-create-update-rz5hs" Dec 07 16:18:48 crc kubenswrapper[4716]: I1207 16:18:48.572864 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f94a6cb-d8ed-47f6-a90a-fa705985ecaf-config-data\") pod \"keystone-db-sync-rxd6q\" (UID: \"6f94a6cb-d8ed-47f6-a90a-fa705985ecaf\") " pod="openstack/keystone-db-sync-rxd6q" Dec 07 16:18:48 crc kubenswrapper[4716]: I1207 16:18:48.572951 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f94a6cb-d8ed-47f6-a90a-fa705985ecaf-combined-ca-bundle\") pod \"keystone-db-sync-rxd6q\" (UID: \"6f94a6cb-d8ed-47f6-a90a-fa705985ecaf\") " pod="openstack/keystone-db-sync-rxd6q" Dec 07 16:18:48 crc kubenswrapper[4716]: I1207 16:18:48.573006 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mmm2z\" (UniqueName: \"kubernetes.io/projected/6f94a6cb-d8ed-47f6-a90a-fa705985ecaf-kube-api-access-mmm2z\") pod \"keystone-db-sync-rxd6q\" (UID: \"6f94a6cb-d8ed-47f6-a90a-fa705985ecaf\") " pod="openstack/keystone-db-sync-rxd6q" Dec 07 16:18:48 crc kubenswrapper[4716]: I1207 16:18:48.573199 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-jh7l6" Dec 07 16:18:48 crc kubenswrapper[4716]: I1207 16:18:48.578941 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f94a6cb-d8ed-47f6-a90a-fa705985ecaf-combined-ca-bundle\") pod \"keystone-db-sync-rxd6q\" (UID: \"6f94a6cb-d8ed-47f6-a90a-fa705985ecaf\") " pod="openstack/keystone-db-sync-rxd6q" Dec 07 16:18:48 crc kubenswrapper[4716]: I1207 16:18:48.579009 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f94a6cb-d8ed-47f6-a90a-fa705985ecaf-config-data\") pod \"keystone-db-sync-rxd6q\" (UID: \"6f94a6cb-d8ed-47f6-a90a-fa705985ecaf\") " pod="openstack/keystone-db-sync-rxd6q" Dec 07 16:18:48 crc kubenswrapper[4716]: I1207 16:18:48.592103 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmm2z\" (UniqueName: \"kubernetes.io/projected/6f94a6cb-d8ed-47f6-a90a-fa705985ecaf-kube-api-access-mmm2z\") pod \"keystone-db-sync-rxd6q\" (UID: \"6f94a6cb-d8ed-47f6-a90a-fa705985ecaf\") " pod="openstack/keystone-db-sync-rxd6q" Dec 07 16:18:48 crc kubenswrapper[4716]: I1207 16:18:48.592499 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-9b37-account-create-update-rz5hs" Dec 07 16:18:48 crc kubenswrapper[4716]: I1207 16:18:48.663069 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-rxd6q" Dec 07 16:18:49 crc kubenswrapper[4716]: I1207 16:18:49.567588 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-t7n4f"] Dec 07 16:18:49 crc kubenswrapper[4716]: I1207 16:18:49.595469 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-9210-account-create-update-lvz68"] Dec 07 16:18:49 crc kubenswrapper[4716]: I1207 16:18:49.765432 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-jct5c"] Dec 07 16:18:49 crc kubenswrapper[4716]: I1207 16:18:49.781036 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-e8e1-account-create-update-frpnc"] Dec 07 16:18:49 crc kubenswrapper[4716]: W1207 16:18:49.907923 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod862acceb_bcf0_4ef4_9a53_df61d6084806.slice/crio-b8834f0627c8880af5e95e4677a1ca35f910398a815a94beedecdba741589aa7 WatchSource:0}: Error finding container b8834f0627c8880af5e95e4677a1ca35f910398a815a94beedecdba741589aa7: Status 404 returned error can't find the container with id b8834f0627c8880af5e95e4677a1ca35f910398a815a94beedecdba741589aa7 Dec 07 16:18:49 crc kubenswrapper[4716]: W1207 16:18:49.919389 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod34a0af5c_2914_4a67_8ede_f178f08d473f.slice/crio-1ec27f5aae2c05c0c0fa732b16b5229d7c0f78e0394d2eacc97e40ab1e56b0d7 WatchSource:0}: Error finding container 1ec27f5aae2c05c0c0fa732b16b5229d7c0f78e0394d2eacc97e40ab1e56b0d7: Status 404 returned error can't find the container with id 1ec27f5aae2c05c0c0fa732b16b5229d7c0f78e0394d2eacc97e40ab1e56b0d7 Dec 07 16:18:49 crc kubenswrapper[4716]: I1207 16:18:49.953415 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-jct5c" event={"ID":"9d1c048e-719c-4406-bfae-1e5a629534ed","Type":"ContainerStarted","Data":"9158e6ebac7561b53ecbebdecad67d30189c62a2aa5813acf13a387662620b14"} Dec 07 16:18:49 crc kubenswrapper[4716]: I1207 16:18:49.965179 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-t7n4f" event={"ID":"50a30fe7-3fb4-4b9d-b186-139c2b48295b","Type":"ContainerStarted","Data":"e68d569a5ad2aaa108a0ec6ee7042287e0c0930a477ba3aecefbcb576e069592"} Dec 07 16:18:50 crc kubenswrapper[4716]: I1207 16:18:50.214737 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-jh7l6"] Dec 07 16:18:50 crc kubenswrapper[4716]: I1207 16:18:50.230094 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-9b37-account-create-update-rz5hs"] Dec 07 16:18:50 crc kubenswrapper[4716]: W1207 16:18:50.235220 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbd43dd05_4a94_4e12_9e2a_788135a9ee76.slice/crio-f58607a1d161754aae7dd5bcc6635f4adfb6fb7263c4f9e5f89230b0baf4e5e1 WatchSource:0}: Error finding container f58607a1d161754aae7dd5bcc6635f4adfb6fb7263c4f9e5f89230b0baf4e5e1: Status 404 returned error can't find the container with id f58607a1d161754aae7dd5bcc6635f4adfb6fb7263c4f9e5f89230b0baf4e5e1 Dec 07 16:18:50 crc kubenswrapper[4716]: I1207 16:18:50.344095 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-rxd6q"] Dec 07 16:18:50 crc kubenswrapper[4716]: W1207 16:18:50.345646 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6f94a6cb_d8ed_47f6_a90a_fa705985ecaf.slice/crio-3e44d674e9fbf27b2d4b7aa1903262a119682971afda860fb232ca214d6ba5de WatchSource:0}: Error finding container 3e44d674e9fbf27b2d4b7aa1903262a119682971afda860fb232ca214d6ba5de: Status 404 returned error can't find the container with id 3e44d674e9fbf27b2d4b7aa1903262a119682971afda860fb232ca214d6ba5de Dec 07 16:18:50 crc kubenswrapper[4716]: I1207 16:18:50.480086 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-74f6bcbc87-2cv95" Dec 07 16:18:50 crc kubenswrapper[4716]: I1207 16:18:50.533858 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-nj89r"] Dec 07 16:18:50 crc kubenswrapper[4716]: I1207 16:18:50.534153 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-698758b865-nj89r" podUID="666f9fee-767e-40d6-ac44-5abe506e3266" containerName="dnsmasq-dns" containerID="cri-o://ebe51569d6f1b0c4d1019e4f87d89008f07452249943e222bf04cf1af5375d29" gracePeriod=10 Dec 07 16:18:50 crc kubenswrapper[4716]: I1207 16:18:50.978769 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-e8e1-account-create-update-frpnc" event={"ID":"862acceb-bcf0-4ef4-9a53-df61d6084806","Type":"ContainerStarted","Data":"b8834f0627c8880af5e95e4677a1ca35f910398a815a94beedecdba741589aa7"} Dec 07 16:18:50 crc kubenswrapper[4716]: I1207 16:18:50.979936 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-jh7l6" event={"ID":"0b362916-6a42-4fae-88be-f364dc650240","Type":"ContainerStarted","Data":"f4c661b4078d957555ac63efbe36443861255872076f418d7649a96c3e0c04f5"} Dec 07 16:18:50 crc kubenswrapper[4716]: I1207 16:18:50.981428 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-9b37-account-create-update-rz5hs" event={"ID":"bd43dd05-4a94-4e12-9e2a-788135a9ee76","Type":"ContainerStarted","Data":"f58607a1d161754aae7dd5bcc6635f4adfb6fb7263c4f9e5f89230b0baf4e5e1"} Dec 07 16:18:50 crc kubenswrapper[4716]: I1207 16:18:50.982511 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-9210-account-create-update-lvz68" event={"ID":"34a0af5c-2914-4a67-8ede-f178f08d473f","Type":"ContainerStarted","Data":"1ec27f5aae2c05c0c0fa732b16b5229d7c0f78e0394d2eacc97e40ab1e56b0d7"} Dec 07 16:18:50 crc kubenswrapper[4716]: I1207 16:18:50.984109 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-rxd6q" event={"ID":"6f94a6cb-d8ed-47f6-a90a-fa705985ecaf","Type":"ContainerStarted","Data":"3e44d674e9fbf27b2d4b7aa1903262a119682971afda860fb232ca214d6ba5de"} Dec 07 16:18:51 crc kubenswrapper[4716]: I1207 16:18:51.995145 4716 generic.go:334] "Generic (PLEG): container finished" podID="50a30fe7-3fb4-4b9d-b186-139c2b48295b" containerID="9451289360f0da6e6a1c2c0b7496a3f85c00d5f1d96494d9b9b614d5cd4bd237" exitCode=0 Dec 07 16:18:51 crc kubenswrapper[4716]: I1207 16:18:51.995253 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-t7n4f" event={"ID":"50a30fe7-3fb4-4b9d-b186-139c2b48295b","Type":"ContainerDied","Data":"9451289360f0da6e6a1c2c0b7496a3f85c00d5f1d96494d9b9b614d5cd4bd237"} Dec 07 16:18:51 crc kubenswrapper[4716]: I1207 16:18:51.998346 4716 generic.go:334] "Generic (PLEG): container finished" podID="9d1c048e-719c-4406-bfae-1e5a629534ed" containerID="17f5dc7f54db07da92aa7837da9cf1b03ba4e247d65da80dd553c5fe36814a0f" exitCode=0 Dec 07 16:18:51 crc kubenswrapper[4716]: I1207 16:18:51.998459 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-jct5c" event={"ID":"9d1c048e-719c-4406-bfae-1e5a629534ed","Type":"ContainerDied","Data":"17f5dc7f54db07da92aa7837da9cf1b03ba4e247d65da80dd553c5fe36814a0f"} Dec 07 16:18:52 crc kubenswrapper[4716]: I1207 16:18:52.003322 4716 generic.go:334] "Generic (PLEG): container finished" podID="666f9fee-767e-40d6-ac44-5abe506e3266" containerID="ebe51569d6f1b0c4d1019e4f87d89008f07452249943e222bf04cf1af5375d29" exitCode=0 Dec 07 16:18:52 crc kubenswrapper[4716]: I1207 16:18:52.003411 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-nj89r" event={"ID":"666f9fee-767e-40d6-ac44-5abe506e3266","Type":"ContainerDied","Data":"ebe51569d6f1b0c4d1019e4f87d89008f07452249943e222bf04cf1af5375d29"} Dec 07 16:18:52 crc kubenswrapper[4716]: I1207 16:18:52.005872 4716 generic.go:334] "Generic (PLEG): container finished" podID="862acceb-bcf0-4ef4-9a53-df61d6084806" containerID="ac3d925aa12a9311bcb204061c9de1fc84a0224612badc8a11ad34df2ad5a754" exitCode=0 Dec 07 16:18:52 crc kubenswrapper[4716]: I1207 16:18:52.005943 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-e8e1-account-create-update-frpnc" event={"ID":"862acceb-bcf0-4ef4-9a53-df61d6084806","Type":"ContainerDied","Data":"ac3d925aa12a9311bcb204061c9de1fc84a0224612badc8a11ad34df2ad5a754"} Dec 07 16:18:52 crc kubenswrapper[4716]: I1207 16:18:52.016684 4716 generic.go:334] "Generic (PLEG): container finished" podID="0b362916-6a42-4fae-88be-f364dc650240" containerID="75ce43476b8844c678d48c3ccb9c33485489d20a32ba7acddb269a24adfa5af2" exitCode=0 Dec 07 16:18:52 crc kubenswrapper[4716]: I1207 16:18:52.016826 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-jh7l6" event={"ID":"0b362916-6a42-4fae-88be-f364dc650240","Type":"ContainerDied","Data":"75ce43476b8844c678d48c3ccb9c33485489d20a32ba7acddb269a24adfa5af2"} Dec 07 16:18:52 crc kubenswrapper[4716]: I1207 16:18:52.019354 4716 generic.go:334] "Generic (PLEG): container finished" podID="bd43dd05-4a94-4e12-9e2a-788135a9ee76" containerID="b52f6aad7da0f69d63bf00028fce594bb28e1dc0d3c7d743727abbfd62dff3b7" exitCode=0 Dec 07 16:18:52 crc kubenswrapper[4716]: I1207 16:18:52.019396 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-9b37-account-create-update-rz5hs" event={"ID":"bd43dd05-4a94-4e12-9e2a-788135a9ee76","Type":"ContainerDied","Data":"b52f6aad7da0f69d63bf00028fce594bb28e1dc0d3c7d743727abbfd62dff3b7"} Dec 07 16:18:52 crc kubenswrapper[4716]: I1207 16:18:52.027162 4716 generic.go:334] "Generic (PLEG): container finished" podID="34a0af5c-2914-4a67-8ede-f178f08d473f" containerID="2445d34fc45b616bb4cd2f04302cbcf69cc53f93e76a6c8ef579ce0b0b858aca" exitCode=0 Dec 07 16:18:52 crc kubenswrapper[4716]: I1207 16:18:52.027167 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-9210-account-create-update-lvz68" event={"ID":"34a0af5c-2914-4a67-8ede-f178f08d473f","Type":"ContainerDied","Data":"2445d34fc45b616bb4cd2f04302cbcf69cc53f93e76a6c8ef579ce0b0b858aca"} Dec 07 16:18:52 crc kubenswrapper[4716]: I1207 16:18:52.232766 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-nj89r" Dec 07 16:18:52 crc kubenswrapper[4716]: I1207 16:18:52.326422 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/666f9fee-767e-40d6-ac44-5abe506e3266-config\") pod \"666f9fee-767e-40d6-ac44-5abe506e3266\" (UID: \"666f9fee-767e-40d6-ac44-5abe506e3266\") " Dec 07 16:18:52 crc kubenswrapper[4716]: I1207 16:18:52.326465 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/666f9fee-767e-40d6-ac44-5abe506e3266-ovsdbserver-sb\") pod \"666f9fee-767e-40d6-ac44-5abe506e3266\" (UID: \"666f9fee-767e-40d6-ac44-5abe506e3266\") " Dec 07 16:18:52 crc kubenswrapper[4716]: I1207 16:18:52.326589 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/666f9fee-767e-40d6-ac44-5abe506e3266-ovsdbserver-nb\") pod \"666f9fee-767e-40d6-ac44-5abe506e3266\" (UID: \"666f9fee-767e-40d6-ac44-5abe506e3266\") " Dec 07 16:18:52 crc kubenswrapper[4716]: I1207 16:18:52.326682 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/666f9fee-767e-40d6-ac44-5abe506e3266-dns-svc\") pod \"666f9fee-767e-40d6-ac44-5abe506e3266\" (UID: \"666f9fee-767e-40d6-ac44-5abe506e3266\") " Dec 07 16:18:52 crc kubenswrapper[4716]: I1207 16:18:52.326723 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8wkjm\" (UniqueName: \"kubernetes.io/projected/666f9fee-767e-40d6-ac44-5abe506e3266-kube-api-access-8wkjm\") pod \"666f9fee-767e-40d6-ac44-5abe506e3266\" (UID: \"666f9fee-767e-40d6-ac44-5abe506e3266\") " Dec 07 16:18:52 crc kubenswrapper[4716]: I1207 16:18:52.333188 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/666f9fee-767e-40d6-ac44-5abe506e3266-kube-api-access-8wkjm" (OuterVolumeSpecName: "kube-api-access-8wkjm") pod "666f9fee-767e-40d6-ac44-5abe506e3266" (UID: "666f9fee-767e-40d6-ac44-5abe506e3266"). InnerVolumeSpecName "kube-api-access-8wkjm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:18:52 crc kubenswrapper[4716]: I1207 16:18:52.371264 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/666f9fee-767e-40d6-ac44-5abe506e3266-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "666f9fee-767e-40d6-ac44-5abe506e3266" (UID: "666f9fee-767e-40d6-ac44-5abe506e3266"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:18:52 crc kubenswrapper[4716]: I1207 16:18:52.371726 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/666f9fee-767e-40d6-ac44-5abe506e3266-config" (OuterVolumeSpecName: "config") pod "666f9fee-767e-40d6-ac44-5abe506e3266" (UID: "666f9fee-767e-40d6-ac44-5abe506e3266"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:18:52 crc kubenswrapper[4716]: I1207 16:18:52.373858 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/666f9fee-767e-40d6-ac44-5abe506e3266-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "666f9fee-767e-40d6-ac44-5abe506e3266" (UID: "666f9fee-767e-40d6-ac44-5abe506e3266"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:18:52 crc kubenswrapper[4716]: I1207 16:18:52.375353 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/666f9fee-767e-40d6-ac44-5abe506e3266-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "666f9fee-767e-40d6-ac44-5abe506e3266" (UID: "666f9fee-767e-40d6-ac44-5abe506e3266"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:18:52 crc kubenswrapper[4716]: I1207 16:18:52.428062 4716 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/666f9fee-767e-40d6-ac44-5abe506e3266-config\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:52 crc kubenswrapper[4716]: I1207 16:18:52.428112 4716 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/666f9fee-767e-40d6-ac44-5abe506e3266-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:52 crc kubenswrapper[4716]: I1207 16:18:52.428125 4716 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/666f9fee-767e-40d6-ac44-5abe506e3266-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:52 crc kubenswrapper[4716]: I1207 16:18:52.428134 4716 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/666f9fee-767e-40d6-ac44-5abe506e3266-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:52 crc kubenswrapper[4716]: I1207 16:18:52.428144 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8wkjm\" (UniqueName: \"kubernetes.io/projected/666f9fee-767e-40d6-ac44-5abe506e3266-kube-api-access-8wkjm\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:52 crc kubenswrapper[4716]: I1207 16:18:52.760901 4716 patch_prober.go:28] interesting pod/machine-config-daemon-zcxxp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 16:18:52 crc kubenswrapper[4716]: I1207 16:18:52.760959 4716 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 16:18:53 crc kubenswrapper[4716]: I1207 16:18:53.037303 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-nj89r" event={"ID":"666f9fee-767e-40d6-ac44-5abe506e3266","Type":"ContainerDied","Data":"79480ac09df21f06646f2b716c79bb5b771610a22a6eb43478fbb6cb63124dad"} Dec 07 16:18:53 crc kubenswrapper[4716]: I1207 16:18:53.037354 4716 scope.go:117] "RemoveContainer" containerID="ebe51569d6f1b0c4d1019e4f87d89008f07452249943e222bf04cf1af5375d29" Dec 07 16:18:53 crc kubenswrapper[4716]: I1207 16:18:53.037418 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-nj89r" Dec 07 16:18:53 crc kubenswrapper[4716]: I1207 16:18:53.073501 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-nj89r"] Dec 07 16:18:53 crc kubenswrapper[4716]: I1207 16:18:53.079355 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-698758b865-nj89r"] Dec 07 16:18:53 crc kubenswrapper[4716]: I1207 16:18:53.670610 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="666f9fee-767e-40d6-ac44-5abe506e3266" path="/var/lib/kubelet/pods/666f9fee-767e-40d6-ac44-5abe506e3266/volumes" Dec 07 16:18:55 crc kubenswrapper[4716]: I1207 16:18:55.358127 4716 scope.go:117] "RemoveContainer" containerID="f5ea6b68fe720e1798d5e831a08cd1ed5dbe813349bb3eb0aad5c6346dd997cf" Dec 07 16:18:55 crc kubenswrapper[4716]: I1207 16:18:55.553415 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-jh7l6" Dec 07 16:18:55 crc kubenswrapper[4716]: I1207 16:18:55.574110 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-t7n4f" Dec 07 16:18:55 crc kubenswrapper[4716]: I1207 16:18:55.587032 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0b362916-6a42-4fae-88be-f364dc650240-operator-scripts\") pod \"0b362916-6a42-4fae-88be-f364dc650240\" (UID: \"0b362916-6a42-4fae-88be-f364dc650240\") " Dec 07 16:18:55 crc kubenswrapper[4716]: I1207 16:18:55.587138 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mjgh9\" (UniqueName: \"kubernetes.io/projected/0b362916-6a42-4fae-88be-f364dc650240-kube-api-access-mjgh9\") pod \"0b362916-6a42-4fae-88be-f364dc650240\" (UID: \"0b362916-6a42-4fae-88be-f364dc650240\") " Dec 07 16:18:55 crc kubenswrapper[4716]: I1207 16:18:55.589598 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b362916-6a42-4fae-88be-f364dc650240-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0b362916-6a42-4fae-88be-f364dc650240" (UID: "0b362916-6a42-4fae-88be-f364dc650240"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:18:55 crc kubenswrapper[4716]: I1207 16:18:55.602203 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b362916-6a42-4fae-88be-f364dc650240-kube-api-access-mjgh9" (OuterVolumeSpecName: "kube-api-access-mjgh9") pod "0b362916-6a42-4fae-88be-f364dc650240" (UID: "0b362916-6a42-4fae-88be-f364dc650240"). InnerVolumeSpecName "kube-api-access-mjgh9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:18:55 crc kubenswrapper[4716]: I1207 16:18:55.605413 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-9210-account-create-update-lvz68" Dec 07 16:18:55 crc kubenswrapper[4716]: I1207 16:18:55.626492 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-jct5c" Dec 07 16:18:55 crc kubenswrapper[4716]: I1207 16:18:55.627017 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-9b37-account-create-update-rz5hs" Dec 07 16:18:55 crc kubenswrapper[4716]: I1207 16:18:55.636135 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-e8e1-account-create-update-frpnc" Dec 07 16:18:55 crc kubenswrapper[4716]: I1207 16:18:55.696850 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5jjbx\" (UniqueName: \"kubernetes.io/projected/34a0af5c-2914-4a67-8ede-f178f08d473f-kube-api-access-5jjbx\") pod \"34a0af5c-2914-4a67-8ede-f178f08d473f\" (UID: \"34a0af5c-2914-4a67-8ede-f178f08d473f\") " Dec 07 16:18:55 crc kubenswrapper[4716]: I1207 16:18:55.696905 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/50a30fe7-3fb4-4b9d-b186-139c2b48295b-operator-scripts\") pod \"50a30fe7-3fb4-4b9d-b186-139c2b48295b\" (UID: \"50a30fe7-3fb4-4b9d-b186-139c2b48295b\") " Dec 07 16:18:55 crc kubenswrapper[4716]: I1207 16:18:55.696961 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9d1c048e-719c-4406-bfae-1e5a629534ed-operator-scripts\") pod \"9d1c048e-719c-4406-bfae-1e5a629534ed\" (UID: \"9d1c048e-719c-4406-bfae-1e5a629534ed\") " Dec 07 16:18:55 crc kubenswrapper[4716]: I1207 16:18:55.697005 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9qjjg\" (UniqueName: \"kubernetes.io/projected/bd43dd05-4a94-4e12-9e2a-788135a9ee76-kube-api-access-9qjjg\") pod \"bd43dd05-4a94-4e12-9e2a-788135a9ee76\" (UID: \"bd43dd05-4a94-4e12-9e2a-788135a9ee76\") " Dec 07 16:18:55 crc kubenswrapper[4716]: I1207 16:18:55.697063 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4mzpj\" (UniqueName: \"kubernetes.io/projected/9d1c048e-719c-4406-bfae-1e5a629534ed-kube-api-access-4mzpj\") pod \"9d1c048e-719c-4406-bfae-1e5a629534ed\" (UID: \"9d1c048e-719c-4406-bfae-1e5a629534ed\") " Dec 07 16:18:55 crc kubenswrapper[4716]: I1207 16:18:55.697175 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rbfnz\" (UniqueName: \"kubernetes.io/projected/862acceb-bcf0-4ef4-9a53-df61d6084806-kube-api-access-rbfnz\") pod \"862acceb-bcf0-4ef4-9a53-df61d6084806\" (UID: \"862acceb-bcf0-4ef4-9a53-df61d6084806\") " Dec 07 16:18:55 crc kubenswrapper[4716]: I1207 16:18:55.697240 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bd43dd05-4a94-4e12-9e2a-788135a9ee76-operator-scripts\") pod \"bd43dd05-4a94-4e12-9e2a-788135a9ee76\" (UID: \"bd43dd05-4a94-4e12-9e2a-788135a9ee76\") " Dec 07 16:18:55 crc kubenswrapper[4716]: I1207 16:18:55.697285 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/34a0af5c-2914-4a67-8ede-f178f08d473f-operator-scripts\") pod \"34a0af5c-2914-4a67-8ede-f178f08d473f\" (UID: \"34a0af5c-2914-4a67-8ede-f178f08d473f\") " Dec 07 16:18:55 crc kubenswrapper[4716]: I1207 16:18:55.697365 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/862acceb-bcf0-4ef4-9a53-df61d6084806-operator-scripts\") pod \"862acceb-bcf0-4ef4-9a53-df61d6084806\" (UID: \"862acceb-bcf0-4ef4-9a53-df61d6084806\") " Dec 07 16:18:55 crc kubenswrapper[4716]: I1207 16:18:55.697435 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8xtlm\" (UniqueName: \"kubernetes.io/projected/50a30fe7-3fb4-4b9d-b186-139c2b48295b-kube-api-access-8xtlm\") pod \"50a30fe7-3fb4-4b9d-b186-139c2b48295b\" (UID: \"50a30fe7-3fb4-4b9d-b186-139c2b48295b\") " Dec 07 16:18:55 crc kubenswrapper[4716]: I1207 16:18:55.698624 4716 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0b362916-6a42-4fae-88be-f364dc650240-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:55 crc kubenswrapper[4716]: I1207 16:18:55.698653 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mjgh9\" (UniqueName: \"kubernetes.io/projected/0b362916-6a42-4fae-88be-f364dc650240-kube-api-access-mjgh9\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:55 crc kubenswrapper[4716]: I1207 16:18:55.701848 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd43dd05-4a94-4e12-9e2a-788135a9ee76-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "bd43dd05-4a94-4e12-9e2a-788135a9ee76" (UID: "bd43dd05-4a94-4e12-9e2a-788135a9ee76"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:18:55 crc kubenswrapper[4716]: I1207 16:18:55.705312 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34a0af5c-2914-4a67-8ede-f178f08d473f-kube-api-access-5jjbx" (OuterVolumeSpecName: "kube-api-access-5jjbx") pod "34a0af5c-2914-4a67-8ede-f178f08d473f" (UID: "34a0af5c-2914-4a67-8ede-f178f08d473f"). InnerVolumeSpecName "kube-api-access-5jjbx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:18:55 crc kubenswrapper[4716]: I1207 16:18:55.706248 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50a30fe7-3fb4-4b9d-b186-139c2b48295b-kube-api-access-8xtlm" (OuterVolumeSpecName: "kube-api-access-8xtlm") pod "50a30fe7-3fb4-4b9d-b186-139c2b48295b" (UID: "50a30fe7-3fb4-4b9d-b186-139c2b48295b"). InnerVolumeSpecName "kube-api-access-8xtlm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:18:55 crc kubenswrapper[4716]: I1207 16:18:55.706527 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d1c048e-719c-4406-bfae-1e5a629534ed-kube-api-access-4mzpj" (OuterVolumeSpecName: "kube-api-access-4mzpj") pod "9d1c048e-719c-4406-bfae-1e5a629534ed" (UID: "9d1c048e-719c-4406-bfae-1e5a629534ed"). InnerVolumeSpecName "kube-api-access-4mzpj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:18:55 crc kubenswrapper[4716]: I1207 16:18:55.706546 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d1c048e-719c-4406-bfae-1e5a629534ed-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9d1c048e-719c-4406-bfae-1e5a629534ed" (UID: "9d1c048e-719c-4406-bfae-1e5a629534ed"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:18:55 crc kubenswrapper[4716]: I1207 16:18:55.706589 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/862acceb-bcf0-4ef4-9a53-df61d6084806-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "862acceb-bcf0-4ef4-9a53-df61d6084806" (UID: "862acceb-bcf0-4ef4-9a53-df61d6084806"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:18:55 crc kubenswrapper[4716]: I1207 16:18:55.706972 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/50a30fe7-3fb4-4b9d-b186-139c2b48295b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "50a30fe7-3fb4-4b9d-b186-139c2b48295b" (UID: "50a30fe7-3fb4-4b9d-b186-139c2b48295b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:18:55 crc kubenswrapper[4716]: I1207 16:18:55.707018 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/34a0af5c-2914-4a67-8ede-f178f08d473f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "34a0af5c-2914-4a67-8ede-f178f08d473f" (UID: "34a0af5c-2914-4a67-8ede-f178f08d473f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:18:55 crc kubenswrapper[4716]: I1207 16:18:55.707599 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/862acceb-bcf0-4ef4-9a53-df61d6084806-kube-api-access-rbfnz" (OuterVolumeSpecName: "kube-api-access-rbfnz") pod "862acceb-bcf0-4ef4-9a53-df61d6084806" (UID: "862acceb-bcf0-4ef4-9a53-df61d6084806"). InnerVolumeSpecName "kube-api-access-rbfnz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:18:55 crc kubenswrapper[4716]: I1207 16:18:55.709826 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd43dd05-4a94-4e12-9e2a-788135a9ee76-kube-api-access-9qjjg" (OuterVolumeSpecName: "kube-api-access-9qjjg") pod "bd43dd05-4a94-4e12-9e2a-788135a9ee76" (UID: "bd43dd05-4a94-4e12-9e2a-788135a9ee76"). InnerVolumeSpecName "kube-api-access-9qjjg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:18:55 crc kubenswrapper[4716]: I1207 16:18:55.800797 4716 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/34a0af5c-2914-4a67-8ede-f178f08d473f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:55 crc kubenswrapper[4716]: I1207 16:18:55.800828 4716 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/862acceb-bcf0-4ef4-9a53-df61d6084806-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:55 crc kubenswrapper[4716]: I1207 16:18:55.800838 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8xtlm\" (UniqueName: \"kubernetes.io/projected/50a30fe7-3fb4-4b9d-b186-139c2b48295b-kube-api-access-8xtlm\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:55 crc kubenswrapper[4716]: I1207 16:18:55.800848 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5jjbx\" (UniqueName: \"kubernetes.io/projected/34a0af5c-2914-4a67-8ede-f178f08d473f-kube-api-access-5jjbx\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:55 crc kubenswrapper[4716]: I1207 16:18:55.800856 4716 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/50a30fe7-3fb4-4b9d-b186-139c2b48295b-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:55 crc kubenswrapper[4716]: I1207 16:18:55.800865 4716 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9d1c048e-719c-4406-bfae-1e5a629534ed-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:55 crc kubenswrapper[4716]: I1207 16:18:55.800873 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9qjjg\" (UniqueName: \"kubernetes.io/projected/bd43dd05-4a94-4e12-9e2a-788135a9ee76-kube-api-access-9qjjg\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:55 crc kubenswrapper[4716]: I1207 16:18:55.800882 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4mzpj\" (UniqueName: \"kubernetes.io/projected/9d1c048e-719c-4406-bfae-1e5a629534ed-kube-api-access-4mzpj\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:55 crc kubenswrapper[4716]: I1207 16:18:55.800891 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rbfnz\" (UniqueName: \"kubernetes.io/projected/862acceb-bcf0-4ef4-9a53-df61d6084806-kube-api-access-rbfnz\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:55 crc kubenswrapper[4716]: I1207 16:18:55.800900 4716 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bd43dd05-4a94-4e12-9e2a-788135a9ee76-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 16:18:56 crc kubenswrapper[4716]: I1207 16:18:56.071915 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-9b37-account-create-update-rz5hs" event={"ID":"bd43dd05-4a94-4e12-9e2a-788135a9ee76","Type":"ContainerDied","Data":"f58607a1d161754aae7dd5bcc6635f4adfb6fb7263c4f9e5f89230b0baf4e5e1"} Dec 07 16:18:56 crc kubenswrapper[4716]: I1207 16:18:56.071958 4716 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f58607a1d161754aae7dd5bcc6635f4adfb6fb7263c4f9e5f89230b0baf4e5e1" Dec 07 16:18:56 crc kubenswrapper[4716]: I1207 16:18:56.072028 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-9b37-account-create-update-rz5hs" Dec 07 16:18:56 crc kubenswrapper[4716]: I1207 16:18:56.077314 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-9210-account-create-update-lvz68" event={"ID":"34a0af5c-2914-4a67-8ede-f178f08d473f","Type":"ContainerDied","Data":"1ec27f5aae2c05c0c0fa732b16b5229d7c0f78e0394d2eacc97e40ab1e56b0d7"} Dec 07 16:18:56 crc kubenswrapper[4716]: I1207 16:18:56.077367 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-9210-account-create-update-lvz68" Dec 07 16:18:56 crc kubenswrapper[4716]: I1207 16:18:56.077380 4716 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1ec27f5aae2c05c0c0fa732b16b5229d7c0f78e0394d2eacc97e40ab1e56b0d7" Dec 07 16:18:56 crc kubenswrapper[4716]: I1207 16:18:56.079506 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-rxd6q" event={"ID":"6f94a6cb-d8ed-47f6-a90a-fa705985ecaf","Type":"ContainerStarted","Data":"44c6371bdef3dc8c2a62f43268f22b3bae06501f812fba3cc8545e388fd6080a"} Dec 07 16:18:56 crc kubenswrapper[4716]: I1207 16:18:56.081124 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-jct5c" event={"ID":"9d1c048e-719c-4406-bfae-1e5a629534ed","Type":"ContainerDied","Data":"9158e6ebac7561b53ecbebdecad67d30189c62a2aa5813acf13a387662620b14"} Dec 07 16:18:56 crc kubenswrapper[4716]: I1207 16:18:56.081153 4716 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9158e6ebac7561b53ecbebdecad67d30189c62a2aa5813acf13a387662620b14" Dec 07 16:18:56 crc kubenswrapper[4716]: I1207 16:18:56.081199 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-jct5c" Dec 07 16:18:56 crc kubenswrapper[4716]: I1207 16:18:56.082809 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-t7n4f" Dec 07 16:18:56 crc kubenswrapper[4716]: I1207 16:18:56.082825 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-t7n4f" event={"ID":"50a30fe7-3fb4-4b9d-b186-139c2b48295b","Type":"ContainerDied","Data":"e68d569a5ad2aaa108a0ec6ee7042287e0c0930a477ba3aecefbcb576e069592"} Dec 07 16:18:56 crc kubenswrapper[4716]: I1207 16:18:56.082861 4716 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e68d569a5ad2aaa108a0ec6ee7042287e0c0930a477ba3aecefbcb576e069592" Dec 07 16:18:56 crc kubenswrapper[4716]: I1207 16:18:56.085572 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-e8e1-account-create-update-frpnc" event={"ID":"862acceb-bcf0-4ef4-9a53-df61d6084806","Type":"ContainerDied","Data":"b8834f0627c8880af5e95e4677a1ca35f910398a815a94beedecdba741589aa7"} Dec 07 16:18:56 crc kubenswrapper[4716]: I1207 16:18:56.085616 4716 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b8834f0627c8880af5e95e4677a1ca35f910398a815a94beedecdba741589aa7" Dec 07 16:18:56 crc kubenswrapper[4716]: I1207 16:18:56.085704 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-e8e1-account-create-update-frpnc" Dec 07 16:18:56 crc kubenswrapper[4716]: I1207 16:18:56.087832 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-jh7l6" event={"ID":"0b362916-6a42-4fae-88be-f364dc650240","Type":"ContainerDied","Data":"f4c661b4078d957555ac63efbe36443861255872076f418d7649a96c3e0c04f5"} Dec 07 16:18:56 crc kubenswrapper[4716]: I1207 16:18:56.087869 4716 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f4c661b4078d957555ac63efbe36443861255872076f418d7649a96c3e0c04f5" Dec 07 16:18:56 crc kubenswrapper[4716]: I1207 16:18:56.087942 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-jh7l6" Dec 07 16:18:56 crc kubenswrapper[4716]: I1207 16:18:56.096843 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-rxd6q" podStartSLOduration=3.023511052 podStartE2EDuration="8.096822525s" podCreationTimestamp="2025-12-07 16:18:48 +0000 UTC" firstStartedPulling="2025-12-07 16:18:50.348106897 +0000 UTC m=+993.038391809" lastFinishedPulling="2025-12-07 16:18:55.42141836 +0000 UTC m=+998.111703282" observedRunningTime="2025-12-07 16:18:56.095394156 +0000 UTC m=+998.785679058" watchObservedRunningTime="2025-12-07 16:18:56.096822525 +0000 UTC m=+998.787107437" Dec 07 16:18:59 crc kubenswrapper[4716]: I1207 16:18:59.125958 4716 generic.go:334] "Generic (PLEG): container finished" podID="6f94a6cb-d8ed-47f6-a90a-fa705985ecaf" containerID="44c6371bdef3dc8c2a62f43268f22b3bae06501f812fba3cc8545e388fd6080a" exitCode=0 Dec 07 16:18:59 crc kubenswrapper[4716]: I1207 16:18:59.126040 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-rxd6q" event={"ID":"6f94a6cb-d8ed-47f6-a90a-fa705985ecaf","Type":"ContainerDied","Data":"44c6371bdef3dc8c2a62f43268f22b3bae06501f812fba3cc8545e388fd6080a"} Dec 07 16:19:00 crc kubenswrapper[4716]: I1207 16:19:00.454293 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-rxd6q" Dec 07 16:19:00 crc kubenswrapper[4716]: I1207 16:19:00.579327 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mmm2z\" (UniqueName: \"kubernetes.io/projected/6f94a6cb-d8ed-47f6-a90a-fa705985ecaf-kube-api-access-mmm2z\") pod \"6f94a6cb-d8ed-47f6-a90a-fa705985ecaf\" (UID: \"6f94a6cb-d8ed-47f6-a90a-fa705985ecaf\") " Dec 07 16:19:00 crc kubenswrapper[4716]: I1207 16:19:00.579459 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f94a6cb-d8ed-47f6-a90a-fa705985ecaf-config-data\") pod \"6f94a6cb-d8ed-47f6-a90a-fa705985ecaf\" (UID: \"6f94a6cb-d8ed-47f6-a90a-fa705985ecaf\") " Dec 07 16:19:00 crc kubenswrapper[4716]: I1207 16:19:00.579615 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f94a6cb-d8ed-47f6-a90a-fa705985ecaf-combined-ca-bundle\") pod \"6f94a6cb-d8ed-47f6-a90a-fa705985ecaf\" (UID: \"6f94a6cb-d8ed-47f6-a90a-fa705985ecaf\") " Dec 07 16:19:00 crc kubenswrapper[4716]: I1207 16:19:00.584371 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f94a6cb-d8ed-47f6-a90a-fa705985ecaf-kube-api-access-mmm2z" (OuterVolumeSpecName: "kube-api-access-mmm2z") pod "6f94a6cb-d8ed-47f6-a90a-fa705985ecaf" (UID: "6f94a6cb-d8ed-47f6-a90a-fa705985ecaf"). InnerVolumeSpecName "kube-api-access-mmm2z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:19:00 crc kubenswrapper[4716]: I1207 16:19:00.603238 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f94a6cb-d8ed-47f6-a90a-fa705985ecaf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6f94a6cb-d8ed-47f6-a90a-fa705985ecaf" (UID: "6f94a6cb-d8ed-47f6-a90a-fa705985ecaf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:19:00 crc kubenswrapper[4716]: I1207 16:19:00.625343 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f94a6cb-d8ed-47f6-a90a-fa705985ecaf-config-data" (OuterVolumeSpecName: "config-data") pod "6f94a6cb-d8ed-47f6-a90a-fa705985ecaf" (UID: "6f94a6cb-d8ed-47f6-a90a-fa705985ecaf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:19:00 crc kubenswrapper[4716]: I1207 16:19:00.682340 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mmm2z\" (UniqueName: \"kubernetes.io/projected/6f94a6cb-d8ed-47f6-a90a-fa705985ecaf-kube-api-access-mmm2z\") on node \"crc\" DevicePath \"\"" Dec 07 16:19:00 crc kubenswrapper[4716]: I1207 16:19:00.682393 4716 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f94a6cb-d8ed-47f6-a90a-fa705985ecaf-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 16:19:00 crc kubenswrapper[4716]: I1207 16:19:00.682407 4716 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f94a6cb-d8ed-47f6-a90a-fa705985ecaf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.147496 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-rxd6q" event={"ID":"6f94a6cb-d8ed-47f6-a90a-fa705985ecaf","Type":"ContainerDied","Data":"3e44d674e9fbf27b2d4b7aa1903262a119682971afda860fb232ca214d6ba5de"} Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.147555 4716 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3e44d674e9fbf27b2d4b7aa1903262a119682971afda860fb232ca214d6ba5de" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.147634 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-rxd6q" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.430670 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-6djm5"] Dec 07 16:19:01 crc kubenswrapper[4716]: E1207 16:19:01.431018 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d1c048e-719c-4406-bfae-1e5a629534ed" containerName="mariadb-database-create" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.431033 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d1c048e-719c-4406-bfae-1e5a629534ed" containerName="mariadb-database-create" Dec 07 16:19:01 crc kubenswrapper[4716]: E1207 16:19:01.431044 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="666f9fee-767e-40d6-ac44-5abe506e3266" containerName="init" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.431050 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="666f9fee-767e-40d6-ac44-5abe506e3266" containerName="init" Dec 07 16:19:01 crc kubenswrapper[4716]: E1207 16:19:01.431056 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="666f9fee-767e-40d6-ac44-5abe506e3266" containerName="dnsmasq-dns" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.431062 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="666f9fee-767e-40d6-ac44-5abe506e3266" containerName="dnsmasq-dns" Dec 07 16:19:01 crc kubenswrapper[4716]: E1207 16:19:01.431090 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd43dd05-4a94-4e12-9e2a-788135a9ee76" containerName="mariadb-account-create-update" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.431096 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd43dd05-4a94-4e12-9e2a-788135a9ee76" containerName="mariadb-account-create-update" Dec 07 16:19:01 crc kubenswrapper[4716]: E1207 16:19:01.431117 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b362916-6a42-4fae-88be-f364dc650240" containerName="mariadb-database-create" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.431123 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b362916-6a42-4fae-88be-f364dc650240" containerName="mariadb-database-create" Dec 07 16:19:01 crc kubenswrapper[4716]: E1207 16:19:01.431132 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="862acceb-bcf0-4ef4-9a53-df61d6084806" containerName="mariadb-account-create-update" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.431137 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="862acceb-bcf0-4ef4-9a53-df61d6084806" containerName="mariadb-account-create-update" Dec 07 16:19:01 crc kubenswrapper[4716]: E1207 16:19:01.431161 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50a30fe7-3fb4-4b9d-b186-139c2b48295b" containerName="mariadb-database-create" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.431167 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="50a30fe7-3fb4-4b9d-b186-139c2b48295b" containerName="mariadb-database-create" Dec 07 16:19:01 crc kubenswrapper[4716]: E1207 16:19:01.431179 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34a0af5c-2914-4a67-8ede-f178f08d473f" containerName="mariadb-account-create-update" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.431184 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="34a0af5c-2914-4a67-8ede-f178f08d473f" containerName="mariadb-account-create-update" Dec 07 16:19:01 crc kubenswrapper[4716]: E1207 16:19:01.431193 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f94a6cb-d8ed-47f6-a90a-fa705985ecaf" containerName="keystone-db-sync" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.431199 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f94a6cb-d8ed-47f6-a90a-fa705985ecaf" containerName="keystone-db-sync" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.431337 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b362916-6a42-4fae-88be-f364dc650240" containerName="mariadb-database-create" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.431352 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="50a30fe7-3fb4-4b9d-b186-139c2b48295b" containerName="mariadb-database-create" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.431362 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f94a6cb-d8ed-47f6-a90a-fa705985ecaf" containerName="keystone-db-sync" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.431372 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="34a0af5c-2914-4a67-8ede-f178f08d473f" containerName="mariadb-account-create-update" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.431384 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd43dd05-4a94-4e12-9e2a-788135a9ee76" containerName="mariadb-account-create-update" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.431399 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="862acceb-bcf0-4ef4-9a53-df61d6084806" containerName="mariadb-account-create-update" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.431407 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="666f9fee-767e-40d6-ac44-5abe506e3266" containerName="dnsmasq-dns" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.431415 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d1c048e-719c-4406-bfae-1e5a629534ed" containerName="mariadb-database-create" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.432190 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-6djm5" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.454425 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-26zm5"] Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.455582 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-26zm5" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.458612 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.458909 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.459051 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-lhq7l" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.459193 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.473392 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.478382 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-6djm5"] Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.498355 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0829dcfe-e23e-472a-a01a-d49351ba1f7a-combined-ca-bundle\") pod \"keystone-bootstrap-26zm5\" (UID: \"0829dcfe-e23e-472a-a01a-d49351ba1f7a\") " pod="openstack/keystone-bootstrap-26zm5" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.498613 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0829dcfe-e23e-472a-a01a-d49351ba1f7a-config-data\") pod \"keystone-bootstrap-26zm5\" (UID: \"0829dcfe-e23e-472a-a01a-d49351ba1f7a\") " pod="openstack/keystone-bootstrap-26zm5" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.498648 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1af7e22b-2903-4d7a-bb2b-f565bd44d03b-dns-svc\") pod \"dnsmasq-dns-847c4cc679-6djm5\" (UID: \"1af7e22b-2903-4d7a-bb2b-f565bd44d03b\") " pod="openstack/dnsmasq-dns-847c4cc679-6djm5" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.498669 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-slrzp\" (UniqueName: \"kubernetes.io/projected/0829dcfe-e23e-472a-a01a-d49351ba1f7a-kube-api-access-slrzp\") pod \"keystone-bootstrap-26zm5\" (UID: \"0829dcfe-e23e-472a-a01a-d49351ba1f7a\") " pod="openstack/keystone-bootstrap-26zm5" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.498719 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1af7e22b-2903-4d7a-bb2b-f565bd44d03b-config\") pod \"dnsmasq-dns-847c4cc679-6djm5\" (UID: \"1af7e22b-2903-4d7a-bb2b-f565bd44d03b\") " pod="openstack/dnsmasq-dns-847c4cc679-6djm5" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.498736 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1af7e22b-2903-4d7a-bb2b-f565bd44d03b-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-6djm5\" (UID: \"1af7e22b-2903-4d7a-bb2b-f565bd44d03b\") " pod="openstack/dnsmasq-dns-847c4cc679-6djm5" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.498766 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1af7e22b-2903-4d7a-bb2b-f565bd44d03b-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-6djm5\" (UID: \"1af7e22b-2903-4d7a-bb2b-f565bd44d03b\") " pod="openstack/dnsmasq-dns-847c4cc679-6djm5" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.498783 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1af7e22b-2903-4d7a-bb2b-f565bd44d03b-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-6djm5\" (UID: \"1af7e22b-2903-4d7a-bb2b-f565bd44d03b\") " pod="openstack/dnsmasq-dns-847c4cc679-6djm5" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.498810 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9vrkr\" (UniqueName: \"kubernetes.io/projected/1af7e22b-2903-4d7a-bb2b-f565bd44d03b-kube-api-access-9vrkr\") pod \"dnsmasq-dns-847c4cc679-6djm5\" (UID: \"1af7e22b-2903-4d7a-bb2b-f565bd44d03b\") " pod="openstack/dnsmasq-dns-847c4cc679-6djm5" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.498845 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0829dcfe-e23e-472a-a01a-d49351ba1f7a-fernet-keys\") pod \"keystone-bootstrap-26zm5\" (UID: \"0829dcfe-e23e-472a-a01a-d49351ba1f7a\") " pod="openstack/keystone-bootstrap-26zm5" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.498861 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0829dcfe-e23e-472a-a01a-d49351ba1f7a-credential-keys\") pod \"keystone-bootstrap-26zm5\" (UID: \"0829dcfe-e23e-472a-a01a-d49351ba1f7a\") " pod="openstack/keystone-bootstrap-26zm5" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.498883 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0829dcfe-e23e-472a-a01a-d49351ba1f7a-scripts\") pod \"keystone-bootstrap-26zm5\" (UID: \"0829dcfe-e23e-472a-a01a-d49351ba1f7a\") " pod="openstack/keystone-bootstrap-26zm5" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.517410 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-26zm5"] Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.586828 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-8f6b84b89-znwn9"] Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.588055 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-8f6b84b89-znwn9" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.600212 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9vrkr\" (UniqueName: \"kubernetes.io/projected/1af7e22b-2903-4d7a-bb2b-f565bd44d03b-kube-api-access-9vrkr\") pod \"dnsmasq-dns-847c4cc679-6djm5\" (UID: \"1af7e22b-2903-4d7a-bb2b-f565bd44d03b\") " pod="openstack/dnsmasq-dns-847c4cc679-6djm5" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.600265 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0829dcfe-e23e-472a-a01a-d49351ba1f7a-fernet-keys\") pod \"keystone-bootstrap-26zm5\" (UID: \"0829dcfe-e23e-472a-a01a-d49351ba1f7a\") " pod="openstack/keystone-bootstrap-26zm5" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.600289 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0829dcfe-e23e-472a-a01a-d49351ba1f7a-credential-keys\") pod \"keystone-bootstrap-26zm5\" (UID: \"0829dcfe-e23e-472a-a01a-d49351ba1f7a\") " pod="openstack/keystone-bootstrap-26zm5" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.600312 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0829dcfe-e23e-472a-a01a-d49351ba1f7a-scripts\") pod \"keystone-bootstrap-26zm5\" (UID: \"0829dcfe-e23e-472a-a01a-d49351ba1f7a\") " pod="openstack/keystone-bootstrap-26zm5" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.600342 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0829dcfe-e23e-472a-a01a-d49351ba1f7a-combined-ca-bundle\") pod \"keystone-bootstrap-26zm5\" (UID: \"0829dcfe-e23e-472a-a01a-d49351ba1f7a\") " pod="openstack/keystone-bootstrap-26zm5" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.600361 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0829dcfe-e23e-472a-a01a-d49351ba1f7a-config-data\") pod \"keystone-bootstrap-26zm5\" (UID: \"0829dcfe-e23e-472a-a01a-d49351ba1f7a\") " pod="openstack/keystone-bootstrap-26zm5" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.600387 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1af7e22b-2903-4d7a-bb2b-f565bd44d03b-dns-svc\") pod \"dnsmasq-dns-847c4cc679-6djm5\" (UID: \"1af7e22b-2903-4d7a-bb2b-f565bd44d03b\") " pod="openstack/dnsmasq-dns-847c4cc679-6djm5" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.600406 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-slrzp\" (UniqueName: \"kubernetes.io/projected/0829dcfe-e23e-472a-a01a-d49351ba1f7a-kube-api-access-slrzp\") pod \"keystone-bootstrap-26zm5\" (UID: \"0829dcfe-e23e-472a-a01a-d49351ba1f7a\") " pod="openstack/keystone-bootstrap-26zm5" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.600449 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1af7e22b-2903-4d7a-bb2b-f565bd44d03b-config\") pod \"dnsmasq-dns-847c4cc679-6djm5\" (UID: \"1af7e22b-2903-4d7a-bb2b-f565bd44d03b\") " pod="openstack/dnsmasq-dns-847c4cc679-6djm5" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.600467 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1af7e22b-2903-4d7a-bb2b-f565bd44d03b-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-6djm5\" (UID: \"1af7e22b-2903-4d7a-bb2b-f565bd44d03b\") " pod="openstack/dnsmasq-dns-847c4cc679-6djm5" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.600498 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1af7e22b-2903-4d7a-bb2b-f565bd44d03b-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-6djm5\" (UID: \"1af7e22b-2903-4d7a-bb2b-f565bd44d03b\") " pod="openstack/dnsmasq-dns-847c4cc679-6djm5" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.600518 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1af7e22b-2903-4d7a-bb2b-f565bd44d03b-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-6djm5\" (UID: \"1af7e22b-2903-4d7a-bb2b-f565bd44d03b\") " pod="openstack/dnsmasq-dns-847c4cc679-6djm5" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.601342 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1af7e22b-2903-4d7a-bb2b-f565bd44d03b-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-6djm5\" (UID: \"1af7e22b-2903-4d7a-bb2b-f565bd44d03b\") " pod="openstack/dnsmasq-dns-847c4cc679-6djm5" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.601762 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1af7e22b-2903-4d7a-bb2b-f565bd44d03b-dns-svc\") pod \"dnsmasq-dns-847c4cc679-6djm5\" (UID: \"1af7e22b-2903-4d7a-bb2b-f565bd44d03b\") " pod="openstack/dnsmasq-dns-847c4cc679-6djm5" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.640433 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1af7e22b-2903-4d7a-bb2b-f565bd44d03b-config\") pod \"dnsmasq-dns-847c4cc679-6djm5\" (UID: \"1af7e22b-2903-4d7a-bb2b-f565bd44d03b\") " pod="openstack/dnsmasq-dns-847c4cc679-6djm5" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.640822 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1af7e22b-2903-4d7a-bb2b-f565bd44d03b-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-6djm5\" (UID: \"1af7e22b-2903-4d7a-bb2b-f565bd44d03b\") " pod="openstack/dnsmasq-dns-847c4cc679-6djm5" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.641529 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0829dcfe-e23e-472a-a01a-d49351ba1f7a-combined-ca-bundle\") pod \"keystone-bootstrap-26zm5\" (UID: \"0829dcfe-e23e-472a-a01a-d49351ba1f7a\") " pod="openstack/keystone-bootstrap-26zm5" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.642370 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0829dcfe-e23e-472a-a01a-d49351ba1f7a-fernet-keys\") pod \"keystone-bootstrap-26zm5\" (UID: \"0829dcfe-e23e-472a-a01a-d49351ba1f7a\") " pod="openstack/keystone-bootstrap-26zm5" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.642877 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.643118 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.643491 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-shs9m" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.643980 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.649640 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0829dcfe-e23e-472a-a01a-d49351ba1f7a-credential-keys\") pod \"keystone-bootstrap-26zm5\" (UID: \"0829dcfe-e23e-472a-a01a-d49351ba1f7a\") " pod="openstack/keystone-bootstrap-26zm5" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.651756 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0829dcfe-e23e-472a-a01a-d49351ba1f7a-config-data\") pod \"keystone-bootstrap-26zm5\" (UID: \"0829dcfe-e23e-472a-a01a-d49351ba1f7a\") " pod="openstack/keystone-bootstrap-26zm5" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.653503 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1af7e22b-2903-4d7a-bb2b-f565bd44d03b-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-6djm5\" (UID: \"1af7e22b-2903-4d7a-bb2b-f565bd44d03b\") " pod="openstack/dnsmasq-dns-847c4cc679-6djm5" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.681727 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-slrzp\" (UniqueName: \"kubernetes.io/projected/0829dcfe-e23e-472a-a01a-d49351ba1f7a-kube-api-access-slrzp\") pod \"keystone-bootstrap-26zm5\" (UID: \"0829dcfe-e23e-472a-a01a-d49351ba1f7a\") " pod="openstack/keystone-bootstrap-26zm5" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.729527 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0829dcfe-e23e-472a-a01a-d49351ba1f7a-scripts\") pod \"keystone-bootstrap-26zm5\" (UID: \"0829dcfe-e23e-472a-a01a-d49351ba1f7a\") " pod="openstack/keystone-bootstrap-26zm5" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.771555 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-8f6b84b89-znwn9"] Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.778324 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9vrkr\" (UniqueName: \"kubernetes.io/projected/1af7e22b-2903-4d7a-bb2b-f565bd44d03b-kube-api-access-9vrkr\") pod \"dnsmasq-dns-847c4cc679-6djm5\" (UID: \"1af7e22b-2903-4d7a-bb2b-f565bd44d03b\") " pod="openstack/dnsmasq-dns-847c4cc679-6djm5" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.818594 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-gqg6b"] Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.819926 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-6djm5" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.821010 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff3e516d-0469-4e03-9d71-fed1add36b54-logs\") pod \"horizon-8f6b84b89-znwn9\" (UID: \"ff3e516d-0469-4e03-9d71-fed1add36b54\") " pod="openstack/horizon-8f6b84b89-znwn9" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.822667 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-26zm5" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.825185 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ff3e516d-0469-4e03-9d71-fed1add36b54-horizon-secret-key\") pod \"horizon-8f6b84b89-znwn9\" (UID: \"ff3e516d-0469-4e03-9d71-fed1add36b54\") " pod="openstack/horizon-8f6b84b89-znwn9" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.825283 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n47p4\" (UniqueName: \"kubernetes.io/projected/ff3e516d-0469-4e03-9d71-fed1add36b54-kube-api-access-n47p4\") pod \"horizon-8f6b84b89-znwn9\" (UID: \"ff3e516d-0469-4e03-9d71-fed1add36b54\") " pod="openstack/horizon-8f6b84b89-znwn9" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.825341 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ff3e516d-0469-4e03-9d71-fed1add36b54-config-data\") pod \"horizon-8f6b84b89-znwn9\" (UID: \"ff3e516d-0469-4e03-9d71-fed1add36b54\") " pod="openstack/horizon-8f6b84b89-znwn9" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.825404 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ff3e516d-0469-4e03-9d71-fed1add36b54-scripts\") pod \"horizon-8f6b84b89-znwn9\" (UID: \"ff3e516d-0469-4e03-9d71-fed1add36b54\") " pod="openstack/horizon-8f6b84b89-znwn9" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.825721 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-gqg6b" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.844388 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.844589 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.844703 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-cmrq8" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.854012 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-gqg6b"] Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.905163 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-kl4tr"] Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.906265 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-kl4tr" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.910005 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-pm9qx" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.910254 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.923709 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.930480 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f9413953-056c-4d73-b534-12e6816e6fb4-db-sync-config-data\") pod \"cinder-db-sync-gqg6b\" (UID: \"f9413953-056c-4d73-b534-12e6816e6fb4\") " pod="openstack/cinder-db-sync-gqg6b" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.930514 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9413953-056c-4d73-b534-12e6816e6fb4-combined-ca-bundle\") pod \"cinder-db-sync-gqg6b\" (UID: \"f9413953-056c-4d73-b534-12e6816e6fb4\") " pod="openstack/cinder-db-sync-gqg6b" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.930549 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ff3e516d-0469-4e03-9d71-fed1add36b54-horizon-secret-key\") pod \"horizon-8f6b84b89-znwn9\" (UID: \"ff3e516d-0469-4e03-9d71-fed1add36b54\") " pod="openstack/horizon-8f6b84b89-znwn9" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.930567 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrnjl\" (UniqueName: \"kubernetes.io/projected/f9413953-056c-4d73-b534-12e6816e6fb4-kube-api-access-nrnjl\") pod \"cinder-db-sync-gqg6b\" (UID: \"f9413953-056c-4d73-b534-12e6816e6fb4\") " pod="openstack/cinder-db-sync-gqg6b" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.930596 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f9413953-056c-4d73-b534-12e6816e6fb4-etc-machine-id\") pod \"cinder-db-sync-gqg6b\" (UID: \"f9413953-056c-4d73-b534-12e6816e6fb4\") " pod="openstack/cinder-db-sync-gqg6b" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.930616 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n47p4\" (UniqueName: \"kubernetes.io/projected/ff3e516d-0469-4e03-9d71-fed1add36b54-kube-api-access-n47p4\") pod \"horizon-8f6b84b89-znwn9\" (UID: \"ff3e516d-0469-4e03-9d71-fed1add36b54\") " pod="openstack/horizon-8f6b84b89-znwn9" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.930643 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ff3e516d-0469-4e03-9d71-fed1add36b54-config-data\") pod \"horizon-8f6b84b89-znwn9\" (UID: \"ff3e516d-0469-4e03-9d71-fed1add36b54\") " pod="openstack/horizon-8f6b84b89-znwn9" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.930674 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ff3e516d-0469-4e03-9d71-fed1add36b54-scripts\") pod \"horizon-8f6b84b89-znwn9\" (UID: \"ff3e516d-0469-4e03-9d71-fed1add36b54\") " pod="openstack/horizon-8f6b84b89-znwn9" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.930712 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9413953-056c-4d73-b534-12e6816e6fb4-config-data\") pod \"cinder-db-sync-gqg6b\" (UID: \"f9413953-056c-4d73-b534-12e6816e6fb4\") " pod="openstack/cinder-db-sync-gqg6b" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.930733 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff3e516d-0469-4e03-9d71-fed1add36b54-logs\") pod \"horizon-8f6b84b89-znwn9\" (UID: \"ff3e516d-0469-4e03-9d71-fed1add36b54\") " pod="openstack/horizon-8f6b84b89-znwn9" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.930749 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9413953-056c-4d73-b534-12e6816e6fb4-scripts\") pod \"cinder-db-sync-gqg6b\" (UID: \"f9413953-056c-4d73-b534-12e6816e6fb4\") " pod="openstack/cinder-db-sync-gqg6b" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.933384 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ff3e516d-0469-4e03-9d71-fed1add36b54-scripts\") pod \"horizon-8f6b84b89-znwn9\" (UID: \"ff3e516d-0469-4e03-9d71-fed1add36b54\") " pod="openstack/horizon-8f6b84b89-znwn9" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.933687 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff3e516d-0469-4e03-9d71-fed1add36b54-logs\") pod \"horizon-8f6b84b89-znwn9\" (UID: \"ff3e516d-0469-4e03-9d71-fed1add36b54\") " pod="openstack/horizon-8f6b84b89-znwn9" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.943690 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ff3e516d-0469-4e03-9d71-fed1add36b54-horizon-secret-key\") pod \"horizon-8f6b84b89-znwn9\" (UID: \"ff3e516d-0469-4e03-9d71-fed1add36b54\") " pod="openstack/horizon-8f6b84b89-znwn9" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.963592 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ff3e516d-0469-4e03-9d71-fed1add36b54-config-data\") pod \"horizon-8f6b84b89-znwn9\" (UID: \"ff3e516d-0469-4e03-9d71-fed1add36b54\") " pod="openstack/horizon-8f6b84b89-znwn9" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.977757 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n47p4\" (UniqueName: \"kubernetes.io/projected/ff3e516d-0469-4e03-9d71-fed1add36b54-kube-api-access-n47p4\") pod \"horizon-8f6b84b89-znwn9\" (UID: \"ff3e516d-0469-4e03-9d71-fed1add36b54\") " pod="openstack/horizon-8f6b84b89-znwn9" Dec 07 16:19:01 crc kubenswrapper[4716]: I1207 16:19:01.986891 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-kl4tr"] Dec 07 16:19:02 crc kubenswrapper[4716]: I1207 16:19:02.027294 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-qqklx"] Dec 07 16:19:02 crc kubenswrapper[4716]: I1207 16:19:02.028414 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-qqklx" Dec 07 16:19:02 crc kubenswrapper[4716]: I1207 16:19:02.032254 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/925be76f-fe07-4eb8-982a-02a0a002ea58-combined-ca-bundle\") pod \"neutron-db-sync-kl4tr\" (UID: \"925be76f-fe07-4eb8-982a-02a0a002ea58\") " pod="openstack/neutron-db-sync-kl4tr" Dec 07 16:19:02 crc kubenswrapper[4716]: I1207 16:19:02.032295 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9413953-056c-4d73-b534-12e6816e6fb4-config-data\") pod \"cinder-db-sync-gqg6b\" (UID: \"f9413953-056c-4d73-b534-12e6816e6fb4\") " pod="openstack/cinder-db-sync-gqg6b" Dec 07 16:19:02 crc kubenswrapper[4716]: I1207 16:19:02.032324 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9413953-056c-4d73-b534-12e6816e6fb4-scripts\") pod \"cinder-db-sync-gqg6b\" (UID: \"f9413953-056c-4d73-b534-12e6816e6fb4\") " pod="openstack/cinder-db-sync-gqg6b" Dec 07 16:19:02 crc kubenswrapper[4716]: I1207 16:19:02.032364 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s4nf9\" (UniqueName: \"kubernetes.io/projected/925be76f-fe07-4eb8-982a-02a0a002ea58-kube-api-access-s4nf9\") pod \"neutron-db-sync-kl4tr\" (UID: \"925be76f-fe07-4eb8-982a-02a0a002ea58\") " pod="openstack/neutron-db-sync-kl4tr" Dec 07 16:19:02 crc kubenswrapper[4716]: I1207 16:19:02.032389 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f9413953-056c-4d73-b534-12e6816e6fb4-db-sync-config-data\") pod \"cinder-db-sync-gqg6b\" (UID: \"f9413953-056c-4d73-b534-12e6816e6fb4\") " pod="openstack/cinder-db-sync-gqg6b" Dec 07 16:19:02 crc kubenswrapper[4716]: I1207 16:19:02.032404 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9413953-056c-4d73-b534-12e6816e6fb4-combined-ca-bundle\") pod \"cinder-db-sync-gqg6b\" (UID: \"f9413953-056c-4d73-b534-12e6816e6fb4\") " pod="openstack/cinder-db-sync-gqg6b" Dec 07 16:19:02 crc kubenswrapper[4716]: I1207 16:19:02.032429 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrnjl\" (UniqueName: \"kubernetes.io/projected/f9413953-056c-4d73-b534-12e6816e6fb4-kube-api-access-nrnjl\") pod \"cinder-db-sync-gqg6b\" (UID: \"f9413953-056c-4d73-b534-12e6816e6fb4\") " pod="openstack/cinder-db-sync-gqg6b" Dec 07 16:19:02 crc kubenswrapper[4716]: I1207 16:19:02.032448 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/925be76f-fe07-4eb8-982a-02a0a002ea58-config\") pod \"neutron-db-sync-kl4tr\" (UID: \"925be76f-fe07-4eb8-982a-02a0a002ea58\") " pod="openstack/neutron-db-sync-kl4tr" Dec 07 16:19:02 crc kubenswrapper[4716]: I1207 16:19:02.032470 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f9413953-056c-4d73-b534-12e6816e6fb4-etc-machine-id\") pod \"cinder-db-sync-gqg6b\" (UID: \"f9413953-056c-4d73-b534-12e6816e6fb4\") " pod="openstack/cinder-db-sync-gqg6b" Dec 07 16:19:02 crc kubenswrapper[4716]: I1207 16:19:02.032568 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f9413953-056c-4d73-b534-12e6816e6fb4-etc-machine-id\") pod \"cinder-db-sync-gqg6b\" (UID: \"f9413953-056c-4d73-b534-12e6816e6fb4\") " pod="openstack/cinder-db-sync-gqg6b" Dec 07 16:19:02 crc kubenswrapper[4716]: I1207 16:19:02.038447 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-8f6b84b89-znwn9" Dec 07 16:19:02 crc kubenswrapper[4716]: I1207 16:19:02.040984 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9413953-056c-4d73-b534-12e6816e6fb4-combined-ca-bundle\") pod \"cinder-db-sync-gqg6b\" (UID: \"f9413953-056c-4d73-b534-12e6816e6fb4\") " pod="openstack/cinder-db-sync-gqg6b" Dec 07 16:19:02 crc kubenswrapper[4716]: I1207 16:19:02.041589 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f9413953-056c-4d73-b534-12e6816e6fb4-db-sync-config-data\") pod \"cinder-db-sync-gqg6b\" (UID: \"f9413953-056c-4d73-b534-12e6816e6fb4\") " pod="openstack/cinder-db-sync-gqg6b" Dec 07 16:19:02 crc kubenswrapper[4716]: I1207 16:19:02.041753 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9413953-056c-4d73-b534-12e6816e6fb4-scripts\") pod \"cinder-db-sync-gqg6b\" (UID: \"f9413953-056c-4d73-b534-12e6816e6fb4\") " pod="openstack/cinder-db-sync-gqg6b" Dec 07 16:19:02 crc kubenswrapper[4716]: I1207 16:19:02.042108 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9413953-056c-4d73-b534-12e6816e6fb4-config-data\") pod \"cinder-db-sync-gqg6b\" (UID: \"f9413953-056c-4d73-b534-12e6816e6fb4\") " pod="openstack/cinder-db-sync-gqg6b" Dec 07 16:19:02 crc kubenswrapper[4716]: I1207 16:19:02.048312 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 07 16:19:02 crc kubenswrapper[4716]: I1207 16:19:02.048508 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-dtr4s" Dec 07 16:19:02 crc kubenswrapper[4716]: I1207 16:19:02.048626 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 07 16:19:02 crc kubenswrapper[4716]: I1207 16:19:02.049038 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-qqklx"] Dec 07 16:19:02 crc kubenswrapper[4716]: I1207 16:19:02.109699 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrnjl\" (UniqueName: \"kubernetes.io/projected/f9413953-056c-4d73-b534-12e6816e6fb4-kube-api-access-nrnjl\") pod \"cinder-db-sync-gqg6b\" (UID: \"f9413953-056c-4d73-b534-12e6816e6fb4\") " pod="openstack/cinder-db-sync-gqg6b" Dec 07 16:19:02 crc kubenswrapper[4716]: I1207 16:19:02.110650 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-tk47q"] Dec 07 16:19:02 crc kubenswrapper[4716]: I1207 16:19:02.111802 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-tk47q" Dec 07 16:19:02 crc kubenswrapper[4716]: I1207 16:19:02.117455 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 07 16:19:02 crc kubenswrapper[4716]: I1207 16:19:02.117878 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-kxsqg" Dec 07 16:19:02 crc kubenswrapper[4716]: I1207 16:19:02.130417 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-tk47q"] Dec 07 16:19:02 crc kubenswrapper[4716]: I1207 16:19:02.133943 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6354ab8c-f35e-450f-96f9-8e305f778a54-combined-ca-bundle\") pod \"placement-db-sync-qqklx\" (UID: \"6354ab8c-f35e-450f-96f9-8e305f778a54\") " pod="openstack/placement-db-sync-qqklx" Dec 07 16:19:02 crc kubenswrapper[4716]: I1207 16:19:02.133987 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6354ab8c-f35e-450f-96f9-8e305f778a54-scripts\") pod \"placement-db-sync-qqklx\" (UID: \"6354ab8c-f35e-450f-96f9-8e305f778a54\") " pod="openstack/placement-db-sync-qqklx" Dec 07 16:19:02 crc kubenswrapper[4716]: I1207 16:19:02.134007 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6354ab8c-f35e-450f-96f9-8e305f778a54-logs\") pod \"placement-db-sync-qqklx\" (UID: \"6354ab8c-f35e-450f-96f9-8e305f778a54\") " pod="openstack/placement-db-sync-qqklx" Dec 07 16:19:02 crc kubenswrapper[4716]: I1207 16:19:02.134045 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s4nf9\" (UniqueName: \"kubernetes.io/projected/925be76f-fe07-4eb8-982a-02a0a002ea58-kube-api-access-s4nf9\") pod \"neutron-db-sync-kl4tr\" (UID: \"925be76f-fe07-4eb8-982a-02a0a002ea58\") " pod="openstack/neutron-db-sync-kl4tr" Dec 07 16:19:02 crc kubenswrapper[4716]: I1207 16:19:02.134120 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72mp2\" (UniqueName: \"kubernetes.io/projected/6354ab8c-f35e-450f-96f9-8e305f778a54-kube-api-access-72mp2\") pod \"placement-db-sync-qqklx\" (UID: \"6354ab8c-f35e-450f-96f9-8e305f778a54\") " pod="openstack/placement-db-sync-qqklx" Dec 07 16:19:02 crc kubenswrapper[4716]: I1207 16:19:02.134140 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/925be76f-fe07-4eb8-982a-02a0a002ea58-config\") pod \"neutron-db-sync-kl4tr\" (UID: \"925be76f-fe07-4eb8-982a-02a0a002ea58\") " pod="openstack/neutron-db-sync-kl4tr" Dec 07 16:19:02 crc kubenswrapper[4716]: I1207 16:19:02.134199 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6354ab8c-f35e-450f-96f9-8e305f778a54-config-data\") pod \"placement-db-sync-qqklx\" (UID: \"6354ab8c-f35e-450f-96f9-8e305f778a54\") " pod="openstack/placement-db-sync-qqklx" Dec 07 16:19:02 crc kubenswrapper[4716]: I1207 16:19:02.134236 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/925be76f-fe07-4eb8-982a-02a0a002ea58-combined-ca-bundle\") pod \"neutron-db-sync-kl4tr\" (UID: \"925be76f-fe07-4eb8-982a-02a0a002ea58\") " pod="openstack/neutron-db-sync-kl4tr" Dec 07 16:19:02 crc kubenswrapper[4716]: I1207 16:19:02.138690 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/925be76f-fe07-4eb8-982a-02a0a002ea58-combined-ca-bundle\") pod \"neutron-db-sync-kl4tr\" (UID: \"925be76f-fe07-4eb8-982a-02a0a002ea58\") " pod="openstack/neutron-db-sync-kl4tr" Dec 07 16:19:02 crc kubenswrapper[4716]: I1207 16:19:02.153454 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/925be76f-fe07-4eb8-982a-02a0a002ea58-config\") pod \"neutron-db-sync-kl4tr\" (UID: \"925be76f-fe07-4eb8-982a-02a0a002ea58\") " pod="openstack/neutron-db-sync-kl4tr" Dec 07 16:19:02 crc kubenswrapper[4716]: I1207 16:19:02.153514 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-6djm5"] Dec 07 16:19:02 crc kubenswrapper[4716]: I1207 16:19:02.172627 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s4nf9\" (UniqueName: \"kubernetes.io/projected/925be76f-fe07-4eb8-982a-02a0a002ea58-kube-api-access-s4nf9\") pod \"neutron-db-sync-kl4tr\" (UID: \"925be76f-fe07-4eb8-982a-02a0a002ea58\") " pod="openstack/neutron-db-sync-kl4tr" Dec 07 16:19:02 crc kubenswrapper[4716]: I1207 16:19:02.202763 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7d8b488bdf-fvlzm"] Dec 07 16:19:02 crc kubenswrapper[4716]: I1207 16:19:02.204232 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7d8b488bdf-fvlzm" Dec 07 16:19:02 crc kubenswrapper[4716]: I1207 16:19:02.218706 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-cq5v8"] Dec 07 16:19:02 crc kubenswrapper[4716]: I1207 16:19:02.220167 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-cq5v8" Dec 07 16:19:02 crc kubenswrapper[4716]: I1207 16:19:02.235266 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgdkv\" (UniqueName: \"kubernetes.io/projected/717a326f-b1ac-4fbe-b1c8-b669081dab80-kube-api-access-qgdkv\") pod \"barbican-db-sync-tk47q\" (UID: \"717a326f-b1ac-4fbe-b1c8-b669081dab80\") " pod="openstack/barbican-db-sync-tk47q" Dec 07 16:19:02 crc kubenswrapper[4716]: I1207 16:19:02.235335 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/717a326f-b1ac-4fbe-b1c8-b669081dab80-combined-ca-bundle\") pod \"barbican-db-sync-tk47q\" (UID: \"717a326f-b1ac-4fbe-b1c8-b669081dab80\") " pod="openstack/barbican-db-sync-tk47q" Dec 07 16:19:02 crc kubenswrapper[4716]: I1207 16:19:02.235373 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-72mp2\" (UniqueName: \"kubernetes.io/projected/6354ab8c-f35e-450f-96f9-8e305f778a54-kube-api-access-72mp2\") pod \"placement-db-sync-qqklx\" (UID: \"6354ab8c-f35e-450f-96f9-8e305f778a54\") " pod="openstack/placement-db-sync-qqklx" Dec 07 16:19:02 crc kubenswrapper[4716]: I1207 16:19:02.235432 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6354ab8c-f35e-450f-96f9-8e305f778a54-config-data\") pod \"placement-db-sync-qqklx\" (UID: \"6354ab8c-f35e-450f-96f9-8e305f778a54\") " pod="openstack/placement-db-sync-qqklx" Dec 07 16:19:02 crc kubenswrapper[4716]: I1207 16:19:02.235475 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/717a326f-b1ac-4fbe-b1c8-b669081dab80-db-sync-config-data\") pod \"barbican-db-sync-tk47q\" (UID: \"717a326f-b1ac-4fbe-b1c8-b669081dab80\") " pod="openstack/barbican-db-sync-tk47q" Dec 07 16:19:02 crc kubenswrapper[4716]: I1207 16:19:02.235495 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6354ab8c-f35e-450f-96f9-8e305f778a54-combined-ca-bundle\") pod \"placement-db-sync-qqklx\" (UID: \"6354ab8c-f35e-450f-96f9-8e305f778a54\") " pod="openstack/placement-db-sync-qqklx" Dec 07 16:19:02 crc kubenswrapper[4716]: I1207 16:19:02.235512 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6354ab8c-f35e-450f-96f9-8e305f778a54-scripts\") pod \"placement-db-sync-qqklx\" (UID: \"6354ab8c-f35e-450f-96f9-8e305f778a54\") " pod="openstack/placement-db-sync-qqklx" Dec 07 16:19:02 crc kubenswrapper[4716]: I1207 16:19:02.235529 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6354ab8c-f35e-450f-96f9-8e305f778a54-logs\") pod \"placement-db-sync-qqklx\" (UID: \"6354ab8c-f35e-450f-96f9-8e305f778a54\") " pod="openstack/placement-db-sync-qqklx" Dec 07 16:19:02 crc kubenswrapper[4716]: I1207 16:19:02.237877 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6354ab8c-f35e-450f-96f9-8e305f778a54-logs\") pod \"placement-db-sync-qqklx\" (UID: \"6354ab8c-f35e-450f-96f9-8e305f778a54\") " pod="openstack/placement-db-sync-qqklx" Dec 07 16:19:02 crc kubenswrapper[4716]: I1207 16:19:02.239624 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 07 16:19:02 crc kubenswrapper[4716]: I1207 16:19:02.244524 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6354ab8c-f35e-450f-96f9-8e305f778a54-config-data\") pod \"placement-db-sync-qqklx\" (UID: \"6354ab8c-f35e-450f-96f9-8e305f778a54\") " pod="openstack/placement-db-sync-qqklx" Dec 07 16:19:02 crc kubenswrapper[4716]: I1207 16:19:02.247952 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6354ab8c-f35e-450f-96f9-8e305f778a54-scripts\") pod \"placement-db-sync-qqklx\" (UID: \"6354ab8c-f35e-450f-96f9-8e305f778a54\") " pod="openstack/placement-db-sync-qqklx" Dec 07 16:19:02 crc kubenswrapper[4716]: I1207 16:19:02.250629 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6354ab8c-f35e-450f-96f9-8e305f778a54-combined-ca-bundle\") pod \"placement-db-sync-qqklx\" (UID: \"6354ab8c-f35e-450f-96f9-8e305f778a54\") " pod="openstack/placement-db-sync-qqklx" Dec 07 16:19:02 crc kubenswrapper[4716]: I1207 16:19:02.252860 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7d8b488bdf-fvlzm"] Dec 07 16:19:02 crc kubenswrapper[4716]: I1207 16:19:02.253028 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.259452 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.259620 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.272052 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-cq5v8"] Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.292555 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.293617 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-72mp2\" (UniqueName: \"kubernetes.io/projected/6354ab8c-f35e-450f-96f9-8e305f778a54-kube-api-access-72mp2\") pod \"placement-db-sync-qqklx\" (UID: \"6354ab8c-f35e-450f-96f9-8e305f778a54\") " pod="openstack/placement-db-sync-qqklx" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.294033 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-gqg6b" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.301363 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.308014 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.311362 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-2hfps" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.312409 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.312537 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.312681 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.312816 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.338033 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-kl4tr" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.340602 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hq68h\" (UniqueName: \"kubernetes.io/projected/2ee00bf0-1412-4fca-a570-6600db549923-kube-api-access-hq68h\") pod \"ceilometer-0\" (UID: \"2ee00bf0-1412-4fca-a570-6600db549923\") " pod="openstack/ceilometer-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.340638 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d7db6aff-6831-49ea-9d51-24ec9ac40a88-scripts\") pod \"horizon-7d8b488bdf-fvlzm\" (UID: \"d7db6aff-6831-49ea-9d51-24ec9ac40a88\") " pod="openstack/horizon-7d8b488bdf-fvlzm" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.340663 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qgdkv\" (UniqueName: \"kubernetes.io/projected/717a326f-b1ac-4fbe-b1c8-b669081dab80-kube-api-access-qgdkv\") pod \"barbican-db-sync-tk47q\" (UID: \"717a326f-b1ac-4fbe-b1c8-b669081dab80\") " pod="openstack/barbican-db-sync-tk47q" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.340683 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ee00bf0-1412-4fca-a570-6600db549923-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2ee00bf0-1412-4fca-a570-6600db549923\") " pod="openstack/ceilometer-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.340711 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5ced3e75-e271-433b-ad77-fe9101c18be5-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-cq5v8\" (UID: \"5ced3e75-e271-433b-ad77-fe9101c18be5\") " pod="openstack/dnsmasq-dns-785d8bcb8c-cq5v8" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.340728 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5ced3e75-e271-433b-ad77-fe9101c18be5-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-cq5v8\" (UID: \"5ced3e75-e271-433b-ad77-fe9101c18be5\") " pod="openstack/dnsmasq-dns-785d8bcb8c-cq5v8" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.340745 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cnh2x\" (UniqueName: \"kubernetes.io/projected/d7db6aff-6831-49ea-9d51-24ec9ac40a88-kube-api-access-cnh2x\") pod \"horizon-7d8b488bdf-fvlzm\" (UID: \"d7db6aff-6831-49ea-9d51-24ec9ac40a88\") " pod="openstack/horizon-7d8b488bdf-fvlzm" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.340770 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2ee00bf0-1412-4fca-a570-6600db549923-run-httpd\") pod \"ceilometer-0\" (UID: \"2ee00bf0-1412-4fca-a570-6600db549923\") " pod="openstack/ceilometer-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.340790 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/717a326f-b1ac-4fbe-b1c8-b669081dab80-combined-ca-bundle\") pod \"barbican-db-sync-tk47q\" (UID: \"717a326f-b1ac-4fbe-b1c8-b669081dab80\") " pod="openstack/barbican-db-sync-tk47q" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.340812 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d7db6aff-6831-49ea-9d51-24ec9ac40a88-config-data\") pod \"horizon-7d8b488bdf-fvlzm\" (UID: \"d7db6aff-6831-49ea-9d51-24ec9ac40a88\") " pod="openstack/horizon-7d8b488bdf-fvlzm" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.340848 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2ee00bf0-1412-4fca-a570-6600db549923-log-httpd\") pod \"ceilometer-0\" (UID: \"2ee00bf0-1412-4fca-a570-6600db549923\") " pod="openstack/ceilometer-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.340882 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nzlqz\" (UniqueName: \"kubernetes.io/projected/5ced3e75-e271-433b-ad77-fe9101c18be5-kube-api-access-nzlqz\") pod \"dnsmasq-dns-785d8bcb8c-cq5v8\" (UID: \"5ced3e75-e271-433b-ad77-fe9101c18be5\") " pod="openstack/dnsmasq-dns-785d8bcb8c-cq5v8" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.340917 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2ee00bf0-1412-4fca-a570-6600db549923-scripts\") pod \"ceilometer-0\" (UID: \"2ee00bf0-1412-4fca-a570-6600db549923\") " pod="openstack/ceilometer-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.340935 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d7db6aff-6831-49ea-9d51-24ec9ac40a88-horizon-secret-key\") pod \"horizon-7d8b488bdf-fvlzm\" (UID: \"d7db6aff-6831-49ea-9d51-24ec9ac40a88\") " pod="openstack/horizon-7d8b488bdf-fvlzm" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.340989 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ee00bf0-1412-4fca-a570-6600db549923-config-data\") pod \"ceilometer-0\" (UID: \"2ee00bf0-1412-4fca-a570-6600db549923\") " pod="openstack/ceilometer-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.341011 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d7db6aff-6831-49ea-9d51-24ec9ac40a88-logs\") pod \"horizon-7d8b488bdf-fvlzm\" (UID: \"d7db6aff-6831-49ea-9d51-24ec9ac40a88\") " pod="openstack/horizon-7d8b488bdf-fvlzm" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.341030 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2ee00bf0-1412-4fca-a570-6600db549923-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2ee00bf0-1412-4fca-a570-6600db549923\") " pod="openstack/ceilometer-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.341047 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5ced3e75-e271-433b-ad77-fe9101c18be5-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-cq5v8\" (UID: \"5ced3e75-e271-433b-ad77-fe9101c18be5\") " pod="openstack/dnsmasq-dns-785d8bcb8c-cq5v8" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.341096 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ced3e75-e271-433b-ad77-fe9101c18be5-config\") pod \"dnsmasq-dns-785d8bcb8c-cq5v8\" (UID: \"5ced3e75-e271-433b-ad77-fe9101c18be5\") " pod="openstack/dnsmasq-dns-785d8bcb8c-cq5v8" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.341118 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/717a326f-b1ac-4fbe-b1c8-b669081dab80-db-sync-config-data\") pod \"barbican-db-sync-tk47q\" (UID: \"717a326f-b1ac-4fbe-b1c8-b669081dab80\") " pod="openstack/barbican-db-sync-tk47q" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.341152 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5ced3e75-e271-433b-ad77-fe9101c18be5-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-cq5v8\" (UID: \"5ced3e75-e271-433b-ad77-fe9101c18be5\") " pod="openstack/dnsmasq-dns-785d8bcb8c-cq5v8" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.351556 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/717a326f-b1ac-4fbe-b1c8-b669081dab80-db-sync-config-data\") pod \"barbican-db-sync-tk47q\" (UID: \"717a326f-b1ac-4fbe-b1c8-b669081dab80\") " pod="openstack/barbican-db-sync-tk47q" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.352295 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/717a326f-b1ac-4fbe-b1c8-b669081dab80-combined-ca-bundle\") pod \"barbican-db-sync-tk47q\" (UID: \"717a326f-b1ac-4fbe-b1c8-b669081dab80\") " pod="openstack/barbican-db-sync-tk47q" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.358565 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qgdkv\" (UniqueName: \"kubernetes.io/projected/717a326f-b1ac-4fbe-b1c8-b669081dab80-kube-api-access-qgdkv\") pod \"barbican-db-sync-tk47q\" (UID: \"717a326f-b1ac-4fbe-b1c8-b669081dab80\") " pod="openstack/barbican-db-sync-tk47q" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.374566 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-qqklx" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.442340 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"43211ed2-3960-41b4-a4dc-1207a5b0d2bc\") " pod="openstack/glance-default-external-api-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.442415 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5ced3e75-e271-433b-ad77-fe9101c18be5-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-cq5v8\" (UID: \"5ced3e75-e271-433b-ad77-fe9101c18be5\") " pod="openstack/dnsmasq-dns-785d8bcb8c-cq5v8" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.442469 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hq68h\" (UniqueName: \"kubernetes.io/projected/2ee00bf0-1412-4fca-a570-6600db549923-kube-api-access-hq68h\") pod \"ceilometer-0\" (UID: \"2ee00bf0-1412-4fca-a570-6600db549923\") " pod="openstack/ceilometer-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.442489 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d7db6aff-6831-49ea-9d51-24ec9ac40a88-scripts\") pod \"horizon-7d8b488bdf-fvlzm\" (UID: \"d7db6aff-6831-49ea-9d51-24ec9ac40a88\") " pod="openstack/horizon-7d8b488bdf-fvlzm" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.442508 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/43211ed2-3960-41b4-a4dc-1207a5b0d2bc-scripts\") pod \"glance-default-external-api-0\" (UID: \"43211ed2-3960-41b4-a4dc-1207a5b0d2bc\") " pod="openstack/glance-default-external-api-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.442546 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ee00bf0-1412-4fca-a570-6600db549923-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2ee00bf0-1412-4fca-a570-6600db549923\") " pod="openstack/ceilometer-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.442573 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5ced3e75-e271-433b-ad77-fe9101c18be5-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-cq5v8\" (UID: \"5ced3e75-e271-433b-ad77-fe9101c18be5\") " pod="openstack/dnsmasq-dns-785d8bcb8c-cq5v8" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.442590 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5ced3e75-e271-433b-ad77-fe9101c18be5-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-cq5v8\" (UID: \"5ced3e75-e271-433b-ad77-fe9101c18be5\") " pod="openstack/dnsmasq-dns-785d8bcb8c-cq5v8" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.442629 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43211ed2-3960-41b4-a4dc-1207a5b0d2bc-config-data\") pod \"glance-default-external-api-0\" (UID: \"43211ed2-3960-41b4-a4dc-1207a5b0d2bc\") " pod="openstack/glance-default-external-api-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.442647 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cnh2x\" (UniqueName: \"kubernetes.io/projected/d7db6aff-6831-49ea-9d51-24ec9ac40a88-kube-api-access-cnh2x\") pod \"horizon-7d8b488bdf-fvlzm\" (UID: \"d7db6aff-6831-49ea-9d51-24ec9ac40a88\") " pod="openstack/horizon-7d8b488bdf-fvlzm" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.442669 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43211ed2-3960-41b4-a4dc-1207a5b0d2bc-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"43211ed2-3960-41b4-a4dc-1207a5b0d2bc\") " pod="openstack/glance-default-external-api-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.442714 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2ee00bf0-1412-4fca-a570-6600db549923-run-httpd\") pod \"ceilometer-0\" (UID: \"2ee00bf0-1412-4fca-a570-6600db549923\") " pod="openstack/ceilometer-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.442730 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d7db6aff-6831-49ea-9d51-24ec9ac40a88-config-data\") pod \"horizon-7d8b488bdf-fvlzm\" (UID: \"d7db6aff-6831-49ea-9d51-24ec9ac40a88\") " pod="openstack/horizon-7d8b488bdf-fvlzm" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.442770 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2ee00bf0-1412-4fca-a570-6600db549923-log-httpd\") pod \"ceilometer-0\" (UID: \"2ee00bf0-1412-4fca-a570-6600db549923\") " pod="openstack/ceilometer-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.442787 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/43211ed2-3960-41b4-a4dc-1207a5b0d2bc-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"43211ed2-3960-41b4-a4dc-1207a5b0d2bc\") " pod="openstack/glance-default-external-api-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.442809 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nzlqz\" (UniqueName: \"kubernetes.io/projected/5ced3e75-e271-433b-ad77-fe9101c18be5-kube-api-access-nzlqz\") pod \"dnsmasq-dns-785d8bcb8c-cq5v8\" (UID: \"5ced3e75-e271-433b-ad77-fe9101c18be5\") " pod="openstack/dnsmasq-dns-785d8bcb8c-cq5v8" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.442846 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2g9tv\" (UniqueName: \"kubernetes.io/projected/43211ed2-3960-41b4-a4dc-1207a5b0d2bc-kube-api-access-2g9tv\") pod \"glance-default-external-api-0\" (UID: \"43211ed2-3960-41b4-a4dc-1207a5b0d2bc\") " pod="openstack/glance-default-external-api-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.442870 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2ee00bf0-1412-4fca-a570-6600db549923-scripts\") pod \"ceilometer-0\" (UID: \"2ee00bf0-1412-4fca-a570-6600db549923\") " pod="openstack/ceilometer-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.442888 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d7db6aff-6831-49ea-9d51-24ec9ac40a88-horizon-secret-key\") pod \"horizon-7d8b488bdf-fvlzm\" (UID: \"d7db6aff-6831-49ea-9d51-24ec9ac40a88\") " pod="openstack/horizon-7d8b488bdf-fvlzm" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.443173 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ee00bf0-1412-4fca-a570-6600db549923-config-data\") pod \"ceilometer-0\" (UID: \"2ee00bf0-1412-4fca-a570-6600db549923\") " pod="openstack/ceilometer-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.443199 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/43211ed2-3960-41b4-a4dc-1207a5b0d2bc-logs\") pod \"glance-default-external-api-0\" (UID: \"43211ed2-3960-41b4-a4dc-1207a5b0d2bc\") " pod="openstack/glance-default-external-api-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.443234 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/43211ed2-3960-41b4-a4dc-1207a5b0d2bc-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"43211ed2-3960-41b4-a4dc-1207a5b0d2bc\") " pod="openstack/glance-default-external-api-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.443253 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d7db6aff-6831-49ea-9d51-24ec9ac40a88-logs\") pod \"horizon-7d8b488bdf-fvlzm\" (UID: \"d7db6aff-6831-49ea-9d51-24ec9ac40a88\") " pod="openstack/horizon-7d8b488bdf-fvlzm" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.443272 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2ee00bf0-1412-4fca-a570-6600db549923-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2ee00bf0-1412-4fca-a570-6600db549923\") " pod="openstack/ceilometer-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.443305 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5ced3e75-e271-433b-ad77-fe9101c18be5-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-cq5v8\" (UID: \"5ced3e75-e271-433b-ad77-fe9101c18be5\") " pod="openstack/dnsmasq-dns-785d8bcb8c-cq5v8" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.443335 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ced3e75-e271-433b-ad77-fe9101c18be5-config\") pod \"dnsmasq-dns-785d8bcb8c-cq5v8\" (UID: \"5ced3e75-e271-433b-ad77-fe9101c18be5\") " pod="openstack/dnsmasq-dns-785d8bcb8c-cq5v8" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.445249 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ced3e75-e271-433b-ad77-fe9101c18be5-config\") pod \"dnsmasq-dns-785d8bcb8c-cq5v8\" (UID: \"5ced3e75-e271-433b-ad77-fe9101c18be5\") " pod="openstack/dnsmasq-dns-785d8bcb8c-cq5v8" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.445940 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5ced3e75-e271-433b-ad77-fe9101c18be5-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-cq5v8\" (UID: \"5ced3e75-e271-433b-ad77-fe9101c18be5\") " pod="openstack/dnsmasq-dns-785d8bcb8c-cq5v8" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.446258 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2ee00bf0-1412-4fca-a570-6600db549923-log-httpd\") pod \"ceilometer-0\" (UID: \"2ee00bf0-1412-4fca-a570-6600db549923\") " pod="openstack/ceilometer-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.447212 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d7db6aff-6831-49ea-9d51-24ec9ac40a88-scripts\") pod \"horizon-7d8b488bdf-fvlzm\" (UID: \"d7db6aff-6831-49ea-9d51-24ec9ac40a88\") " pod="openstack/horizon-7d8b488bdf-fvlzm" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.450033 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2ee00bf0-1412-4fca-a570-6600db549923-scripts\") pod \"ceilometer-0\" (UID: \"2ee00bf0-1412-4fca-a570-6600db549923\") " pod="openstack/ceilometer-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.452211 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ee00bf0-1412-4fca-a570-6600db549923-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2ee00bf0-1412-4fca-a570-6600db549923\") " pod="openstack/ceilometer-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.453533 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2ee00bf0-1412-4fca-a570-6600db549923-run-httpd\") pod \"ceilometer-0\" (UID: \"2ee00bf0-1412-4fca-a570-6600db549923\") " pod="openstack/ceilometer-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.453976 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5ced3e75-e271-433b-ad77-fe9101c18be5-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-cq5v8\" (UID: \"5ced3e75-e271-433b-ad77-fe9101c18be5\") " pod="openstack/dnsmasq-dns-785d8bcb8c-cq5v8" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.454755 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5ced3e75-e271-433b-ad77-fe9101c18be5-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-cq5v8\" (UID: \"5ced3e75-e271-433b-ad77-fe9101c18be5\") " pod="openstack/dnsmasq-dns-785d8bcb8c-cq5v8" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.455312 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d7db6aff-6831-49ea-9d51-24ec9ac40a88-logs\") pod \"horizon-7d8b488bdf-fvlzm\" (UID: \"d7db6aff-6831-49ea-9d51-24ec9ac40a88\") " pod="openstack/horizon-7d8b488bdf-fvlzm" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.460469 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d7db6aff-6831-49ea-9d51-24ec9ac40a88-config-data\") pod \"horizon-7d8b488bdf-fvlzm\" (UID: \"d7db6aff-6831-49ea-9d51-24ec9ac40a88\") " pod="openstack/horizon-7d8b488bdf-fvlzm" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.461000 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-tk47q" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.462102 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d7db6aff-6831-49ea-9d51-24ec9ac40a88-horizon-secret-key\") pod \"horizon-7d8b488bdf-fvlzm\" (UID: \"d7db6aff-6831-49ea-9d51-24ec9ac40a88\") " pod="openstack/horizon-7d8b488bdf-fvlzm" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.465105 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nzlqz\" (UniqueName: \"kubernetes.io/projected/5ced3e75-e271-433b-ad77-fe9101c18be5-kube-api-access-nzlqz\") pod \"dnsmasq-dns-785d8bcb8c-cq5v8\" (UID: \"5ced3e75-e271-433b-ad77-fe9101c18be5\") " pod="openstack/dnsmasq-dns-785d8bcb8c-cq5v8" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.466560 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5ced3e75-e271-433b-ad77-fe9101c18be5-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-cq5v8\" (UID: \"5ced3e75-e271-433b-ad77-fe9101c18be5\") " pod="openstack/dnsmasq-dns-785d8bcb8c-cq5v8" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.469227 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ee00bf0-1412-4fca-a570-6600db549923-config-data\") pod \"ceilometer-0\" (UID: \"2ee00bf0-1412-4fca-a570-6600db549923\") " pod="openstack/ceilometer-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.471195 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hq68h\" (UniqueName: \"kubernetes.io/projected/2ee00bf0-1412-4fca-a570-6600db549923-kube-api-access-hq68h\") pod \"ceilometer-0\" (UID: \"2ee00bf0-1412-4fca-a570-6600db549923\") " pod="openstack/ceilometer-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.471886 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2ee00bf0-1412-4fca-a570-6600db549923-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2ee00bf0-1412-4fca-a570-6600db549923\") " pod="openstack/ceilometer-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.472613 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cnh2x\" (UniqueName: \"kubernetes.io/projected/d7db6aff-6831-49ea-9d51-24ec9ac40a88-kube-api-access-cnh2x\") pod \"horizon-7d8b488bdf-fvlzm\" (UID: \"d7db6aff-6831-49ea-9d51-24ec9ac40a88\") " pod="openstack/horizon-7d8b488bdf-fvlzm" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.544750 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/43211ed2-3960-41b4-a4dc-1207a5b0d2bc-scripts\") pod \"glance-default-external-api-0\" (UID: \"43211ed2-3960-41b4-a4dc-1207a5b0d2bc\") " pod="openstack/glance-default-external-api-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.544803 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43211ed2-3960-41b4-a4dc-1207a5b0d2bc-config-data\") pod \"glance-default-external-api-0\" (UID: \"43211ed2-3960-41b4-a4dc-1207a5b0d2bc\") " pod="openstack/glance-default-external-api-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.544827 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43211ed2-3960-41b4-a4dc-1207a5b0d2bc-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"43211ed2-3960-41b4-a4dc-1207a5b0d2bc\") " pod="openstack/glance-default-external-api-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.544857 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/43211ed2-3960-41b4-a4dc-1207a5b0d2bc-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"43211ed2-3960-41b4-a4dc-1207a5b0d2bc\") " pod="openstack/glance-default-external-api-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.544885 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2g9tv\" (UniqueName: \"kubernetes.io/projected/43211ed2-3960-41b4-a4dc-1207a5b0d2bc-kube-api-access-2g9tv\") pod \"glance-default-external-api-0\" (UID: \"43211ed2-3960-41b4-a4dc-1207a5b0d2bc\") " pod="openstack/glance-default-external-api-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.544954 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/43211ed2-3960-41b4-a4dc-1207a5b0d2bc-logs\") pod \"glance-default-external-api-0\" (UID: \"43211ed2-3960-41b4-a4dc-1207a5b0d2bc\") " pod="openstack/glance-default-external-api-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.545063 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/43211ed2-3960-41b4-a4dc-1207a5b0d2bc-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"43211ed2-3960-41b4-a4dc-1207a5b0d2bc\") " pod="openstack/glance-default-external-api-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.545132 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"43211ed2-3960-41b4-a4dc-1207a5b0d2bc\") " pod="openstack/glance-default-external-api-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.545560 4716 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"43211ed2-3960-41b4-a4dc-1207a5b0d2bc\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-external-api-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.546424 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/43211ed2-3960-41b4-a4dc-1207a5b0d2bc-logs\") pod \"glance-default-external-api-0\" (UID: \"43211ed2-3960-41b4-a4dc-1207a5b0d2bc\") " pod="openstack/glance-default-external-api-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.546823 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/43211ed2-3960-41b4-a4dc-1207a5b0d2bc-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"43211ed2-3960-41b4-a4dc-1207a5b0d2bc\") " pod="openstack/glance-default-external-api-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.553345 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43211ed2-3960-41b4-a4dc-1207a5b0d2bc-config-data\") pod \"glance-default-external-api-0\" (UID: \"43211ed2-3960-41b4-a4dc-1207a5b0d2bc\") " pod="openstack/glance-default-external-api-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.556515 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/43211ed2-3960-41b4-a4dc-1207a5b0d2bc-scripts\") pod \"glance-default-external-api-0\" (UID: \"43211ed2-3960-41b4-a4dc-1207a5b0d2bc\") " pod="openstack/glance-default-external-api-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.558654 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/43211ed2-3960-41b4-a4dc-1207a5b0d2bc-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"43211ed2-3960-41b4-a4dc-1207a5b0d2bc\") " pod="openstack/glance-default-external-api-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.572580 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2g9tv\" (UniqueName: \"kubernetes.io/projected/43211ed2-3960-41b4-a4dc-1207a5b0d2bc-kube-api-access-2g9tv\") pod \"glance-default-external-api-0\" (UID: \"43211ed2-3960-41b4-a4dc-1207a5b0d2bc\") " pod="openstack/glance-default-external-api-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.605709 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43211ed2-3960-41b4-a4dc-1207a5b0d2bc-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"43211ed2-3960-41b4-a4dc-1207a5b0d2bc\") " pod="openstack/glance-default-external-api-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.612501 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"43211ed2-3960-41b4-a4dc-1207a5b0d2bc\") " pod="openstack/glance-default-external-api-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.656171 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7d8b488bdf-fvlzm" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.656864 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-cq5v8" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.676845 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.695540 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.800726 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.801998 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.803690 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.804005 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.815205 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.852711 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e\") " pod="openstack/glance-default-internal-api-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.852776 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e\") " pod="openstack/glance-default-internal-api-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.852938 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e\") " pod="openstack/glance-default-internal-api-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.853018 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e\") " pod="openstack/glance-default-internal-api-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.853064 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e-logs\") pod \"glance-default-internal-api-0\" (UID: \"d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e\") " pod="openstack/glance-default-internal-api-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.853187 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e\") " pod="openstack/glance-default-internal-api-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.853247 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dbf6q\" (UniqueName: \"kubernetes.io/projected/d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e-kube-api-access-dbf6q\") pod \"glance-default-internal-api-0\" (UID: \"d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e\") " pod="openstack/glance-default-internal-api-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.853379 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e\") " pod="openstack/glance-default-internal-api-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.954781 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e\") " pod="openstack/glance-default-internal-api-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.954829 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e\") " pod="openstack/glance-default-internal-api-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.954850 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e-logs\") pod \"glance-default-internal-api-0\" (UID: \"d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e\") " pod="openstack/glance-default-internal-api-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.954897 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e\") " pod="openstack/glance-default-internal-api-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.954915 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dbf6q\" (UniqueName: \"kubernetes.io/projected/d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e-kube-api-access-dbf6q\") pod \"glance-default-internal-api-0\" (UID: \"d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e\") " pod="openstack/glance-default-internal-api-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.954968 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e\") " pod="openstack/glance-default-internal-api-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.955027 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e\") " pod="openstack/glance-default-internal-api-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.955047 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e\") " pod="openstack/glance-default-internal-api-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.955576 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e-logs\") pod \"glance-default-internal-api-0\" (UID: \"d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e\") " pod="openstack/glance-default-internal-api-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.955757 4716 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-internal-api-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.955882 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e\") " pod="openstack/glance-default-internal-api-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.959536 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e\") " pod="openstack/glance-default-internal-api-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.960760 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e\") " pod="openstack/glance-default-internal-api-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.962556 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e\") " pod="openstack/glance-default-internal-api-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.972567 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e\") " pod="openstack/glance-default-internal-api-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.974917 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dbf6q\" (UniqueName: \"kubernetes.io/projected/d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e-kube-api-access-dbf6q\") pod \"glance-default-internal-api-0\" (UID: \"d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e\") " pod="openstack/glance-default-internal-api-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:02.982222 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e\") " pod="openstack/glance-default-internal-api-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:03.155511 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:03.734997 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:03.797959 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-8f6b84b89-znwn9"] Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:03.857162 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7d76599897-4mhqt"] Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:03.858958 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:03.859055 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7d76599897-4mhqt" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:03.873243 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5c76ced8-9170-4ccc-a497-87388e8546dd-config-data\") pod \"horizon-7d76599897-4mhqt\" (UID: \"5c76ced8-9170-4ccc-a497-87388e8546dd\") " pod="openstack/horizon-7d76599897-4mhqt" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:03.873319 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8w7nq\" (UniqueName: \"kubernetes.io/projected/5c76ced8-9170-4ccc-a497-87388e8546dd-kube-api-access-8w7nq\") pod \"horizon-7d76599897-4mhqt\" (UID: \"5c76ced8-9170-4ccc-a497-87388e8546dd\") " pod="openstack/horizon-7d76599897-4mhqt" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:03.873370 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5c76ced8-9170-4ccc-a497-87388e8546dd-scripts\") pod \"horizon-7d76599897-4mhqt\" (UID: \"5c76ced8-9170-4ccc-a497-87388e8546dd\") " pod="openstack/horizon-7d76599897-4mhqt" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:03.873416 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c76ced8-9170-4ccc-a497-87388e8546dd-logs\") pod \"horizon-7d76599897-4mhqt\" (UID: \"5c76ced8-9170-4ccc-a497-87388e8546dd\") " pod="openstack/horizon-7d76599897-4mhqt" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:03.873499 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5c76ced8-9170-4ccc-a497-87388e8546dd-horizon-secret-key\") pod \"horizon-7d76599897-4mhqt\" (UID: \"5c76ced8-9170-4ccc-a497-87388e8546dd\") " pod="openstack/horizon-7d76599897-4mhqt" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:03.893165 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7d76599897-4mhqt"] Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:03.978155 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5c76ced8-9170-4ccc-a497-87388e8546dd-config-data\") pod \"horizon-7d76599897-4mhqt\" (UID: \"5c76ced8-9170-4ccc-a497-87388e8546dd\") " pod="openstack/horizon-7d76599897-4mhqt" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:03.978219 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8w7nq\" (UniqueName: \"kubernetes.io/projected/5c76ced8-9170-4ccc-a497-87388e8546dd-kube-api-access-8w7nq\") pod \"horizon-7d76599897-4mhqt\" (UID: \"5c76ced8-9170-4ccc-a497-87388e8546dd\") " pod="openstack/horizon-7d76599897-4mhqt" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:03.978312 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5c76ced8-9170-4ccc-a497-87388e8546dd-scripts\") pod \"horizon-7d76599897-4mhqt\" (UID: \"5c76ced8-9170-4ccc-a497-87388e8546dd\") " pod="openstack/horizon-7d76599897-4mhqt" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:03.978347 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c76ced8-9170-4ccc-a497-87388e8546dd-logs\") pod \"horizon-7d76599897-4mhqt\" (UID: \"5c76ced8-9170-4ccc-a497-87388e8546dd\") " pod="openstack/horizon-7d76599897-4mhqt" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:03.978402 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5c76ced8-9170-4ccc-a497-87388e8546dd-horizon-secret-key\") pod \"horizon-7d76599897-4mhqt\" (UID: \"5c76ced8-9170-4ccc-a497-87388e8546dd\") " pod="openstack/horizon-7d76599897-4mhqt" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:03.979579 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c76ced8-9170-4ccc-a497-87388e8546dd-logs\") pod \"horizon-7d76599897-4mhqt\" (UID: \"5c76ced8-9170-4ccc-a497-87388e8546dd\") " pod="openstack/horizon-7d76599897-4mhqt" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:03.979870 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5c76ced8-9170-4ccc-a497-87388e8546dd-scripts\") pod \"horizon-7d76599897-4mhqt\" (UID: \"5c76ced8-9170-4ccc-a497-87388e8546dd\") " pod="openstack/horizon-7d76599897-4mhqt" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:03.980066 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5c76ced8-9170-4ccc-a497-87388e8546dd-config-data\") pod \"horizon-7d76599897-4mhqt\" (UID: \"5c76ced8-9170-4ccc-a497-87388e8546dd\") " pod="openstack/horizon-7d76599897-4mhqt" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:03.993395 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5c76ced8-9170-4ccc-a497-87388e8546dd-horizon-secret-key\") pod \"horizon-7d76599897-4mhqt\" (UID: \"5c76ced8-9170-4ccc-a497-87388e8546dd\") " pod="openstack/horizon-7d76599897-4mhqt" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:03.995838 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8w7nq\" (UniqueName: \"kubernetes.io/projected/5c76ced8-9170-4ccc-a497-87388e8546dd-kube-api-access-8w7nq\") pod \"horizon-7d76599897-4mhqt\" (UID: \"5c76ced8-9170-4ccc-a497-87388e8546dd\") " pod="openstack/horizon-7d76599897-4mhqt" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:04.186505 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7d76599897-4mhqt" Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:04.258166 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:04.369143 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-6djm5"] Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:04.380161 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-26zm5"] Dec 07 16:19:04 crc kubenswrapper[4716]: W1207 16:19:04.388296 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0829dcfe_e23e_472a_a01a_d49351ba1f7a.slice/crio-3f4fc79efa766d7ca00151321768a20ad163a8bdafe72fedbcfd442ddf7c6ba0 WatchSource:0}: Error finding container 3f4fc79efa766d7ca00151321768a20ad163a8bdafe72fedbcfd442ddf7c6ba0: Status 404 returned error can't find the container with id 3f4fc79efa766d7ca00151321768a20ad163a8bdafe72fedbcfd442ddf7c6ba0 Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:04.769267 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-kl4tr"] Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:04.788410 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-qqklx"] Dec 07 16:19:04 crc kubenswrapper[4716]: W1207 16:19:04.813034 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod925be76f_fe07_4eb8_982a_02a0a002ea58.slice/crio-a41869ed94b7de3cafdecb2332a0dd22951bc27f6dc071e3b2b8d6fafde9e4ab WatchSource:0}: Error finding container a41869ed94b7de3cafdecb2332a0dd22951bc27f6dc071e3b2b8d6fafde9e4ab: Status 404 returned error can't find the container with id a41869ed94b7de3cafdecb2332a0dd22951bc27f6dc071e3b2b8d6fafde9e4ab Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:04.816502 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-8f6b84b89-znwn9"] Dec 07 16:19:04 crc kubenswrapper[4716]: W1207 16:19:04.821255 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod717a326f_b1ac_4fbe_b1c8_b669081dab80.slice/crio-15f2b878a373b032d52f19fadc7e27b03a3417197ec85e6a2ffcfeb5203f21c8 WatchSource:0}: Error finding container 15f2b878a373b032d52f19fadc7e27b03a3417197ec85e6a2ffcfeb5203f21c8: Status 404 returned error can't find the container with id 15f2b878a373b032d52f19fadc7e27b03a3417197ec85e6a2ffcfeb5203f21c8 Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:04.853777 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:04.892604 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-gqg6b"] Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:04.906160 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-cq5v8"] Dec 07 16:19:04 crc kubenswrapper[4716]: W1207 16:19:04.919467 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5c76ced8_9170_4ccc_a497_87388e8546dd.slice/crio-b23cf3815dfabcc10c50609de078972833d877d6fe2f266489c3f08fd714a42c WatchSource:0}: Error finding container b23cf3815dfabcc10c50609de078972833d877d6fe2f266489c3f08fd714a42c: Status 404 returned error can't find the container with id b23cf3815dfabcc10c50609de078972833d877d6fe2f266489c3f08fd714a42c Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:04.925550 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-tk47q"] Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:04.938559 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7d8b488bdf-fvlzm"] Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:04.973898 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 07 16:19:04 crc kubenswrapper[4716]: I1207 16:19:04.980552 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7d76599897-4mhqt"] Dec 07 16:19:05 crc kubenswrapper[4716]: I1207 16:19:05.196764 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-tk47q" event={"ID":"717a326f-b1ac-4fbe-b1c8-b669081dab80","Type":"ContainerStarted","Data":"15f2b878a373b032d52f19fadc7e27b03a3417197ec85e6a2ffcfeb5203f21c8"} Dec 07 16:19:05 crc kubenswrapper[4716]: I1207 16:19:05.198241 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"43211ed2-3960-41b4-a4dc-1207a5b0d2bc","Type":"ContainerStarted","Data":"1ca9f64e2db9a3b8d9290607ee0ff9ee2d5c2d77aabdf1b4a71036e5b4781802"} Dec 07 16:19:05 crc kubenswrapper[4716]: I1207 16:19:05.200662 4716 generic.go:334] "Generic (PLEG): container finished" podID="5ced3e75-e271-433b-ad77-fe9101c18be5" containerID="04594e235359bf223f892f9d9da526ca2df5e5e5ccdd1a9c7517fa66eb8ca47c" exitCode=0 Dec 07 16:19:05 crc kubenswrapper[4716]: I1207 16:19:05.200707 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-cq5v8" event={"ID":"5ced3e75-e271-433b-ad77-fe9101c18be5","Type":"ContainerDied","Data":"04594e235359bf223f892f9d9da526ca2df5e5e5ccdd1a9c7517fa66eb8ca47c"} Dec 07 16:19:05 crc kubenswrapper[4716]: I1207 16:19:05.200735 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-cq5v8" event={"ID":"5ced3e75-e271-433b-ad77-fe9101c18be5","Type":"ContainerStarted","Data":"1306ff3757f816247c4c49b1313e0ecbee33cc633d87c01a1da6f9bff3f01822"} Dec 07 16:19:05 crc kubenswrapper[4716]: I1207 16:19:05.202320 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-gqg6b" event={"ID":"f9413953-056c-4d73-b534-12e6816e6fb4","Type":"ContainerStarted","Data":"73ef3cdbf94738d92af0605456c808535aff71aa960d87d8943550a902ca6331"} Dec 07 16:19:05 crc kubenswrapper[4716]: I1207 16:19:05.204918 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-8f6b84b89-znwn9" event={"ID":"ff3e516d-0469-4e03-9d71-fed1add36b54","Type":"ContainerStarted","Data":"3722bd5a37148297b80d2466823ef2b5a4826285863eeff61619558b85b50c6d"} Dec 07 16:19:05 crc kubenswrapper[4716]: I1207 16:19:05.206564 4716 generic.go:334] "Generic (PLEG): container finished" podID="1af7e22b-2903-4d7a-bb2b-f565bd44d03b" containerID="040363d15e81674e3c171a80e9f51c850ee79a715574ce17e4ebb01b79ee6094" exitCode=0 Dec 07 16:19:05 crc kubenswrapper[4716]: I1207 16:19:05.206623 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-6djm5" event={"ID":"1af7e22b-2903-4d7a-bb2b-f565bd44d03b","Type":"ContainerDied","Data":"040363d15e81674e3c171a80e9f51c850ee79a715574ce17e4ebb01b79ee6094"} Dec 07 16:19:05 crc kubenswrapper[4716]: I1207 16:19:05.206638 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-6djm5" event={"ID":"1af7e22b-2903-4d7a-bb2b-f565bd44d03b","Type":"ContainerStarted","Data":"e6a709ba91d278cca1171cbca36223dad59798c8aae3bfa0996b30956f5cce94"} Dec 07 16:19:05 crc kubenswrapper[4716]: I1207 16:19:05.214704 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-26zm5" event={"ID":"0829dcfe-e23e-472a-a01a-d49351ba1f7a","Type":"ContainerStarted","Data":"49e2ed79567e1481a18a12c7c4b8d770a72017bbcff0fed08ff63919176cc886"} Dec 07 16:19:05 crc kubenswrapper[4716]: I1207 16:19:05.215056 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-26zm5" event={"ID":"0829dcfe-e23e-472a-a01a-d49351ba1f7a","Type":"ContainerStarted","Data":"3f4fc79efa766d7ca00151321768a20ad163a8bdafe72fedbcfd442ddf7c6ba0"} Dec 07 16:19:05 crc kubenswrapper[4716]: I1207 16:19:05.221909 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2ee00bf0-1412-4fca-a570-6600db549923","Type":"ContainerStarted","Data":"b0681f205b22e31369df55d644f1428738617f1f21c4866b0798f167f6aee150"} Dec 07 16:19:05 crc kubenswrapper[4716]: I1207 16:19:05.223769 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-kl4tr" event={"ID":"925be76f-fe07-4eb8-982a-02a0a002ea58","Type":"ContainerStarted","Data":"f899907d58ebda8ca729b889e0d1a75de6da61c3d90e525bba1d8e11925ba380"} Dec 07 16:19:05 crc kubenswrapper[4716]: I1207 16:19:05.223793 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-kl4tr" event={"ID":"925be76f-fe07-4eb8-982a-02a0a002ea58","Type":"ContainerStarted","Data":"a41869ed94b7de3cafdecb2332a0dd22951bc27f6dc071e3b2b8d6fafde9e4ab"} Dec 07 16:19:05 crc kubenswrapper[4716]: I1207 16:19:05.224968 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-qqklx" event={"ID":"6354ab8c-f35e-450f-96f9-8e305f778a54","Type":"ContainerStarted","Data":"a615881d89bc426d8e5e0ae90cf47c8e936a3781f1552654e42c25608b198950"} Dec 07 16:19:05 crc kubenswrapper[4716]: I1207 16:19:05.227328 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d76599897-4mhqt" event={"ID":"5c76ced8-9170-4ccc-a497-87388e8546dd","Type":"ContainerStarted","Data":"b23cf3815dfabcc10c50609de078972833d877d6fe2f266489c3f08fd714a42c"} Dec 07 16:19:05 crc kubenswrapper[4716]: I1207 16:19:05.229779 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d8b488bdf-fvlzm" event={"ID":"d7db6aff-6831-49ea-9d51-24ec9ac40a88","Type":"ContainerStarted","Data":"05f9d64947bebe40459dd781411ba7578a23432192b23ecd871939ba96d606d4"} Dec 07 16:19:05 crc kubenswrapper[4716]: I1207 16:19:05.267231 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-26zm5" podStartSLOduration=4.267208682 podStartE2EDuration="4.267208682s" podCreationTimestamp="2025-12-07 16:19:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:19:05.253500699 +0000 UTC m=+1007.943785611" watchObservedRunningTime="2025-12-07 16:19:05.267208682 +0000 UTC m=+1007.957493594" Dec 07 16:19:05 crc kubenswrapper[4716]: I1207 16:19:05.280479 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-kl4tr" podStartSLOduration=4.280456161 podStartE2EDuration="4.280456161s" podCreationTimestamp="2025-12-07 16:19:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:19:05.267394097 +0000 UTC m=+1007.957679009" watchObservedRunningTime="2025-12-07 16:19:05.280456161 +0000 UTC m=+1007.970741083" Dec 07 16:19:05 crc kubenswrapper[4716]: I1207 16:19:05.561911 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-6djm5" Dec 07 16:19:05 crc kubenswrapper[4716]: I1207 16:19:05.606226 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1af7e22b-2903-4d7a-bb2b-f565bd44d03b-dns-svc\") pod \"1af7e22b-2903-4d7a-bb2b-f565bd44d03b\" (UID: \"1af7e22b-2903-4d7a-bb2b-f565bd44d03b\") " Dec 07 16:19:05 crc kubenswrapper[4716]: I1207 16:19:05.606264 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1af7e22b-2903-4d7a-bb2b-f565bd44d03b-dns-swift-storage-0\") pod \"1af7e22b-2903-4d7a-bb2b-f565bd44d03b\" (UID: \"1af7e22b-2903-4d7a-bb2b-f565bd44d03b\") " Dec 07 16:19:05 crc kubenswrapper[4716]: I1207 16:19:05.606300 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9vrkr\" (UniqueName: \"kubernetes.io/projected/1af7e22b-2903-4d7a-bb2b-f565bd44d03b-kube-api-access-9vrkr\") pod \"1af7e22b-2903-4d7a-bb2b-f565bd44d03b\" (UID: \"1af7e22b-2903-4d7a-bb2b-f565bd44d03b\") " Dec 07 16:19:05 crc kubenswrapper[4716]: I1207 16:19:05.606348 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1af7e22b-2903-4d7a-bb2b-f565bd44d03b-ovsdbserver-sb\") pod \"1af7e22b-2903-4d7a-bb2b-f565bd44d03b\" (UID: \"1af7e22b-2903-4d7a-bb2b-f565bd44d03b\") " Dec 07 16:19:05 crc kubenswrapper[4716]: I1207 16:19:05.606369 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1af7e22b-2903-4d7a-bb2b-f565bd44d03b-config\") pod \"1af7e22b-2903-4d7a-bb2b-f565bd44d03b\" (UID: \"1af7e22b-2903-4d7a-bb2b-f565bd44d03b\") " Dec 07 16:19:05 crc kubenswrapper[4716]: I1207 16:19:05.606407 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1af7e22b-2903-4d7a-bb2b-f565bd44d03b-ovsdbserver-nb\") pod \"1af7e22b-2903-4d7a-bb2b-f565bd44d03b\" (UID: \"1af7e22b-2903-4d7a-bb2b-f565bd44d03b\") " Dec 07 16:19:05 crc kubenswrapper[4716]: I1207 16:19:05.611410 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1af7e22b-2903-4d7a-bb2b-f565bd44d03b-kube-api-access-9vrkr" (OuterVolumeSpecName: "kube-api-access-9vrkr") pod "1af7e22b-2903-4d7a-bb2b-f565bd44d03b" (UID: "1af7e22b-2903-4d7a-bb2b-f565bd44d03b"). InnerVolumeSpecName "kube-api-access-9vrkr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:19:05 crc kubenswrapper[4716]: I1207 16:19:05.635843 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1af7e22b-2903-4d7a-bb2b-f565bd44d03b-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "1af7e22b-2903-4d7a-bb2b-f565bd44d03b" (UID: "1af7e22b-2903-4d7a-bb2b-f565bd44d03b"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:19:05 crc kubenswrapper[4716]: I1207 16:19:05.635883 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1af7e22b-2903-4d7a-bb2b-f565bd44d03b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1af7e22b-2903-4d7a-bb2b-f565bd44d03b" (UID: "1af7e22b-2903-4d7a-bb2b-f565bd44d03b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:19:05 crc kubenswrapper[4716]: I1207 16:19:05.655656 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1af7e22b-2903-4d7a-bb2b-f565bd44d03b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "1af7e22b-2903-4d7a-bb2b-f565bd44d03b" (UID: "1af7e22b-2903-4d7a-bb2b-f565bd44d03b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:19:05 crc kubenswrapper[4716]: I1207 16:19:05.681816 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1af7e22b-2903-4d7a-bb2b-f565bd44d03b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "1af7e22b-2903-4d7a-bb2b-f565bd44d03b" (UID: "1af7e22b-2903-4d7a-bb2b-f565bd44d03b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:19:05 crc kubenswrapper[4716]: I1207 16:19:05.703214 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1af7e22b-2903-4d7a-bb2b-f565bd44d03b-config" (OuterVolumeSpecName: "config") pod "1af7e22b-2903-4d7a-bb2b-f565bd44d03b" (UID: "1af7e22b-2903-4d7a-bb2b-f565bd44d03b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:19:05 crc kubenswrapper[4716]: I1207 16:19:05.708551 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9vrkr\" (UniqueName: \"kubernetes.io/projected/1af7e22b-2903-4d7a-bb2b-f565bd44d03b-kube-api-access-9vrkr\") on node \"crc\" DevicePath \"\"" Dec 07 16:19:05 crc kubenswrapper[4716]: I1207 16:19:05.708592 4716 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1af7e22b-2903-4d7a-bb2b-f565bd44d03b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 07 16:19:05 crc kubenswrapper[4716]: I1207 16:19:05.708606 4716 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1af7e22b-2903-4d7a-bb2b-f565bd44d03b-config\") on node \"crc\" DevicePath \"\"" Dec 07 16:19:05 crc kubenswrapper[4716]: I1207 16:19:05.708620 4716 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1af7e22b-2903-4d7a-bb2b-f565bd44d03b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 07 16:19:05 crc kubenswrapper[4716]: I1207 16:19:05.708632 4716 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1af7e22b-2903-4d7a-bb2b-f565bd44d03b-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 07 16:19:05 crc kubenswrapper[4716]: I1207 16:19:05.708644 4716 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1af7e22b-2903-4d7a-bb2b-f565bd44d03b-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 07 16:19:05 crc kubenswrapper[4716]: I1207 16:19:05.863943 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 07 16:19:06 crc kubenswrapper[4716]: I1207 16:19:06.258270 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-6djm5" event={"ID":"1af7e22b-2903-4d7a-bb2b-f565bd44d03b","Type":"ContainerDied","Data":"e6a709ba91d278cca1171cbca36223dad59798c8aae3bfa0996b30956f5cce94"} Dec 07 16:19:06 crc kubenswrapper[4716]: I1207 16:19:06.258623 4716 scope.go:117] "RemoveContainer" containerID="040363d15e81674e3c171a80e9f51c850ee79a715574ce17e4ebb01b79ee6094" Dec 07 16:19:06 crc kubenswrapper[4716]: I1207 16:19:06.258790 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-6djm5" Dec 07 16:19:06 crc kubenswrapper[4716]: I1207 16:19:06.281331 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"43211ed2-3960-41b4-a4dc-1207a5b0d2bc","Type":"ContainerStarted","Data":"38509b7504f5aabae26ab0ffb9fd3aadcfe7a3b509c4bb131c367e5b8cbf7d3c"} Dec 07 16:19:06 crc kubenswrapper[4716]: I1207 16:19:06.287088 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e","Type":"ContainerStarted","Data":"c84d6e94e9b20a68481533baa58835f93e883fce8a7ca85bb5b0780d31ad9189"} Dec 07 16:19:06 crc kubenswrapper[4716]: I1207 16:19:06.296391 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-cq5v8" event={"ID":"5ced3e75-e271-433b-ad77-fe9101c18be5","Type":"ContainerStarted","Data":"8aaa9cd86647e78761528be99baca4b3a306d7e641067604de34f07fb03d3f7e"} Dec 07 16:19:06 crc kubenswrapper[4716]: I1207 16:19:06.296424 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-785d8bcb8c-cq5v8" Dec 07 16:19:06 crc kubenswrapper[4716]: I1207 16:19:06.337340 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-6djm5"] Dec 07 16:19:06 crc kubenswrapper[4716]: I1207 16:19:06.344524 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-6djm5"] Dec 07 16:19:06 crc kubenswrapper[4716]: I1207 16:19:06.353066 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-785d8bcb8c-cq5v8" podStartSLOduration=4.35304723 podStartE2EDuration="4.35304723s" podCreationTimestamp="2025-12-07 16:19:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:19:06.350398628 +0000 UTC m=+1009.040683540" watchObservedRunningTime="2025-12-07 16:19:06.35304723 +0000 UTC m=+1009.043332142" Dec 07 16:19:07 crc kubenswrapper[4716]: I1207 16:19:07.315716 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e","Type":"ContainerStarted","Data":"7a783303abd3c25533a394e48b039cc8b2b83fbb8687254194f0b6a5eb45156b"} Dec 07 16:19:07 crc kubenswrapper[4716]: I1207 16:19:07.328030 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"43211ed2-3960-41b4-a4dc-1207a5b0d2bc","Type":"ContainerStarted","Data":"83e3fba825b5a95d45367566ff2f290a9309499d1a555060d5b08b31407af7c3"} Dec 07 16:19:07 crc kubenswrapper[4716]: I1207 16:19:07.328378 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="43211ed2-3960-41b4-a4dc-1207a5b0d2bc" containerName="glance-log" containerID="cri-o://38509b7504f5aabae26ab0ffb9fd3aadcfe7a3b509c4bb131c367e5b8cbf7d3c" gracePeriod=30 Dec 07 16:19:07 crc kubenswrapper[4716]: I1207 16:19:07.328496 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="43211ed2-3960-41b4-a4dc-1207a5b0d2bc" containerName="glance-httpd" containerID="cri-o://83e3fba825b5a95d45367566ff2f290a9309499d1a555060d5b08b31407af7c3" gracePeriod=30 Dec 07 16:19:07 crc kubenswrapper[4716]: I1207 16:19:07.362491 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.362469813 podStartE2EDuration="5.362469813s" podCreationTimestamp="2025-12-07 16:19:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:19:07.352145782 +0000 UTC m=+1010.042430694" watchObservedRunningTime="2025-12-07 16:19:07.362469813 +0000 UTC m=+1010.052754725" Dec 07 16:19:07 crc kubenswrapper[4716]: I1207 16:19:07.681430 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1af7e22b-2903-4d7a-bb2b-f565bd44d03b" path="/var/lib/kubelet/pods/1af7e22b-2903-4d7a-bb2b-f565bd44d03b/volumes" Dec 07 16:19:08 crc kubenswrapper[4716]: I1207 16:19:08.350645 4716 generic.go:334] "Generic (PLEG): container finished" podID="43211ed2-3960-41b4-a4dc-1207a5b0d2bc" containerID="83e3fba825b5a95d45367566ff2f290a9309499d1a555060d5b08b31407af7c3" exitCode=0 Dec 07 16:19:08 crc kubenswrapper[4716]: I1207 16:19:08.350670 4716 generic.go:334] "Generic (PLEG): container finished" podID="43211ed2-3960-41b4-a4dc-1207a5b0d2bc" containerID="38509b7504f5aabae26ab0ffb9fd3aadcfe7a3b509c4bb131c367e5b8cbf7d3c" exitCode=143 Dec 07 16:19:08 crc kubenswrapper[4716]: I1207 16:19:08.350727 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"43211ed2-3960-41b4-a4dc-1207a5b0d2bc","Type":"ContainerDied","Data":"83e3fba825b5a95d45367566ff2f290a9309499d1a555060d5b08b31407af7c3"} Dec 07 16:19:08 crc kubenswrapper[4716]: I1207 16:19:08.350785 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"43211ed2-3960-41b4-a4dc-1207a5b0d2bc","Type":"ContainerDied","Data":"38509b7504f5aabae26ab0ffb9fd3aadcfe7a3b509c4bb131c367e5b8cbf7d3c"} Dec 07 16:19:08 crc kubenswrapper[4716]: I1207 16:19:08.354575 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e","Type":"ContainerStarted","Data":"ebf7134443f761827f287b60f8be3332e9509b00b529ef0c8a668ae9d14d372c"} Dec 07 16:19:08 crc kubenswrapper[4716]: I1207 16:19:08.354738 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e" containerName="glance-log" containerID="cri-o://7a783303abd3c25533a394e48b039cc8b2b83fbb8687254194f0b6a5eb45156b" gracePeriod=30 Dec 07 16:19:08 crc kubenswrapper[4716]: I1207 16:19:08.354797 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e" containerName="glance-httpd" containerID="cri-o://ebf7134443f761827f287b60f8be3332e9509b00b529ef0c8a668ae9d14d372c" gracePeriod=30 Dec 07 16:19:08 crc kubenswrapper[4716]: I1207 16:19:08.387088 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=7.387054647 podStartE2EDuration="7.387054647s" podCreationTimestamp="2025-12-07 16:19:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:19:08.376928382 +0000 UTC m=+1011.067213294" watchObservedRunningTime="2025-12-07 16:19:08.387054647 +0000 UTC m=+1011.077339559" Dec 07 16:19:09 crc kubenswrapper[4716]: I1207 16:19:09.410372 4716 generic.go:334] "Generic (PLEG): container finished" podID="0829dcfe-e23e-472a-a01a-d49351ba1f7a" containerID="49e2ed79567e1481a18a12c7c4b8d770a72017bbcff0fed08ff63919176cc886" exitCode=0 Dec 07 16:19:09 crc kubenswrapper[4716]: I1207 16:19:09.410523 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-26zm5" event={"ID":"0829dcfe-e23e-472a-a01a-d49351ba1f7a","Type":"ContainerDied","Data":"49e2ed79567e1481a18a12c7c4b8d770a72017bbcff0fed08ff63919176cc886"} Dec 07 16:19:09 crc kubenswrapper[4716]: I1207 16:19:09.428393 4716 generic.go:334] "Generic (PLEG): container finished" podID="d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e" containerID="ebf7134443f761827f287b60f8be3332e9509b00b529ef0c8a668ae9d14d372c" exitCode=0 Dec 07 16:19:09 crc kubenswrapper[4716]: I1207 16:19:09.428425 4716 generic.go:334] "Generic (PLEG): container finished" podID="d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e" containerID="7a783303abd3c25533a394e48b039cc8b2b83fbb8687254194f0b6a5eb45156b" exitCode=143 Dec 07 16:19:09 crc kubenswrapper[4716]: I1207 16:19:09.428464 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e","Type":"ContainerDied","Data":"ebf7134443f761827f287b60f8be3332e9509b00b529ef0c8a668ae9d14d372c"} Dec 07 16:19:09 crc kubenswrapper[4716]: I1207 16:19:09.428489 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e","Type":"ContainerDied","Data":"7a783303abd3c25533a394e48b039cc8b2b83fbb8687254194f0b6a5eb45156b"} Dec 07 16:19:10 crc kubenswrapper[4716]: I1207 16:19:10.353581 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7d8b488bdf-fvlzm"] Dec 07 16:19:10 crc kubenswrapper[4716]: I1207 16:19:10.390601 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-86c7567d4-99rx9"] Dec 07 16:19:10 crc kubenswrapper[4716]: E1207 16:19:10.390958 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1af7e22b-2903-4d7a-bb2b-f565bd44d03b" containerName="init" Dec 07 16:19:10 crc kubenswrapper[4716]: I1207 16:19:10.390970 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="1af7e22b-2903-4d7a-bb2b-f565bd44d03b" containerName="init" Dec 07 16:19:10 crc kubenswrapper[4716]: I1207 16:19:10.391146 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="1af7e22b-2903-4d7a-bb2b-f565bd44d03b" containerName="init" Dec 07 16:19:10 crc kubenswrapper[4716]: I1207 16:19:10.391998 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-86c7567d4-99rx9" Dec 07 16:19:10 crc kubenswrapper[4716]: I1207 16:19:10.394621 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Dec 07 16:19:10 crc kubenswrapper[4716]: I1207 16:19:10.412664 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-86c7567d4-99rx9"] Dec 07 16:19:10 crc kubenswrapper[4716]: I1207 16:19:10.429230 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7d76599897-4mhqt"] Dec 07 16:19:10 crc kubenswrapper[4716]: I1207 16:19:10.451784 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7f9b78866d-9g9kv"] Dec 07 16:19:10 crc kubenswrapper[4716]: I1207 16:19:10.455819 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7f9b78866d-9g9kv" Dec 07 16:19:10 crc kubenswrapper[4716]: I1207 16:19:10.503428 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c291e114-7940-46bc-91d3-a8fa256549ff-combined-ca-bundle\") pod \"horizon-86c7567d4-99rx9\" (UID: \"c291e114-7940-46bc-91d3-a8fa256549ff\") " pod="openstack/horizon-86c7567d4-99rx9" Dec 07 16:19:10 crc kubenswrapper[4716]: I1207 16:19:10.503724 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8xcrh\" (UniqueName: \"kubernetes.io/projected/c291e114-7940-46bc-91d3-a8fa256549ff-kube-api-access-8xcrh\") pod \"horizon-86c7567d4-99rx9\" (UID: \"c291e114-7940-46bc-91d3-a8fa256549ff\") " pod="openstack/horizon-86c7567d4-99rx9" Dec 07 16:19:10 crc kubenswrapper[4716]: I1207 16:19:10.503748 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c291e114-7940-46bc-91d3-a8fa256549ff-scripts\") pod \"horizon-86c7567d4-99rx9\" (UID: \"c291e114-7940-46bc-91d3-a8fa256549ff\") " pod="openstack/horizon-86c7567d4-99rx9" Dec 07 16:19:10 crc kubenswrapper[4716]: I1207 16:19:10.503767 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c291e114-7940-46bc-91d3-a8fa256549ff-horizon-secret-key\") pod \"horizon-86c7567d4-99rx9\" (UID: \"c291e114-7940-46bc-91d3-a8fa256549ff\") " pod="openstack/horizon-86c7567d4-99rx9" Dec 07 16:19:10 crc kubenswrapper[4716]: I1207 16:19:10.503785 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/c291e114-7940-46bc-91d3-a8fa256549ff-horizon-tls-certs\") pod \"horizon-86c7567d4-99rx9\" (UID: \"c291e114-7940-46bc-91d3-a8fa256549ff\") " pod="openstack/horizon-86c7567d4-99rx9" Dec 07 16:19:10 crc kubenswrapper[4716]: I1207 16:19:10.503862 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c291e114-7940-46bc-91d3-a8fa256549ff-config-data\") pod \"horizon-86c7567d4-99rx9\" (UID: \"c291e114-7940-46bc-91d3-a8fa256549ff\") " pod="openstack/horizon-86c7567d4-99rx9" Dec 07 16:19:10 crc kubenswrapper[4716]: I1207 16:19:10.503891 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c291e114-7940-46bc-91d3-a8fa256549ff-logs\") pod \"horizon-86c7567d4-99rx9\" (UID: \"c291e114-7940-46bc-91d3-a8fa256549ff\") " pod="openstack/horizon-86c7567d4-99rx9" Dec 07 16:19:10 crc kubenswrapper[4716]: I1207 16:19:10.504011 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7f9b78866d-9g9kv"] Dec 07 16:19:10 crc kubenswrapper[4716]: I1207 16:19:10.605196 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c291e114-7940-46bc-91d3-a8fa256549ff-config-data\") pod \"horizon-86c7567d4-99rx9\" (UID: \"c291e114-7940-46bc-91d3-a8fa256549ff\") " pod="openstack/horizon-86c7567d4-99rx9" Dec 07 16:19:10 crc kubenswrapper[4716]: I1207 16:19:10.605250 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c291e114-7940-46bc-91d3-a8fa256549ff-logs\") pod \"horizon-86c7567d4-99rx9\" (UID: \"c291e114-7940-46bc-91d3-a8fa256549ff\") " pod="openstack/horizon-86c7567d4-99rx9" Dec 07 16:19:10 crc kubenswrapper[4716]: I1207 16:19:10.605300 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d845ae0b-5d45-4021-a1e9-4b124298b65b-horizon-secret-key\") pod \"horizon-7f9b78866d-9g9kv\" (UID: \"d845ae0b-5d45-4021-a1e9-4b124298b65b\") " pod="openstack/horizon-7f9b78866d-9g9kv" Dec 07 16:19:10 crc kubenswrapper[4716]: I1207 16:19:10.605320 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c291e114-7940-46bc-91d3-a8fa256549ff-combined-ca-bundle\") pod \"horizon-86c7567d4-99rx9\" (UID: \"c291e114-7940-46bc-91d3-a8fa256549ff\") " pod="openstack/horizon-86c7567d4-99rx9" Dec 07 16:19:10 crc kubenswrapper[4716]: I1207 16:19:10.605341 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d845ae0b-5d45-4021-a1e9-4b124298b65b-combined-ca-bundle\") pod \"horizon-7f9b78866d-9g9kv\" (UID: \"d845ae0b-5d45-4021-a1e9-4b124298b65b\") " pod="openstack/horizon-7f9b78866d-9g9kv" Dec 07 16:19:10 crc kubenswrapper[4716]: I1207 16:19:10.605360 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8xcrh\" (UniqueName: \"kubernetes.io/projected/c291e114-7940-46bc-91d3-a8fa256549ff-kube-api-access-8xcrh\") pod \"horizon-86c7567d4-99rx9\" (UID: \"c291e114-7940-46bc-91d3-a8fa256549ff\") " pod="openstack/horizon-86c7567d4-99rx9" Dec 07 16:19:10 crc kubenswrapper[4716]: I1207 16:19:10.605379 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c291e114-7940-46bc-91d3-a8fa256549ff-scripts\") pod \"horizon-86c7567d4-99rx9\" (UID: \"c291e114-7940-46bc-91d3-a8fa256549ff\") " pod="openstack/horizon-86c7567d4-99rx9" Dec 07 16:19:10 crc kubenswrapper[4716]: I1207 16:19:10.605400 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c291e114-7940-46bc-91d3-a8fa256549ff-horizon-secret-key\") pod \"horizon-86c7567d4-99rx9\" (UID: \"c291e114-7940-46bc-91d3-a8fa256549ff\") " pod="openstack/horizon-86c7567d4-99rx9" Dec 07 16:19:10 crc kubenswrapper[4716]: I1207 16:19:10.605418 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tqjjd\" (UniqueName: \"kubernetes.io/projected/d845ae0b-5d45-4021-a1e9-4b124298b65b-kube-api-access-tqjjd\") pod \"horizon-7f9b78866d-9g9kv\" (UID: \"d845ae0b-5d45-4021-a1e9-4b124298b65b\") " pod="openstack/horizon-7f9b78866d-9g9kv" Dec 07 16:19:10 crc kubenswrapper[4716]: I1207 16:19:10.605435 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/c291e114-7940-46bc-91d3-a8fa256549ff-horizon-tls-certs\") pod \"horizon-86c7567d4-99rx9\" (UID: \"c291e114-7940-46bc-91d3-a8fa256549ff\") " pod="openstack/horizon-86c7567d4-99rx9" Dec 07 16:19:10 crc kubenswrapper[4716]: I1207 16:19:10.605468 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d845ae0b-5d45-4021-a1e9-4b124298b65b-config-data\") pod \"horizon-7f9b78866d-9g9kv\" (UID: \"d845ae0b-5d45-4021-a1e9-4b124298b65b\") " pod="openstack/horizon-7f9b78866d-9g9kv" Dec 07 16:19:10 crc kubenswrapper[4716]: I1207 16:19:10.605486 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/d845ae0b-5d45-4021-a1e9-4b124298b65b-horizon-tls-certs\") pod \"horizon-7f9b78866d-9g9kv\" (UID: \"d845ae0b-5d45-4021-a1e9-4b124298b65b\") " pod="openstack/horizon-7f9b78866d-9g9kv" Dec 07 16:19:10 crc kubenswrapper[4716]: I1207 16:19:10.605509 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d845ae0b-5d45-4021-a1e9-4b124298b65b-scripts\") pod \"horizon-7f9b78866d-9g9kv\" (UID: \"d845ae0b-5d45-4021-a1e9-4b124298b65b\") " pod="openstack/horizon-7f9b78866d-9g9kv" Dec 07 16:19:10 crc kubenswrapper[4716]: I1207 16:19:10.605539 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d845ae0b-5d45-4021-a1e9-4b124298b65b-logs\") pod \"horizon-7f9b78866d-9g9kv\" (UID: \"d845ae0b-5d45-4021-a1e9-4b124298b65b\") " pod="openstack/horizon-7f9b78866d-9g9kv" Dec 07 16:19:10 crc kubenswrapper[4716]: I1207 16:19:10.606692 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c291e114-7940-46bc-91d3-a8fa256549ff-config-data\") pod \"horizon-86c7567d4-99rx9\" (UID: \"c291e114-7940-46bc-91d3-a8fa256549ff\") " pod="openstack/horizon-86c7567d4-99rx9" Dec 07 16:19:10 crc kubenswrapper[4716]: I1207 16:19:10.606905 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c291e114-7940-46bc-91d3-a8fa256549ff-scripts\") pod \"horizon-86c7567d4-99rx9\" (UID: \"c291e114-7940-46bc-91d3-a8fa256549ff\") " pod="openstack/horizon-86c7567d4-99rx9" Dec 07 16:19:10 crc kubenswrapper[4716]: I1207 16:19:10.607248 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c291e114-7940-46bc-91d3-a8fa256549ff-logs\") pod \"horizon-86c7567d4-99rx9\" (UID: \"c291e114-7940-46bc-91d3-a8fa256549ff\") " pod="openstack/horizon-86c7567d4-99rx9" Dec 07 16:19:10 crc kubenswrapper[4716]: I1207 16:19:10.612411 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/c291e114-7940-46bc-91d3-a8fa256549ff-horizon-tls-certs\") pod \"horizon-86c7567d4-99rx9\" (UID: \"c291e114-7940-46bc-91d3-a8fa256549ff\") " pod="openstack/horizon-86c7567d4-99rx9" Dec 07 16:19:10 crc kubenswrapper[4716]: I1207 16:19:10.616398 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c291e114-7940-46bc-91d3-a8fa256549ff-combined-ca-bundle\") pod \"horizon-86c7567d4-99rx9\" (UID: \"c291e114-7940-46bc-91d3-a8fa256549ff\") " pod="openstack/horizon-86c7567d4-99rx9" Dec 07 16:19:10 crc kubenswrapper[4716]: I1207 16:19:10.622906 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c291e114-7940-46bc-91d3-a8fa256549ff-horizon-secret-key\") pod \"horizon-86c7567d4-99rx9\" (UID: \"c291e114-7940-46bc-91d3-a8fa256549ff\") " pod="openstack/horizon-86c7567d4-99rx9" Dec 07 16:19:10 crc kubenswrapper[4716]: I1207 16:19:10.634536 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8xcrh\" (UniqueName: \"kubernetes.io/projected/c291e114-7940-46bc-91d3-a8fa256549ff-kube-api-access-8xcrh\") pod \"horizon-86c7567d4-99rx9\" (UID: \"c291e114-7940-46bc-91d3-a8fa256549ff\") " pod="openstack/horizon-86c7567d4-99rx9" Dec 07 16:19:10 crc kubenswrapper[4716]: I1207 16:19:10.706746 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d845ae0b-5d45-4021-a1e9-4b124298b65b-combined-ca-bundle\") pod \"horizon-7f9b78866d-9g9kv\" (UID: \"d845ae0b-5d45-4021-a1e9-4b124298b65b\") " pod="openstack/horizon-7f9b78866d-9g9kv" Dec 07 16:19:10 crc kubenswrapper[4716]: I1207 16:19:10.706802 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tqjjd\" (UniqueName: \"kubernetes.io/projected/d845ae0b-5d45-4021-a1e9-4b124298b65b-kube-api-access-tqjjd\") pod \"horizon-7f9b78866d-9g9kv\" (UID: \"d845ae0b-5d45-4021-a1e9-4b124298b65b\") " pod="openstack/horizon-7f9b78866d-9g9kv" Dec 07 16:19:10 crc kubenswrapper[4716]: I1207 16:19:10.706840 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d845ae0b-5d45-4021-a1e9-4b124298b65b-config-data\") pod \"horizon-7f9b78866d-9g9kv\" (UID: \"d845ae0b-5d45-4021-a1e9-4b124298b65b\") " pod="openstack/horizon-7f9b78866d-9g9kv" Dec 07 16:19:10 crc kubenswrapper[4716]: I1207 16:19:10.706859 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/d845ae0b-5d45-4021-a1e9-4b124298b65b-horizon-tls-certs\") pod \"horizon-7f9b78866d-9g9kv\" (UID: \"d845ae0b-5d45-4021-a1e9-4b124298b65b\") " pod="openstack/horizon-7f9b78866d-9g9kv" Dec 07 16:19:10 crc kubenswrapper[4716]: I1207 16:19:10.706882 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d845ae0b-5d45-4021-a1e9-4b124298b65b-scripts\") pod \"horizon-7f9b78866d-9g9kv\" (UID: \"d845ae0b-5d45-4021-a1e9-4b124298b65b\") " pod="openstack/horizon-7f9b78866d-9g9kv" Dec 07 16:19:10 crc kubenswrapper[4716]: I1207 16:19:10.706911 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d845ae0b-5d45-4021-a1e9-4b124298b65b-logs\") pod \"horizon-7f9b78866d-9g9kv\" (UID: \"d845ae0b-5d45-4021-a1e9-4b124298b65b\") " pod="openstack/horizon-7f9b78866d-9g9kv" Dec 07 16:19:10 crc kubenswrapper[4716]: I1207 16:19:10.706982 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d845ae0b-5d45-4021-a1e9-4b124298b65b-horizon-secret-key\") pod \"horizon-7f9b78866d-9g9kv\" (UID: \"d845ae0b-5d45-4021-a1e9-4b124298b65b\") " pod="openstack/horizon-7f9b78866d-9g9kv" Dec 07 16:19:10 crc kubenswrapper[4716]: I1207 16:19:10.708391 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d845ae0b-5d45-4021-a1e9-4b124298b65b-config-data\") pod \"horizon-7f9b78866d-9g9kv\" (UID: \"d845ae0b-5d45-4021-a1e9-4b124298b65b\") " pod="openstack/horizon-7f9b78866d-9g9kv" Dec 07 16:19:10 crc kubenswrapper[4716]: I1207 16:19:10.708703 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d845ae0b-5d45-4021-a1e9-4b124298b65b-logs\") pod \"horizon-7f9b78866d-9g9kv\" (UID: \"d845ae0b-5d45-4021-a1e9-4b124298b65b\") " pod="openstack/horizon-7f9b78866d-9g9kv" Dec 07 16:19:10 crc kubenswrapper[4716]: I1207 16:19:10.708870 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d845ae0b-5d45-4021-a1e9-4b124298b65b-scripts\") pod \"horizon-7f9b78866d-9g9kv\" (UID: \"d845ae0b-5d45-4021-a1e9-4b124298b65b\") " pod="openstack/horizon-7f9b78866d-9g9kv" Dec 07 16:19:10 crc kubenswrapper[4716]: I1207 16:19:10.716503 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d845ae0b-5d45-4021-a1e9-4b124298b65b-horizon-secret-key\") pod \"horizon-7f9b78866d-9g9kv\" (UID: \"d845ae0b-5d45-4021-a1e9-4b124298b65b\") " pod="openstack/horizon-7f9b78866d-9g9kv" Dec 07 16:19:10 crc kubenswrapper[4716]: I1207 16:19:10.716548 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d845ae0b-5d45-4021-a1e9-4b124298b65b-combined-ca-bundle\") pod \"horizon-7f9b78866d-9g9kv\" (UID: \"d845ae0b-5d45-4021-a1e9-4b124298b65b\") " pod="openstack/horizon-7f9b78866d-9g9kv" Dec 07 16:19:10 crc kubenswrapper[4716]: I1207 16:19:10.716793 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-86c7567d4-99rx9" Dec 07 16:19:10 crc kubenswrapper[4716]: I1207 16:19:10.716898 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/d845ae0b-5d45-4021-a1e9-4b124298b65b-horizon-tls-certs\") pod \"horizon-7f9b78866d-9g9kv\" (UID: \"d845ae0b-5d45-4021-a1e9-4b124298b65b\") " pod="openstack/horizon-7f9b78866d-9g9kv" Dec 07 16:19:10 crc kubenswrapper[4716]: I1207 16:19:10.723057 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tqjjd\" (UniqueName: \"kubernetes.io/projected/d845ae0b-5d45-4021-a1e9-4b124298b65b-kube-api-access-tqjjd\") pod \"horizon-7f9b78866d-9g9kv\" (UID: \"d845ae0b-5d45-4021-a1e9-4b124298b65b\") " pod="openstack/horizon-7f9b78866d-9g9kv" Dec 07 16:19:10 crc kubenswrapper[4716]: I1207 16:19:10.787286 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7f9b78866d-9g9kv" Dec 07 16:19:12 crc kubenswrapper[4716]: I1207 16:19:12.658320 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-785d8bcb8c-cq5v8" Dec 07 16:19:12 crc kubenswrapper[4716]: I1207 16:19:12.716818 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-2cv95"] Dec 07 16:19:12 crc kubenswrapper[4716]: I1207 16:19:12.717206 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-74f6bcbc87-2cv95" podUID="7a57dd80-9afb-4d22-bd71-7104bc574ab0" containerName="dnsmasq-dns" containerID="cri-o://670d21ca07fde13d32e5cc95349a99ca26fd9a0580fa5a23a4f056dd9038b973" gracePeriod=10 Dec 07 16:19:14 crc kubenswrapper[4716]: I1207 16:19:14.493871 4716 generic.go:334] "Generic (PLEG): container finished" podID="7a57dd80-9afb-4d22-bd71-7104bc574ab0" containerID="670d21ca07fde13d32e5cc95349a99ca26fd9a0580fa5a23a4f056dd9038b973" exitCode=0 Dec 07 16:19:14 crc kubenswrapper[4716]: I1207 16:19:14.493944 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-2cv95" event={"ID":"7a57dd80-9afb-4d22-bd71-7104bc574ab0","Type":"ContainerDied","Data":"670d21ca07fde13d32e5cc95349a99ca26fd9a0580fa5a23a4f056dd9038b973"} Dec 07 16:19:15 crc kubenswrapper[4716]: I1207 16:19:15.479979 4716 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-74f6bcbc87-2cv95" podUID="7a57dd80-9afb-4d22-bd71-7104bc574ab0" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.126:5353: connect: connection refused" Dec 07 16:19:16 crc kubenswrapper[4716]: I1207 16:19:16.949353 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 07 16:19:16 crc kubenswrapper[4716]: I1207 16:19:16.958932 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 07 16:19:16 crc kubenswrapper[4716]: I1207 16:19:16.966478 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-26zm5" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.148514 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e-httpd-run\") pod \"d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e\" (UID: \"d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e\") " Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.148610 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/43211ed2-3960-41b4-a4dc-1207a5b0d2bc-public-tls-certs\") pod \"43211ed2-3960-41b4-a4dc-1207a5b0d2bc\" (UID: \"43211ed2-3960-41b4-a4dc-1207a5b0d2bc\") " Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.148633 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/43211ed2-3960-41b4-a4dc-1207a5b0d2bc-httpd-run\") pod \"43211ed2-3960-41b4-a4dc-1207a5b0d2bc\" (UID: \"43211ed2-3960-41b4-a4dc-1207a5b0d2bc\") " Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.148650 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/43211ed2-3960-41b4-a4dc-1207a5b0d2bc-scripts\") pod \"43211ed2-3960-41b4-a4dc-1207a5b0d2bc\" (UID: \"43211ed2-3960-41b4-a4dc-1207a5b0d2bc\") " Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.148672 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0829dcfe-e23e-472a-a01a-d49351ba1f7a-fernet-keys\") pod \"0829dcfe-e23e-472a-a01a-d49351ba1f7a\" (UID: \"0829dcfe-e23e-472a-a01a-d49351ba1f7a\") " Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.148700 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/43211ed2-3960-41b4-a4dc-1207a5b0d2bc-logs\") pod \"43211ed2-3960-41b4-a4dc-1207a5b0d2bc\" (UID: \"43211ed2-3960-41b4-a4dc-1207a5b0d2bc\") " Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.148742 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e-logs\") pod \"d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e\" (UID: \"d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e\") " Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.148757 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0829dcfe-e23e-472a-a01a-d49351ba1f7a-config-data\") pod \"0829dcfe-e23e-472a-a01a-d49351ba1f7a\" (UID: \"0829dcfe-e23e-472a-a01a-d49351ba1f7a\") " Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.148804 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0829dcfe-e23e-472a-a01a-d49351ba1f7a-scripts\") pod \"0829dcfe-e23e-472a-a01a-d49351ba1f7a\" (UID: \"0829dcfe-e23e-472a-a01a-d49351ba1f7a\") " Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.148839 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-slrzp\" (UniqueName: \"kubernetes.io/projected/0829dcfe-e23e-472a-a01a-d49351ba1f7a-kube-api-access-slrzp\") pod \"0829dcfe-e23e-472a-a01a-d49351ba1f7a\" (UID: \"0829dcfe-e23e-472a-a01a-d49351ba1f7a\") " Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.148878 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"43211ed2-3960-41b4-a4dc-1207a5b0d2bc\" (UID: \"43211ed2-3960-41b4-a4dc-1207a5b0d2bc\") " Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.148894 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e\" (UID: \"d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e\") " Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.148932 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e-config-data\") pod \"d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e\" (UID: \"d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e\") " Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.148947 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e-internal-tls-certs\") pod \"d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e\" (UID: \"d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e\") " Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.148964 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0829dcfe-e23e-472a-a01a-d49351ba1f7a-combined-ca-bundle\") pod \"0829dcfe-e23e-472a-a01a-d49351ba1f7a\" (UID: \"0829dcfe-e23e-472a-a01a-d49351ba1f7a\") " Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.148983 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2g9tv\" (UniqueName: \"kubernetes.io/projected/43211ed2-3960-41b4-a4dc-1207a5b0d2bc-kube-api-access-2g9tv\") pod \"43211ed2-3960-41b4-a4dc-1207a5b0d2bc\" (UID: \"43211ed2-3960-41b4-a4dc-1207a5b0d2bc\") " Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.149003 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43211ed2-3960-41b4-a4dc-1207a5b0d2bc-combined-ca-bundle\") pod \"43211ed2-3960-41b4-a4dc-1207a5b0d2bc\" (UID: \"43211ed2-3960-41b4-a4dc-1207a5b0d2bc\") " Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.149023 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbf6q\" (UniqueName: \"kubernetes.io/projected/d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e-kube-api-access-dbf6q\") pod \"d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e\" (UID: \"d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e\") " Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.149047 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0829dcfe-e23e-472a-a01a-d49351ba1f7a-credential-keys\") pod \"0829dcfe-e23e-472a-a01a-d49351ba1f7a\" (UID: \"0829dcfe-e23e-472a-a01a-d49351ba1f7a\") " Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.149064 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e-scripts\") pod \"d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e\" (UID: \"d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e\") " Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.149109 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43211ed2-3960-41b4-a4dc-1207a5b0d2bc-config-data\") pod \"43211ed2-3960-41b4-a4dc-1207a5b0d2bc\" (UID: \"43211ed2-3960-41b4-a4dc-1207a5b0d2bc\") " Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.149125 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e-combined-ca-bundle\") pod \"d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e\" (UID: \"d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e\") " Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.149167 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/43211ed2-3960-41b4-a4dc-1207a5b0d2bc-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "43211ed2-3960-41b4-a4dc-1207a5b0d2bc" (UID: "43211ed2-3960-41b4-a4dc-1207a5b0d2bc"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.149167 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e" (UID: "d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.149668 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e-logs" (OuterVolumeSpecName: "logs") pod "d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e" (UID: "d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.149669 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/43211ed2-3960-41b4-a4dc-1207a5b0d2bc-logs" (OuterVolumeSpecName: "logs") pod "43211ed2-3960-41b4-a4dc-1207a5b0d2bc" (UID: "43211ed2-3960-41b4-a4dc-1207a5b0d2bc"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.152308 4716 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.152340 4716 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/43211ed2-3960-41b4-a4dc-1207a5b0d2bc-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.158536 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e-scripts" (OuterVolumeSpecName: "scripts") pod "d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e" (UID: "d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.158552 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0829dcfe-e23e-472a-a01a-d49351ba1f7a-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "0829dcfe-e23e-472a-a01a-d49351ba1f7a" (UID: "0829dcfe-e23e-472a-a01a-d49351ba1f7a"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.158619 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0829dcfe-e23e-472a-a01a-d49351ba1f7a-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "0829dcfe-e23e-472a-a01a-d49351ba1f7a" (UID: "0829dcfe-e23e-472a-a01a-d49351ba1f7a"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.158657 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0829dcfe-e23e-472a-a01a-d49351ba1f7a-kube-api-access-slrzp" (OuterVolumeSpecName: "kube-api-access-slrzp") pod "0829dcfe-e23e-472a-a01a-d49351ba1f7a" (UID: "0829dcfe-e23e-472a-a01a-d49351ba1f7a"). InnerVolumeSpecName "kube-api-access-slrzp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.158670 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43211ed2-3960-41b4-a4dc-1207a5b0d2bc-scripts" (OuterVolumeSpecName: "scripts") pod "43211ed2-3960-41b4-a4dc-1207a5b0d2bc" (UID: "43211ed2-3960-41b4-a4dc-1207a5b0d2bc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.158714 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance") pod "43211ed2-3960-41b4-a4dc-1207a5b0d2bc" (UID: "43211ed2-3960-41b4-a4dc-1207a5b0d2bc"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.175490 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e" (UID: "d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.175561 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43211ed2-3960-41b4-a4dc-1207a5b0d2bc-kube-api-access-2g9tv" (OuterVolumeSpecName: "kube-api-access-2g9tv") pod "43211ed2-3960-41b4-a4dc-1207a5b0d2bc" (UID: "43211ed2-3960-41b4-a4dc-1207a5b0d2bc"). InnerVolumeSpecName "kube-api-access-2g9tv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.175590 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e-kube-api-access-dbf6q" (OuterVolumeSpecName: "kube-api-access-dbf6q") pod "d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e" (UID: "d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e"). InnerVolumeSpecName "kube-api-access-dbf6q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.189638 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0829dcfe-e23e-472a-a01a-d49351ba1f7a-scripts" (OuterVolumeSpecName: "scripts") pod "0829dcfe-e23e-472a-a01a-d49351ba1f7a" (UID: "0829dcfe-e23e-472a-a01a-d49351ba1f7a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.194906 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43211ed2-3960-41b4-a4dc-1207a5b0d2bc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "43211ed2-3960-41b4-a4dc-1207a5b0d2bc" (UID: "43211ed2-3960-41b4-a4dc-1207a5b0d2bc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.194970 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0829dcfe-e23e-472a-a01a-d49351ba1f7a-config-data" (OuterVolumeSpecName: "config-data") pod "0829dcfe-e23e-472a-a01a-d49351ba1f7a" (UID: "0829dcfe-e23e-472a-a01a-d49351ba1f7a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.197425 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e" (UID: "d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.215418 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43211ed2-3960-41b4-a4dc-1207a5b0d2bc-config-data" (OuterVolumeSpecName: "config-data") pod "43211ed2-3960-41b4-a4dc-1207a5b0d2bc" (UID: "43211ed2-3960-41b4-a4dc-1207a5b0d2bc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.217415 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e" (UID: "d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.224106 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0829dcfe-e23e-472a-a01a-d49351ba1f7a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0829dcfe-e23e-472a-a01a-d49351ba1f7a" (UID: "0829dcfe-e23e-472a-a01a-d49351ba1f7a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.232375 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e-config-data" (OuterVolumeSpecName: "config-data") pod "d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e" (UID: "d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.241791 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43211ed2-3960-41b4-a4dc-1207a5b0d2bc-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "43211ed2-3960-41b4-a4dc-1207a5b0d2bc" (UID: "43211ed2-3960-41b4-a4dc-1207a5b0d2bc"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.253162 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-slrzp\" (UniqueName: \"kubernetes.io/projected/0829dcfe-e23e-472a-a01a-d49351ba1f7a-kube-api-access-slrzp\") on node \"crc\" DevicePath \"\"" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.253226 4716 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.253240 4716 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.253249 4716 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.253258 4716 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.253267 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2g9tv\" (UniqueName: \"kubernetes.io/projected/43211ed2-3960-41b4-a4dc-1207a5b0d2bc-kube-api-access-2g9tv\") on node \"crc\" DevicePath \"\"" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.253277 4716 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0829dcfe-e23e-472a-a01a-d49351ba1f7a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.253286 4716 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43211ed2-3960-41b4-a4dc-1207a5b0d2bc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.253296 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbf6q\" (UniqueName: \"kubernetes.io/projected/d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e-kube-api-access-dbf6q\") on node \"crc\" DevicePath \"\"" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.253324 4716 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0829dcfe-e23e-472a-a01a-d49351ba1f7a-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.253334 4716 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.253345 4716 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.253356 4716 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43211ed2-3960-41b4-a4dc-1207a5b0d2bc-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.253368 4716 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/43211ed2-3960-41b4-a4dc-1207a5b0d2bc-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.253378 4716 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/43211ed2-3960-41b4-a4dc-1207a5b0d2bc-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.253388 4716 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0829dcfe-e23e-472a-a01a-d49351ba1f7a-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.253398 4716 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/43211ed2-3960-41b4-a4dc-1207a5b0d2bc-logs\") on node \"crc\" DevicePath \"\"" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.253408 4716 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e-logs\") on node \"crc\" DevicePath \"\"" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.253418 4716 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0829dcfe-e23e-472a-a01a-d49351ba1f7a-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.253428 4716 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0829dcfe-e23e-472a-a01a-d49351ba1f7a-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.271133 4716 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.274187 4716 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.354597 4716 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.354631 4716 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.527661 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e","Type":"ContainerDied","Data":"c84d6e94e9b20a68481533baa58835f93e883fce8a7ca85bb5b0780d31ad9189"} Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.527716 4716 scope.go:117] "RemoveContainer" containerID="ebf7134443f761827f287b60f8be3332e9509b00b529ef0c8a668ae9d14d372c" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.527723 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.530372 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-26zm5" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.530386 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-26zm5" event={"ID":"0829dcfe-e23e-472a-a01a-d49351ba1f7a","Type":"ContainerDied","Data":"3f4fc79efa766d7ca00151321768a20ad163a8bdafe72fedbcfd442ddf7c6ba0"} Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.530422 4716 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3f4fc79efa766d7ca00151321768a20ad163a8bdafe72fedbcfd442ddf7c6ba0" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.534988 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"43211ed2-3960-41b4-a4dc-1207a5b0d2bc","Type":"ContainerDied","Data":"1ca9f64e2db9a3b8d9290607ee0ff9ee2d5c2d77aabdf1b4a71036e5b4781802"} Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.535068 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.594895 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.609350 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.627447 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.634732 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.647468 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 07 16:19:17 crc kubenswrapper[4716]: E1207 16:19:17.647884 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e" containerName="glance-httpd" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.647899 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e" containerName="glance-httpd" Dec 07 16:19:17 crc kubenswrapper[4716]: E1207 16:19:17.647908 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0829dcfe-e23e-472a-a01a-d49351ba1f7a" containerName="keystone-bootstrap" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.647915 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="0829dcfe-e23e-472a-a01a-d49351ba1f7a" containerName="keystone-bootstrap" Dec 07 16:19:17 crc kubenswrapper[4716]: E1207 16:19:17.647926 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43211ed2-3960-41b4-a4dc-1207a5b0d2bc" containerName="glance-log" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.647932 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="43211ed2-3960-41b4-a4dc-1207a5b0d2bc" containerName="glance-log" Dec 07 16:19:17 crc kubenswrapper[4716]: E1207 16:19:17.647942 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43211ed2-3960-41b4-a4dc-1207a5b0d2bc" containerName="glance-httpd" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.647948 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="43211ed2-3960-41b4-a4dc-1207a5b0d2bc" containerName="glance-httpd" Dec 07 16:19:17 crc kubenswrapper[4716]: E1207 16:19:17.647968 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e" containerName="glance-log" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.647974 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e" containerName="glance-log" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.648159 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e" containerName="glance-log" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.648178 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="43211ed2-3960-41b4-a4dc-1207a5b0d2bc" containerName="glance-httpd" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.648190 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e" containerName="glance-httpd" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.648206 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="0829dcfe-e23e-472a-a01a-d49351ba1f7a" containerName="keystone-bootstrap" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.648214 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="43211ed2-3960-41b4-a4dc-1207a5b0d2bc" containerName="glance-log" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.649134 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.652006 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.652927 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.653040 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.653069 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-2hfps" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.656291 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.658954 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.661009 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.665224 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.709942 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43211ed2-3960-41b4-a4dc-1207a5b0d2bc" path="/var/lib/kubelet/pods/43211ed2-3960-41b4-a4dc-1207a5b0d2bc/volumes" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.711095 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e" path="/var/lib/kubelet/pods/d6e1c7d8-3d2d-4ce5-814a-cd302e34f50e/volumes" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.713677 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.713720 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.762422 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5eca408c-c86b-4dd6-a1f9-81ef16e12da4-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5eca408c-c86b-4dd6-a1f9-81ef16e12da4\") " pod="openstack/glance-default-internal-api-0" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.762532 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/afba3ceb-86bf-4aaf-8ac3-b58ab411850c-logs\") pod \"glance-default-external-api-0\" (UID: \"afba3ceb-86bf-4aaf-8ac3-b58ab411850c\") " pod="openstack/glance-default-external-api-0" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.762772 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gcgn6\" (UniqueName: \"kubernetes.io/projected/afba3ceb-86bf-4aaf-8ac3-b58ab411850c-kube-api-access-gcgn6\") pod \"glance-default-external-api-0\" (UID: \"afba3ceb-86bf-4aaf-8ac3-b58ab411850c\") " pod="openstack/glance-default-external-api-0" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.762867 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/afba3ceb-86bf-4aaf-8ac3-b58ab411850c-config-data\") pod \"glance-default-external-api-0\" (UID: \"afba3ceb-86bf-4aaf-8ac3-b58ab411850c\") " pod="openstack/glance-default-external-api-0" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.762916 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/afba3ceb-86bf-4aaf-8ac3-b58ab411850c-scripts\") pod \"glance-default-external-api-0\" (UID: \"afba3ceb-86bf-4aaf-8ac3-b58ab411850c\") " pod="openstack/glance-default-external-api-0" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.762942 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/afba3ceb-86bf-4aaf-8ac3-b58ab411850c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"afba3ceb-86bf-4aaf-8ac3-b58ab411850c\") " pod="openstack/glance-default-external-api-0" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.763015 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hnll7\" (UniqueName: \"kubernetes.io/projected/5eca408c-c86b-4dd6-a1f9-81ef16e12da4-kube-api-access-hnll7\") pod \"glance-default-internal-api-0\" (UID: \"5eca408c-c86b-4dd6-a1f9-81ef16e12da4\") " pod="openstack/glance-default-internal-api-0" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.763093 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afba3ceb-86bf-4aaf-8ac3-b58ab411850c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"afba3ceb-86bf-4aaf-8ac3-b58ab411850c\") " pod="openstack/glance-default-external-api-0" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.763125 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"afba3ceb-86bf-4aaf-8ac3-b58ab411850c\") " pod="openstack/glance-default-external-api-0" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.763143 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5eca408c-c86b-4dd6-a1f9-81ef16e12da4-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5eca408c-c86b-4dd6-a1f9-81ef16e12da4\") " pod="openstack/glance-default-internal-api-0" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.763163 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5eca408c-c86b-4dd6-a1f9-81ef16e12da4-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"5eca408c-c86b-4dd6-a1f9-81ef16e12da4\") " pod="openstack/glance-default-internal-api-0" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.763180 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5eca408c-c86b-4dd6-a1f9-81ef16e12da4-logs\") pod \"glance-default-internal-api-0\" (UID: \"5eca408c-c86b-4dd6-a1f9-81ef16e12da4\") " pod="openstack/glance-default-internal-api-0" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.763261 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5eca408c-c86b-4dd6-a1f9-81ef16e12da4-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5eca408c-c86b-4dd6-a1f9-81ef16e12da4\") " pod="openstack/glance-default-internal-api-0" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.763280 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/afba3ceb-86bf-4aaf-8ac3-b58ab411850c-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"afba3ceb-86bf-4aaf-8ac3-b58ab411850c\") " pod="openstack/glance-default-external-api-0" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.763293 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"5eca408c-c86b-4dd6-a1f9-81ef16e12da4\") " pod="openstack/glance-default-internal-api-0" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.763342 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5eca408c-c86b-4dd6-a1f9-81ef16e12da4-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5eca408c-c86b-4dd6-a1f9-81ef16e12da4\") " pod="openstack/glance-default-internal-api-0" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.865327 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/afba3ceb-86bf-4aaf-8ac3-b58ab411850c-logs\") pod \"glance-default-external-api-0\" (UID: \"afba3ceb-86bf-4aaf-8ac3-b58ab411850c\") " pod="openstack/glance-default-external-api-0" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.865381 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gcgn6\" (UniqueName: \"kubernetes.io/projected/afba3ceb-86bf-4aaf-8ac3-b58ab411850c-kube-api-access-gcgn6\") pod \"glance-default-external-api-0\" (UID: \"afba3ceb-86bf-4aaf-8ac3-b58ab411850c\") " pod="openstack/glance-default-external-api-0" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.865417 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/afba3ceb-86bf-4aaf-8ac3-b58ab411850c-config-data\") pod \"glance-default-external-api-0\" (UID: \"afba3ceb-86bf-4aaf-8ac3-b58ab411850c\") " pod="openstack/glance-default-external-api-0" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.865435 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/afba3ceb-86bf-4aaf-8ac3-b58ab411850c-scripts\") pod \"glance-default-external-api-0\" (UID: \"afba3ceb-86bf-4aaf-8ac3-b58ab411850c\") " pod="openstack/glance-default-external-api-0" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.865458 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/afba3ceb-86bf-4aaf-8ac3-b58ab411850c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"afba3ceb-86bf-4aaf-8ac3-b58ab411850c\") " pod="openstack/glance-default-external-api-0" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.865501 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hnll7\" (UniqueName: \"kubernetes.io/projected/5eca408c-c86b-4dd6-a1f9-81ef16e12da4-kube-api-access-hnll7\") pod \"glance-default-internal-api-0\" (UID: \"5eca408c-c86b-4dd6-a1f9-81ef16e12da4\") " pod="openstack/glance-default-internal-api-0" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.865525 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afba3ceb-86bf-4aaf-8ac3-b58ab411850c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"afba3ceb-86bf-4aaf-8ac3-b58ab411850c\") " pod="openstack/glance-default-external-api-0" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.865550 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"afba3ceb-86bf-4aaf-8ac3-b58ab411850c\") " pod="openstack/glance-default-external-api-0" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.865563 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5eca408c-c86b-4dd6-a1f9-81ef16e12da4-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5eca408c-c86b-4dd6-a1f9-81ef16e12da4\") " pod="openstack/glance-default-internal-api-0" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.865580 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5eca408c-c86b-4dd6-a1f9-81ef16e12da4-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"5eca408c-c86b-4dd6-a1f9-81ef16e12da4\") " pod="openstack/glance-default-internal-api-0" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.865593 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5eca408c-c86b-4dd6-a1f9-81ef16e12da4-logs\") pod \"glance-default-internal-api-0\" (UID: \"5eca408c-c86b-4dd6-a1f9-81ef16e12da4\") " pod="openstack/glance-default-internal-api-0" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.865611 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5eca408c-c86b-4dd6-a1f9-81ef16e12da4-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5eca408c-c86b-4dd6-a1f9-81ef16e12da4\") " pod="openstack/glance-default-internal-api-0" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.865627 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/afba3ceb-86bf-4aaf-8ac3-b58ab411850c-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"afba3ceb-86bf-4aaf-8ac3-b58ab411850c\") " pod="openstack/glance-default-external-api-0" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.865644 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"5eca408c-c86b-4dd6-a1f9-81ef16e12da4\") " pod="openstack/glance-default-internal-api-0" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.865660 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5eca408c-c86b-4dd6-a1f9-81ef16e12da4-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5eca408c-c86b-4dd6-a1f9-81ef16e12da4\") " pod="openstack/glance-default-internal-api-0" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.865684 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5eca408c-c86b-4dd6-a1f9-81ef16e12da4-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5eca408c-c86b-4dd6-a1f9-81ef16e12da4\") " pod="openstack/glance-default-internal-api-0" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.866047 4716 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"afba3ceb-86bf-4aaf-8ac3-b58ab411850c\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-external-api-0" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.866392 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5eca408c-c86b-4dd6-a1f9-81ef16e12da4-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5eca408c-c86b-4dd6-a1f9-81ef16e12da4\") " pod="openstack/glance-default-internal-api-0" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.866475 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/afba3ceb-86bf-4aaf-8ac3-b58ab411850c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"afba3ceb-86bf-4aaf-8ac3-b58ab411850c\") " pod="openstack/glance-default-external-api-0" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.866660 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5eca408c-c86b-4dd6-a1f9-81ef16e12da4-logs\") pod \"glance-default-internal-api-0\" (UID: \"5eca408c-c86b-4dd6-a1f9-81ef16e12da4\") " pod="openstack/glance-default-internal-api-0" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.866902 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/afba3ceb-86bf-4aaf-8ac3-b58ab411850c-logs\") pod \"glance-default-external-api-0\" (UID: \"afba3ceb-86bf-4aaf-8ac3-b58ab411850c\") " pod="openstack/glance-default-external-api-0" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.867224 4716 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"5eca408c-c86b-4dd6-a1f9-81ef16e12da4\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-internal-api-0" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.873329 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5eca408c-c86b-4dd6-a1f9-81ef16e12da4-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"5eca408c-c86b-4dd6-a1f9-81ef16e12da4\") " pod="openstack/glance-default-internal-api-0" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.875640 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/afba3ceb-86bf-4aaf-8ac3-b58ab411850c-scripts\") pod \"glance-default-external-api-0\" (UID: \"afba3ceb-86bf-4aaf-8ac3-b58ab411850c\") " pod="openstack/glance-default-external-api-0" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.876049 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5eca408c-c86b-4dd6-a1f9-81ef16e12da4-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5eca408c-c86b-4dd6-a1f9-81ef16e12da4\") " pod="openstack/glance-default-internal-api-0" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.876207 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5eca408c-c86b-4dd6-a1f9-81ef16e12da4-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5eca408c-c86b-4dd6-a1f9-81ef16e12da4\") " pod="openstack/glance-default-internal-api-0" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.879197 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afba3ceb-86bf-4aaf-8ac3-b58ab411850c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"afba3ceb-86bf-4aaf-8ac3-b58ab411850c\") " pod="openstack/glance-default-external-api-0" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.883795 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5eca408c-c86b-4dd6-a1f9-81ef16e12da4-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5eca408c-c86b-4dd6-a1f9-81ef16e12da4\") " pod="openstack/glance-default-internal-api-0" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.884726 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/afba3ceb-86bf-4aaf-8ac3-b58ab411850c-config-data\") pod \"glance-default-external-api-0\" (UID: \"afba3ceb-86bf-4aaf-8ac3-b58ab411850c\") " pod="openstack/glance-default-external-api-0" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.889600 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/afba3ceb-86bf-4aaf-8ac3-b58ab411850c-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"afba3ceb-86bf-4aaf-8ac3-b58ab411850c\") " pod="openstack/glance-default-external-api-0" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.896559 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gcgn6\" (UniqueName: \"kubernetes.io/projected/afba3ceb-86bf-4aaf-8ac3-b58ab411850c-kube-api-access-gcgn6\") pod \"glance-default-external-api-0\" (UID: \"afba3ceb-86bf-4aaf-8ac3-b58ab411850c\") " pod="openstack/glance-default-external-api-0" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.897217 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hnll7\" (UniqueName: \"kubernetes.io/projected/5eca408c-c86b-4dd6-a1f9-81ef16e12da4-kube-api-access-hnll7\") pod \"glance-default-internal-api-0\" (UID: \"5eca408c-c86b-4dd6-a1f9-81ef16e12da4\") " pod="openstack/glance-default-internal-api-0" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.900534 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"5eca408c-c86b-4dd6-a1f9-81ef16e12da4\") " pod="openstack/glance-default-internal-api-0" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.903196 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"afba3ceb-86bf-4aaf-8ac3-b58ab411850c\") " pod="openstack/glance-default-external-api-0" Dec 07 16:19:17 crc kubenswrapper[4716]: I1207 16:19:17.975318 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 07 16:19:18 crc kubenswrapper[4716]: I1207 16:19:18.016314 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 07 16:19:18 crc kubenswrapper[4716]: I1207 16:19:18.131593 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-26zm5"] Dec 07 16:19:18 crc kubenswrapper[4716]: I1207 16:19:18.139751 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-26zm5"] Dec 07 16:19:18 crc kubenswrapper[4716]: I1207 16:19:18.229917 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-45d84"] Dec 07 16:19:18 crc kubenswrapper[4716]: I1207 16:19:18.231305 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-45d84" Dec 07 16:19:18 crc kubenswrapper[4716]: I1207 16:19:18.245828 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-lhq7l" Dec 07 16:19:18 crc kubenswrapper[4716]: I1207 16:19:18.246119 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 07 16:19:18 crc kubenswrapper[4716]: I1207 16:19:18.246406 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 07 16:19:18 crc kubenswrapper[4716]: I1207 16:19:18.246420 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 07 16:19:18 crc kubenswrapper[4716]: I1207 16:19:18.246119 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-45d84"] Dec 07 16:19:18 crc kubenswrapper[4716]: I1207 16:19:18.247208 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 07 16:19:18 crc kubenswrapper[4716]: I1207 16:19:18.375393 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d103f298-33f8-49df-894b-50b813660371-fernet-keys\") pod \"keystone-bootstrap-45d84\" (UID: \"d103f298-33f8-49df-894b-50b813660371\") " pod="openstack/keystone-bootstrap-45d84" Dec 07 16:19:18 crc kubenswrapper[4716]: I1207 16:19:18.375434 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d103f298-33f8-49df-894b-50b813660371-combined-ca-bundle\") pod \"keystone-bootstrap-45d84\" (UID: \"d103f298-33f8-49df-894b-50b813660371\") " pod="openstack/keystone-bootstrap-45d84" Dec 07 16:19:18 crc kubenswrapper[4716]: I1207 16:19:18.375514 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d103f298-33f8-49df-894b-50b813660371-credential-keys\") pod \"keystone-bootstrap-45d84\" (UID: \"d103f298-33f8-49df-894b-50b813660371\") " pod="openstack/keystone-bootstrap-45d84" Dec 07 16:19:18 crc kubenswrapper[4716]: I1207 16:19:18.375599 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d103f298-33f8-49df-894b-50b813660371-config-data\") pod \"keystone-bootstrap-45d84\" (UID: \"d103f298-33f8-49df-894b-50b813660371\") " pod="openstack/keystone-bootstrap-45d84" Dec 07 16:19:18 crc kubenswrapper[4716]: I1207 16:19:18.375623 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbzs2\" (UniqueName: \"kubernetes.io/projected/d103f298-33f8-49df-894b-50b813660371-kube-api-access-qbzs2\") pod \"keystone-bootstrap-45d84\" (UID: \"d103f298-33f8-49df-894b-50b813660371\") " pod="openstack/keystone-bootstrap-45d84" Dec 07 16:19:18 crc kubenswrapper[4716]: I1207 16:19:18.375644 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d103f298-33f8-49df-894b-50b813660371-scripts\") pod \"keystone-bootstrap-45d84\" (UID: \"d103f298-33f8-49df-894b-50b813660371\") " pod="openstack/keystone-bootstrap-45d84" Dec 07 16:19:18 crc kubenswrapper[4716]: I1207 16:19:18.477581 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d103f298-33f8-49df-894b-50b813660371-fernet-keys\") pod \"keystone-bootstrap-45d84\" (UID: \"d103f298-33f8-49df-894b-50b813660371\") " pod="openstack/keystone-bootstrap-45d84" Dec 07 16:19:18 crc kubenswrapper[4716]: I1207 16:19:18.477636 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d103f298-33f8-49df-894b-50b813660371-combined-ca-bundle\") pod \"keystone-bootstrap-45d84\" (UID: \"d103f298-33f8-49df-894b-50b813660371\") " pod="openstack/keystone-bootstrap-45d84" Dec 07 16:19:18 crc kubenswrapper[4716]: I1207 16:19:18.477717 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d103f298-33f8-49df-894b-50b813660371-credential-keys\") pod \"keystone-bootstrap-45d84\" (UID: \"d103f298-33f8-49df-894b-50b813660371\") " pod="openstack/keystone-bootstrap-45d84" Dec 07 16:19:18 crc kubenswrapper[4716]: I1207 16:19:18.477769 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d103f298-33f8-49df-894b-50b813660371-config-data\") pod \"keystone-bootstrap-45d84\" (UID: \"d103f298-33f8-49df-894b-50b813660371\") " pod="openstack/keystone-bootstrap-45d84" Dec 07 16:19:18 crc kubenswrapper[4716]: I1207 16:19:18.477794 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbzs2\" (UniqueName: \"kubernetes.io/projected/d103f298-33f8-49df-894b-50b813660371-kube-api-access-qbzs2\") pod \"keystone-bootstrap-45d84\" (UID: \"d103f298-33f8-49df-894b-50b813660371\") " pod="openstack/keystone-bootstrap-45d84" Dec 07 16:19:18 crc kubenswrapper[4716]: I1207 16:19:18.477823 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d103f298-33f8-49df-894b-50b813660371-scripts\") pod \"keystone-bootstrap-45d84\" (UID: \"d103f298-33f8-49df-894b-50b813660371\") " pod="openstack/keystone-bootstrap-45d84" Dec 07 16:19:18 crc kubenswrapper[4716]: I1207 16:19:18.486776 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d103f298-33f8-49df-894b-50b813660371-combined-ca-bundle\") pod \"keystone-bootstrap-45d84\" (UID: \"d103f298-33f8-49df-894b-50b813660371\") " pod="openstack/keystone-bootstrap-45d84" Dec 07 16:19:18 crc kubenswrapper[4716]: I1207 16:19:18.494622 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d103f298-33f8-49df-894b-50b813660371-fernet-keys\") pod \"keystone-bootstrap-45d84\" (UID: \"d103f298-33f8-49df-894b-50b813660371\") " pod="openstack/keystone-bootstrap-45d84" Dec 07 16:19:18 crc kubenswrapper[4716]: I1207 16:19:18.498642 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d103f298-33f8-49df-894b-50b813660371-credential-keys\") pod \"keystone-bootstrap-45d84\" (UID: \"d103f298-33f8-49df-894b-50b813660371\") " pod="openstack/keystone-bootstrap-45d84" Dec 07 16:19:18 crc kubenswrapper[4716]: I1207 16:19:18.500474 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d103f298-33f8-49df-894b-50b813660371-scripts\") pod \"keystone-bootstrap-45d84\" (UID: \"d103f298-33f8-49df-894b-50b813660371\") " pod="openstack/keystone-bootstrap-45d84" Dec 07 16:19:18 crc kubenswrapper[4716]: I1207 16:19:18.510640 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbzs2\" (UniqueName: \"kubernetes.io/projected/d103f298-33f8-49df-894b-50b813660371-kube-api-access-qbzs2\") pod \"keystone-bootstrap-45d84\" (UID: \"d103f298-33f8-49df-894b-50b813660371\") " pod="openstack/keystone-bootstrap-45d84" Dec 07 16:19:18 crc kubenswrapper[4716]: I1207 16:19:18.510773 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d103f298-33f8-49df-894b-50b813660371-config-data\") pod \"keystone-bootstrap-45d84\" (UID: \"d103f298-33f8-49df-894b-50b813660371\") " pod="openstack/keystone-bootstrap-45d84" Dec 07 16:19:18 crc kubenswrapper[4716]: I1207 16:19:18.573100 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-45d84" Dec 07 16:19:18 crc kubenswrapper[4716]: E1207 16:19:18.916484 4716 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-placement-api:current-podified" Dec 07 16:19:18 crc kubenswrapper[4716]: E1207 16:19:18.916919 4716 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:placement-db-sync,Image:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/placement,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:placement-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-72mp2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42482,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-db-sync-qqklx_openstack(6354ab8c-f35e-450f-96f9-8e305f778a54): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 07 16:19:18 crc kubenswrapper[4716]: E1207 16:19:18.918121 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/placement-db-sync-qqklx" podUID="6354ab8c-f35e-450f-96f9-8e305f778a54" Dec 07 16:19:18 crc kubenswrapper[4716]: E1207 16:19:18.952194 4716 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Dec 07 16:19:18 crc kubenswrapper[4716]: E1207 16:19:18.952564 4716 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nb6hddh595h599h88h599h58h5c4h558h7fh5b8h677h668h68ch9fh5bh665h598h67fh698h5d5h657h64bhfh688h569hf5h56ch586h65h85h65q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-n47p4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-8f6b84b89-znwn9_openstack(ff3e516d-0469-4e03-9d71-fed1add36b54): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 07 16:19:18 crc kubenswrapper[4716]: E1207 16:19:18.955214 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-8f6b84b89-znwn9" podUID="ff3e516d-0469-4e03-9d71-fed1add36b54" Dec 07 16:19:19 crc kubenswrapper[4716]: E1207 16:19:19.579250 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-placement-api:current-podified\\\"\"" pod="openstack/placement-db-sync-qqklx" podUID="6354ab8c-f35e-450f-96f9-8e305f778a54" Dec 07 16:19:19 crc kubenswrapper[4716]: I1207 16:19:19.668935 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0829dcfe-e23e-472a-a01a-d49351ba1f7a" path="/var/lib/kubelet/pods/0829dcfe-e23e-472a-a01a-d49351ba1f7a/volumes" Dec 07 16:19:22 crc kubenswrapper[4716]: I1207 16:19:22.761114 4716 patch_prober.go:28] interesting pod/machine-config-daemon-zcxxp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 16:19:22 crc kubenswrapper[4716]: I1207 16:19:22.761578 4716 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 16:19:25 crc kubenswrapper[4716]: I1207 16:19:25.479713 4716 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-74f6bcbc87-2cv95" podUID="7a57dd80-9afb-4d22-bd71-7104bc574ab0" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.126:5353: i/o timeout" Dec 07 16:19:29 crc kubenswrapper[4716]: I1207 16:19:29.342387 4716 scope.go:117] "RemoveContainer" containerID="7a783303abd3c25533a394e48b039cc8b2b83fbb8687254194f0b6a5eb45156b" Dec 07 16:19:29 crc kubenswrapper[4716]: I1207 16:19:29.487209 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-2cv95" Dec 07 16:19:29 crc kubenswrapper[4716]: I1207 16:19:29.499675 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-8f6b84b89-znwn9" Dec 07 16:19:29 crc kubenswrapper[4716]: I1207 16:19:29.616737 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n47p4\" (UniqueName: \"kubernetes.io/projected/ff3e516d-0469-4e03-9d71-fed1add36b54-kube-api-access-n47p4\") pod \"ff3e516d-0469-4e03-9d71-fed1add36b54\" (UID: \"ff3e516d-0469-4e03-9d71-fed1add36b54\") " Dec 07 16:19:29 crc kubenswrapper[4716]: I1207 16:19:29.617067 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ff3e516d-0469-4e03-9d71-fed1add36b54-config-data\") pod \"ff3e516d-0469-4e03-9d71-fed1add36b54\" (UID: \"ff3e516d-0469-4e03-9d71-fed1add36b54\") " Dec 07 16:19:29 crc kubenswrapper[4716]: I1207 16:19:29.617207 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dqqlg\" (UniqueName: \"kubernetes.io/projected/7a57dd80-9afb-4d22-bd71-7104bc574ab0-kube-api-access-dqqlg\") pod \"7a57dd80-9afb-4d22-bd71-7104bc574ab0\" (UID: \"7a57dd80-9afb-4d22-bd71-7104bc574ab0\") " Dec 07 16:19:29 crc kubenswrapper[4716]: I1207 16:19:29.617248 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7a57dd80-9afb-4d22-bd71-7104bc574ab0-dns-svc\") pod \"7a57dd80-9afb-4d22-bd71-7104bc574ab0\" (UID: \"7a57dd80-9afb-4d22-bd71-7104bc574ab0\") " Dec 07 16:19:29 crc kubenswrapper[4716]: I1207 16:19:29.617286 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7a57dd80-9afb-4d22-bd71-7104bc574ab0-dns-swift-storage-0\") pod \"7a57dd80-9afb-4d22-bd71-7104bc574ab0\" (UID: \"7a57dd80-9afb-4d22-bd71-7104bc574ab0\") " Dec 07 16:19:29 crc kubenswrapper[4716]: I1207 16:19:29.617312 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7a57dd80-9afb-4d22-bd71-7104bc574ab0-ovsdbserver-sb\") pod \"7a57dd80-9afb-4d22-bd71-7104bc574ab0\" (UID: \"7a57dd80-9afb-4d22-bd71-7104bc574ab0\") " Dec 07 16:19:29 crc kubenswrapper[4716]: I1207 16:19:29.617358 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ff3e516d-0469-4e03-9d71-fed1add36b54-horizon-secret-key\") pod \"ff3e516d-0469-4e03-9d71-fed1add36b54\" (UID: \"ff3e516d-0469-4e03-9d71-fed1add36b54\") " Dec 07 16:19:29 crc kubenswrapper[4716]: I1207 16:19:29.617386 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a57dd80-9afb-4d22-bd71-7104bc574ab0-config\") pod \"7a57dd80-9afb-4d22-bd71-7104bc574ab0\" (UID: \"7a57dd80-9afb-4d22-bd71-7104bc574ab0\") " Dec 07 16:19:29 crc kubenswrapper[4716]: I1207 16:19:29.617467 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ff3e516d-0469-4e03-9d71-fed1add36b54-scripts\") pod \"ff3e516d-0469-4e03-9d71-fed1add36b54\" (UID: \"ff3e516d-0469-4e03-9d71-fed1add36b54\") " Dec 07 16:19:29 crc kubenswrapper[4716]: I1207 16:19:29.617512 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7a57dd80-9afb-4d22-bd71-7104bc574ab0-ovsdbserver-nb\") pod \"7a57dd80-9afb-4d22-bd71-7104bc574ab0\" (UID: \"7a57dd80-9afb-4d22-bd71-7104bc574ab0\") " Dec 07 16:19:29 crc kubenswrapper[4716]: I1207 16:19:29.617571 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff3e516d-0469-4e03-9d71-fed1add36b54-logs\") pod \"ff3e516d-0469-4e03-9d71-fed1add36b54\" (UID: \"ff3e516d-0469-4e03-9d71-fed1add36b54\") " Dec 07 16:19:29 crc kubenswrapper[4716]: I1207 16:19:29.617821 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff3e516d-0469-4e03-9d71-fed1add36b54-config-data" (OuterVolumeSpecName: "config-data") pod "ff3e516d-0469-4e03-9d71-fed1add36b54" (UID: "ff3e516d-0469-4e03-9d71-fed1add36b54"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:19:29 crc kubenswrapper[4716]: I1207 16:19:29.618127 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff3e516d-0469-4e03-9d71-fed1add36b54-logs" (OuterVolumeSpecName: "logs") pod "ff3e516d-0469-4e03-9d71-fed1add36b54" (UID: "ff3e516d-0469-4e03-9d71-fed1add36b54"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:19:29 crc kubenswrapper[4716]: I1207 16:19:29.618307 4716 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ff3e516d-0469-4e03-9d71-fed1add36b54-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 16:19:29 crc kubenswrapper[4716]: I1207 16:19:29.619394 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff3e516d-0469-4e03-9d71-fed1add36b54-scripts" (OuterVolumeSpecName: "scripts") pod "ff3e516d-0469-4e03-9d71-fed1add36b54" (UID: "ff3e516d-0469-4e03-9d71-fed1add36b54"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:19:29 crc kubenswrapper[4716]: I1207 16:19:29.622157 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff3e516d-0469-4e03-9d71-fed1add36b54-kube-api-access-n47p4" (OuterVolumeSpecName: "kube-api-access-n47p4") pod "ff3e516d-0469-4e03-9d71-fed1add36b54" (UID: "ff3e516d-0469-4e03-9d71-fed1add36b54"). InnerVolumeSpecName "kube-api-access-n47p4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:19:29 crc kubenswrapper[4716]: I1207 16:19:29.622354 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff3e516d-0469-4e03-9d71-fed1add36b54-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "ff3e516d-0469-4e03-9d71-fed1add36b54" (UID: "ff3e516d-0469-4e03-9d71-fed1add36b54"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:19:29 crc kubenswrapper[4716]: I1207 16:19:29.633332 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a57dd80-9afb-4d22-bd71-7104bc574ab0-kube-api-access-dqqlg" (OuterVolumeSpecName: "kube-api-access-dqqlg") pod "7a57dd80-9afb-4d22-bd71-7104bc574ab0" (UID: "7a57dd80-9afb-4d22-bd71-7104bc574ab0"). InnerVolumeSpecName "kube-api-access-dqqlg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:19:29 crc kubenswrapper[4716]: I1207 16:19:29.667173 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a57dd80-9afb-4d22-bd71-7104bc574ab0-config" (OuterVolumeSpecName: "config") pod "7a57dd80-9afb-4d22-bd71-7104bc574ab0" (UID: "7a57dd80-9afb-4d22-bd71-7104bc574ab0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:19:29 crc kubenswrapper[4716]: I1207 16:19:29.669318 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a57dd80-9afb-4d22-bd71-7104bc574ab0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7a57dd80-9afb-4d22-bd71-7104bc574ab0" (UID: "7a57dd80-9afb-4d22-bd71-7104bc574ab0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:19:29 crc kubenswrapper[4716]: I1207 16:19:29.671223 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a57dd80-9afb-4d22-bd71-7104bc574ab0-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "7a57dd80-9afb-4d22-bd71-7104bc574ab0" (UID: "7a57dd80-9afb-4d22-bd71-7104bc574ab0"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:19:29 crc kubenswrapper[4716]: I1207 16:19:29.677862 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a57dd80-9afb-4d22-bd71-7104bc574ab0-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7a57dd80-9afb-4d22-bd71-7104bc574ab0" (UID: "7a57dd80-9afb-4d22-bd71-7104bc574ab0"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:19:29 crc kubenswrapper[4716]: I1207 16:19:29.678795 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a57dd80-9afb-4d22-bd71-7104bc574ab0-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7a57dd80-9afb-4d22-bd71-7104bc574ab0" (UID: "7a57dd80-9afb-4d22-bd71-7104bc574ab0"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:19:29 crc kubenswrapper[4716]: I1207 16:19:29.685922 4716 generic.go:334] "Generic (PLEG): container finished" podID="925be76f-fe07-4eb8-982a-02a0a002ea58" containerID="f899907d58ebda8ca729b889e0d1a75de6da61c3d90e525bba1d8e11925ba380" exitCode=0 Dec 07 16:19:29 crc kubenswrapper[4716]: I1207 16:19:29.686004 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-kl4tr" event={"ID":"925be76f-fe07-4eb8-982a-02a0a002ea58","Type":"ContainerDied","Data":"f899907d58ebda8ca729b889e0d1a75de6da61c3d90e525bba1d8e11925ba380"} Dec 07 16:19:29 crc kubenswrapper[4716]: I1207 16:19:29.687351 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-8f6b84b89-znwn9" Dec 07 16:19:29 crc kubenswrapper[4716]: I1207 16:19:29.687362 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-8f6b84b89-znwn9" event={"ID":"ff3e516d-0469-4e03-9d71-fed1add36b54","Type":"ContainerDied","Data":"3722bd5a37148297b80d2466823ef2b5a4826285863eeff61619558b85b50c6d"} Dec 07 16:19:29 crc kubenswrapper[4716]: I1207 16:19:29.692318 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-2cv95" event={"ID":"7a57dd80-9afb-4d22-bd71-7104bc574ab0","Type":"ContainerDied","Data":"57cf0bd73a383215f36e9929371ab9c687a55f290d3c8edd4083c3bed0d35241"} Dec 07 16:19:29 crc kubenswrapper[4716]: I1207 16:19:29.692335 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-2cv95" Dec 07 16:19:29 crc kubenswrapper[4716]: I1207 16:19:29.726443 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dqqlg\" (UniqueName: \"kubernetes.io/projected/7a57dd80-9afb-4d22-bd71-7104bc574ab0-kube-api-access-dqqlg\") on node \"crc\" DevicePath \"\"" Dec 07 16:19:29 crc kubenswrapper[4716]: I1207 16:19:29.726482 4716 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7a57dd80-9afb-4d22-bd71-7104bc574ab0-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 07 16:19:29 crc kubenswrapper[4716]: I1207 16:19:29.726493 4716 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7a57dd80-9afb-4d22-bd71-7104bc574ab0-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 07 16:19:29 crc kubenswrapper[4716]: I1207 16:19:29.726502 4716 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7a57dd80-9afb-4d22-bd71-7104bc574ab0-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 07 16:19:29 crc kubenswrapper[4716]: I1207 16:19:29.726513 4716 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ff3e516d-0469-4e03-9d71-fed1add36b54-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 07 16:19:29 crc kubenswrapper[4716]: I1207 16:19:29.726522 4716 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a57dd80-9afb-4d22-bd71-7104bc574ab0-config\") on node \"crc\" DevicePath \"\"" Dec 07 16:19:29 crc kubenswrapper[4716]: I1207 16:19:29.726530 4716 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ff3e516d-0469-4e03-9d71-fed1add36b54-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 16:19:29 crc kubenswrapper[4716]: I1207 16:19:29.726539 4716 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7a57dd80-9afb-4d22-bd71-7104bc574ab0-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 07 16:19:29 crc kubenswrapper[4716]: I1207 16:19:29.726549 4716 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff3e516d-0469-4e03-9d71-fed1add36b54-logs\") on node \"crc\" DevicePath \"\"" Dec 07 16:19:29 crc kubenswrapper[4716]: I1207 16:19:29.726557 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n47p4\" (UniqueName: \"kubernetes.io/projected/ff3e516d-0469-4e03-9d71-fed1add36b54-kube-api-access-n47p4\") on node \"crc\" DevicePath \"\"" Dec 07 16:19:29 crc kubenswrapper[4716]: I1207 16:19:29.757116 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-8f6b84b89-znwn9"] Dec 07 16:19:29 crc kubenswrapper[4716]: I1207 16:19:29.767239 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-8f6b84b89-znwn9"] Dec 07 16:19:29 crc kubenswrapper[4716]: I1207 16:19:29.775907 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-2cv95"] Dec 07 16:19:29 crc kubenswrapper[4716]: I1207 16:19:29.781203 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-2cv95"] Dec 07 16:19:30 crc kubenswrapper[4716]: I1207 16:19:30.480962 4716 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-74f6bcbc87-2cv95" podUID="7a57dd80-9afb-4d22-bd71-7104bc574ab0" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.126:5353: i/o timeout" Dec 07 16:19:30 crc kubenswrapper[4716]: E1207 16:19:30.584659 4716 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Dec 07 16:19:30 crc kubenswrapper[4716]: E1207 16:19:30.584838 4716 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nrnjl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-gqg6b_openstack(f9413953-056c-4d73-b534-12e6816e6fb4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 07 16:19:30 crc kubenswrapper[4716]: E1207 16:19:30.586461 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-gqg6b" podUID="f9413953-056c-4d73-b534-12e6816e6fb4" Dec 07 16:19:30 crc kubenswrapper[4716]: E1207 16:19:30.711333 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-gqg6b" podUID="f9413953-056c-4d73-b534-12e6816e6fb4" Dec 07 16:19:31 crc kubenswrapper[4716]: E1207 16:19:31.135922 4716 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Dec 07 16:19:31 crc kubenswrapper[4716]: E1207 16:19:31.136376 4716 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qgdkv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-tk47q_openstack(717a326f-b1ac-4fbe-b1c8-b669081dab80): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 07 16:19:31 crc kubenswrapper[4716]: E1207 16:19:31.137547 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-tk47q" podUID="717a326f-b1ac-4fbe-b1c8-b669081dab80" Dec 07 16:19:31 crc kubenswrapper[4716]: I1207 16:19:31.175220 4716 scope.go:117] "RemoveContainer" containerID="83e3fba825b5a95d45367566ff2f290a9309499d1a555060d5b08b31407af7c3" Dec 07 16:19:31 crc kubenswrapper[4716]: I1207 16:19:31.233332 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-kl4tr" Dec 07 16:19:31 crc kubenswrapper[4716]: I1207 16:19:31.326422 4716 scope.go:117] "RemoveContainer" containerID="38509b7504f5aabae26ab0ffb9fd3aadcfe7a3b509c4bb131c367e5b8cbf7d3c" Dec 07 16:19:31 crc kubenswrapper[4716]: I1207 16:19:31.354358 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/925be76f-fe07-4eb8-982a-02a0a002ea58-combined-ca-bundle\") pod \"925be76f-fe07-4eb8-982a-02a0a002ea58\" (UID: \"925be76f-fe07-4eb8-982a-02a0a002ea58\") " Dec 07 16:19:31 crc kubenswrapper[4716]: I1207 16:19:31.354541 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4nf9\" (UniqueName: \"kubernetes.io/projected/925be76f-fe07-4eb8-982a-02a0a002ea58-kube-api-access-s4nf9\") pod \"925be76f-fe07-4eb8-982a-02a0a002ea58\" (UID: \"925be76f-fe07-4eb8-982a-02a0a002ea58\") " Dec 07 16:19:31 crc kubenswrapper[4716]: I1207 16:19:31.354693 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/925be76f-fe07-4eb8-982a-02a0a002ea58-config\") pod \"925be76f-fe07-4eb8-982a-02a0a002ea58\" (UID: \"925be76f-fe07-4eb8-982a-02a0a002ea58\") " Dec 07 16:19:31 crc kubenswrapper[4716]: I1207 16:19:31.360223 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925be76f-fe07-4eb8-982a-02a0a002ea58-kube-api-access-s4nf9" (OuterVolumeSpecName: "kube-api-access-s4nf9") pod "925be76f-fe07-4eb8-982a-02a0a002ea58" (UID: "925be76f-fe07-4eb8-982a-02a0a002ea58"). InnerVolumeSpecName "kube-api-access-s4nf9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:19:31 crc kubenswrapper[4716]: I1207 16:19:31.431222 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925be76f-fe07-4eb8-982a-02a0a002ea58-config" (OuterVolumeSpecName: "config") pod "925be76f-fe07-4eb8-982a-02a0a002ea58" (UID: "925be76f-fe07-4eb8-982a-02a0a002ea58"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:19:31 crc kubenswrapper[4716]: I1207 16:19:31.439495 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925be76f-fe07-4eb8-982a-02a0a002ea58-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "925be76f-fe07-4eb8-982a-02a0a002ea58" (UID: "925be76f-fe07-4eb8-982a-02a0a002ea58"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:19:31 crc kubenswrapper[4716]: I1207 16:19:31.456604 4716 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/925be76f-fe07-4eb8-982a-02a0a002ea58-config\") on node \"crc\" DevicePath \"\"" Dec 07 16:19:31 crc kubenswrapper[4716]: I1207 16:19:31.456640 4716 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/925be76f-fe07-4eb8-982a-02a0a002ea58-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 16:19:31 crc kubenswrapper[4716]: I1207 16:19:31.456653 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4nf9\" (UniqueName: \"kubernetes.io/projected/925be76f-fe07-4eb8-982a-02a0a002ea58-kube-api-access-s4nf9\") on node \"crc\" DevicePath \"\"" Dec 07 16:19:31 crc kubenswrapper[4716]: I1207 16:19:31.513706 4716 scope.go:117] "RemoveContainer" containerID="670d21ca07fde13d32e5cc95349a99ca26fd9a0580fa5a23a4f056dd9038b973" Dec 07 16:19:31 crc kubenswrapper[4716]: I1207 16:19:31.537116 4716 scope.go:117] "RemoveContainer" containerID="80cfd80e2499f2e974a5f55feaab69e7d3d1fc4484c0b6b5aa1f696d3561813c" Dec 07 16:19:31 crc kubenswrapper[4716]: I1207 16:19:31.716295 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a57dd80-9afb-4d22-bd71-7104bc574ab0" path="/var/lib/kubelet/pods/7a57dd80-9afb-4d22-bd71-7104bc574ab0/volumes" Dec 07 16:19:31 crc kubenswrapper[4716]: I1207 16:19:31.717156 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff3e516d-0469-4e03-9d71-fed1add36b54" path="/var/lib/kubelet/pods/ff3e516d-0469-4e03-9d71-fed1add36b54/volumes" Dec 07 16:19:31 crc kubenswrapper[4716]: I1207 16:19:31.717697 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-86c7567d4-99rx9"] Dec 07 16:19:31 crc kubenswrapper[4716]: I1207 16:19:31.742317 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7f9b78866d-9g9kv"] Dec 07 16:19:31 crc kubenswrapper[4716]: I1207 16:19:31.749137 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2ee00bf0-1412-4fca-a570-6600db549923","Type":"ContainerStarted","Data":"65d1e11cc55500273d099448bfdb88d42d28a5fe058e6da987f62dc02869707b"} Dec 07 16:19:31 crc kubenswrapper[4716]: I1207 16:19:31.750996 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-kl4tr" event={"ID":"925be76f-fe07-4eb8-982a-02a0a002ea58","Type":"ContainerDied","Data":"a41869ed94b7de3cafdecb2332a0dd22951bc27f6dc071e3b2b8d6fafde9e4ab"} Dec 07 16:19:31 crc kubenswrapper[4716]: I1207 16:19:31.751022 4716 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a41869ed94b7de3cafdecb2332a0dd22951bc27f6dc071e3b2b8d6fafde9e4ab" Dec 07 16:19:31 crc kubenswrapper[4716]: I1207 16:19:31.751114 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-kl4tr" Dec 07 16:19:31 crc kubenswrapper[4716]: I1207 16:19:31.753599 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-86c7567d4-99rx9" event={"ID":"c291e114-7940-46bc-91d3-a8fa256549ff","Type":"ContainerStarted","Data":"3738b4121edde6afb7201317a053ff7656912527a99eb08cd861bd6904dd5dc4"} Dec 07 16:19:31 crc kubenswrapper[4716]: I1207 16:19:31.756051 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d8b488bdf-fvlzm" event={"ID":"d7db6aff-6831-49ea-9d51-24ec9ac40a88","Type":"ContainerStarted","Data":"d9c99d90e0f76afdd5481a03c49c1a155a22721530d081f1e90302e0f9d9fc9d"} Dec 07 16:19:31 crc kubenswrapper[4716]: I1207 16:19:31.766734 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d76599897-4mhqt" event={"ID":"5c76ced8-9170-4ccc-a497-87388e8546dd","Type":"ContainerStarted","Data":"008406d13a6ea7c73401539a525b2aa9a7d98be0c715633e8d8eac24e5c7ccf6"} Dec 07 16:19:31 crc kubenswrapper[4716]: E1207 16:19:31.841217 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-tk47q" podUID="717a326f-b1ac-4fbe-b1c8-b669081dab80" Dec 07 16:19:31 crc kubenswrapper[4716]: I1207 16:19:31.901450 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-45d84"] Dec 07 16:19:32 crc kubenswrapper[4716]: I1207 16:19:32.046143 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-rtqpb"] Dec 07 16:19:32 crc kubenswrapper[4716]: E1207 16:19:32.046681 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a57dd80-9afb-4d22-bd71-7104bc574ab0" containerName="init" Dec 07 16:19:32 crc kubenswrapper[4716]: I1207 16:19:32.046700 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a57dd80-9afb-4d22-bd71-7104bc574ab0" containerName="init" Dec 07 16:19:32 crc kubenswrapper[4716]: E1207 16:19:32.046733 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="925be76f-fe07-4eb8-982a-02a0a002ea58" containerName="neutron-db-sync" Dec 07 16:19:32 crc kubenswrapper[4716]: I1207 16:19:32.046740 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="925be76f-fe07-4eb8-982a-02a0a002ea58" containerName="neutron-db-sync" Dec 07 16:19:32 crc kubenswrapper[4716]: E1207 16:19:32.046754 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a57dd80-9afb-4d22-bd71-7104bc574ab0" containerName="dnsmasq-dns" Dec 07 16:19:32 crc kubenswrapper[4716]: I1207 16:19:32.046760 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a57dd80-9afb-4d22-bd71-7104bc574ab0" containerName="dnsmasq-dns" Dec 07 16:19:32 crc kubenswrapper[4716]: I1207 16:19:32.046928 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="925be76f-fe07-4eb8-982a-02a0a002ea58" containerName="neutron-db-sync" Dec 07 16:19:32 crc kubenswrapper[4716]: I1207 16:19:32.046943 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a57dd80-9afb-4d22-bd71-7104bc574ab0" containerName="dnsmasq-dns" Dec 07 16:19:32 crc kubenswrapper[4716]: I1207 16:19:32.047963 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-rtqpb" Dec 07 16:19:32 crc kubenswrapper[4716]: I1207 16:19:32.066807 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-rtqpb"] Dec 07 16:19:32 crc kubenswrapper[4716]: I1207 16:19:32.144139 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-69d87fd46d-9k5rq"] Dec 07 16:19:32 crc kubenswrapper[4716]: I1207 16:19:32.145748 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-69d87fd46d-9k5rq" Dec 07 16:19:32 crc kubenswrapper[4716]: I1207 16:19:32.149605 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 07 16:19:32 crc kubenswrapper[4716]: I1207 16:19:32.149923 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 07 16:19:32 crc kubenswrapper[4716]: I1207 16:19:32.150044 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-pm9qx" Dec 07 16:19:32 crc kubenswrapper[4716]: I1207 16:19:32.150177 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Dec 07 16:19:32 crc kubenswrapper[4716]: I1207 16:19:32.185310 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b1a41e8-004e-4562-a8aa-79e422e11e83-config\") pod \"dnsmasq-dns-55f844cf75-rtqpb\" (UID: \"8b1a41e8-004e-4562-a8aa-79e422e11e83\") " pod="openstack/dnsmasq-dns-55f844cf75-rtqpb" Dec 07 16:19:32 crc kubenswrapper[4716]: I1207 16:19:32.185402 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8b1a41e8-004e-4562-a8aa-79e422e11e83-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-rtqpb\" (UID: \"8b1a41e8-004e-4562-a8aa-79e422e11e83\") " pod="openstack/dnsmasq-dns-55f844cf75-rtqpb" Dec 07 16:19:32 crc kubenswrapper[4716]: I1207 16:19:32.185455 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqjwl\" (UniqueName: \"kubernetes.io/projected/8b1a41e8-004e-4562-a8aa-79e422e11e83-kube-api-access-vqjwl\") pod \"dnsmasq-dns-55f844cf75-rtqpb\" (UID: \"8b1a41e8-004e-4562-a8aa-79e422e11e83\") " pod="openstack/dnsmasq-dns-55f844cf75-rtqpb" Dec 07 16:19:32 crc kubenswrapper[4716]: I1207 16:19:32.185552 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8b1a41e8-004e-4562-a8aa-79e422e11e83-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-rtqpb\" (UID: \"8b1a41e8-004e-4562-a8aa-79e422e11e83\") " pod="openstack/dnsmasq-dns-55f844cf75-rtqpb" Dec 07 16:19:32 crc kubenswrapper[4716]: I1207 16:19:32.185587 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8b1a41e8-004e-4562-a8aa-79e422e11e83-dns-svc\") pod \"dnsmasq-dns-55f844cf75-rtqpb\" (UID: \"8b1a41e8-004e-4562-a8aa-79e422e11e83\") " pod="openstack/dnsmasq-dns-55f844cf75-rtqpb" Dec 07 16:19:32 crc kubenswrapper[4716]: I1207 16:19:32.185616 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8b1a41e8-004e-4562-a8aa-79e422e11e83-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-rtqpb\" (UID: \"8b1a41e8-004e-4562-a8aa-79e422e11e83\") " pod="openstack/dnsmasq-dns-55f844cf75-rtqpb" Dec 07 16:19:32 crc kubenswrapper[4716]: I1207 16:19:32.200052 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 07 16:19:32 crc kubenswrapper[4716]: I1207 16:19:32.251097 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-69d87fd46d-9k5rq"] Dec 07 16:19:32 crc kubenswrapper[4716]: I1207 16:19:32.291545 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8b1a41e8-004e-4562-a8aa-79e422e11e83-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-rtqpb\" (UID: \"8b1a41e8-004e-4562-a8aa-79e422e11e83\") " pod="openstack/dnsmasq-dns-55f844cf75-rtqpb" Dec 07 16:19:32 crc kubenswrapper[4716]: I1207 16:19:32.292047 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vqjwl\" (UniqueName: \"kubernetes.io/projected/8b1a41e8-004e-4562-a8aa-79e422e11e83-kube-api-access-vqjwl\") pod \"dnsmasq-dns-55f844cf75-rtqpb\" (UID: \"8b1a41e8-004e-4562-a8aa-79e422e11e83\") " pod="openstack/dnsmasq-dns-55f844cf75-rtqpb" Dec 07 16:19:32 crc kubenswrapper[4716]: I1207 16:19:32.292102 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jxj87\" (UniqueName: \"kubernetes.io/projected/8b125bd6-f08f-442c-96a8-b903160be13a-kube-api-access-jxj87\") pod \"neutron-69d87fd46d-9k5rq\" (UID: \"8b125bd6-f08f-442c-96a8-b903160be13a\") " pod="openstack/neutron-69d87fd46d-9k5rq" Dec 07 16:19:32 crc kubenswrapper[4716]: I1207 16:19:32.292163 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b125bd6-f08f-442c-96a8-b903160be13a-combined-ca-bundle\") pod \"neutron-69d87fd46d-9k5rq\" (UID: \"8b125bd6-f08f-442c-96a8-b903160be13a\") " pod="openstack/neutron-69d87fd46d-9k5rq" Dec 07 16:19:32 crc kubenswrapper[4716]: I1207 16:19:32.292310 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8b1a41e8-004e-4562-a8aa-79e422e11e83-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-rtqpb\" (UID: \"8b1a41e8-004e-4562-a8aa-79e422e11e83\") " pod="openstack/dnsmasq-dns-55f844cf75-rtqpb" Dec 07 16:19:32 crc kubenswrapper[4716]: I1207 16:19:32.292356 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8b1a41e8-004e-4562-a8aa-79e422e11e83-dns-svc\") pod \"dnsmasq-dns-55f844cf75-rtqpb\" (UID: \"8b1a41e8-004e-4562-a8aa-79e422e11e83\") " pod="openstack/dnsmasq-dns-55f844cf75-rtqpb" Dec 07 16:19:32 crc kubenswrapper[4716]: I1207 16:19:32.292385 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8b125bd6-f08f-442c-96a8-b903160be13a-config\") pod \"neutron-69d87fd46d-9k5rq\" (UID: \"8b125bd6-f08f-442c-96a8-b903160be13a\") " pod="openstack/neutron-69d87fd46d-9k5rq" Dec 07 16:19:32 crc kubenswrapper[4716]: I1207 16:19:32.292417 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8b1a41e8-004e-4562-a8aa-79e422e11e83-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-rtqpb\" (UID: \"8b1a41e8-004e-4562-a8aa-79e422e11e83\") " pod="openstack/dnsmasq-dns-55f844cf75-rtqpb" Dec 07 16:19:32 crc kubenswrapper[4716]: I1207 16:19:32.292445 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b125bd6-f08f-442c-96a8-b903160be13a-ovndb-tls-certs\") pod \"neutron-69d87fd46d-9k5rq\" (UID: \"8b125bd6-f08f-442c-96a8-b903160be13a\") " pod="openstack/neutron-69d87fd46d-9k5rq" Dec 07 16:19:32 crc kubenswrapper[4716]: I1207 16:19:32.292537 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/8b125bd6-f08f-442c-96a8-b903160be13a-httpd-config\") pod \"neutron-69d87fd46d-9k5rq\" (UID: \"8b125bd6-f08f-442c-96a8-b903160be13a\") " pod="openstack/neutron-69d87fd46d-9k5rq" Dec 07 16:19:32 crc kubenswrapper[4716]: I1207 16:19:32.293425 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b1a41e8-004e-4562-a8aa-79e422e11e83-config\") pod \"dnsmasq-dns-55f844cf75-rtqpb\" (UID: \"8b1a41e8-004e-4562-a8aa-79e422e11e83\") " pod="openstack/dnsmasq-dns-55f844cf75-rtqpb" Dec 07 16:19:32 crc kubenswrapper[4716]: I1207 16:19:32.293850 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8b1a41e8-004e-4562-a8aa-79e422e11e83-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-rtqpb\" (UID: \"8b1a41e8-004e-4562-a8aa-79e422e11e83\") " pod="openstack/dnsmasq-dns-55f844cf75-rtqpb" Dec 07 16:19:32 crc kubenswrapper[4716]: I1207 16:19:32.294479 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b1a41e8-004e-4562-a8aa-79e422e11e83-config\") pod \"dnsmasq-dns-55f844cf75-rtqpb\" (UID: \"8b1a41e8-004e-4562-a8aa-79e422e11e83\") " pod="openstack/dnsmasq-dns-55f844cf75-rtqpb" Dec 07 16:19:32 crc kubenswrapper[4716]: I1207 16:19:32.294541 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8b1a41e8-004e-4562-a8aa-79e422e11e83-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-rtqpb\" (UID: \"8b1a41e8-004e-4562-a8aa-79e422e11e83\") " pod="openstack/dnsmasq-dns-55f844cf75-rtqpb" Dec 07 16:19:32 crc kubenswrapper[4716]: I1207 16:19:32.295355 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8b1a41e8-004e-4562-a8aa-79e422e11e83-dns-svc\") pod \"dnsmasq-dns-55f844cf75-rtqpb\" (UID: \"8b1a41e8-004e-4562-a8aa-79e422e11e83\") " pod="openstack/dnsmasq-dns-55f844cf75-rtqpb" Dec 07 16:19:32 crc kubenswrapper[4716]: I1207 16:19:32.305810 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8b1a41e8-004e-4562-a8aa-79e422e11e83-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-rtqpb\" (UID: \"8b1a41e8-004e-4562-a8aa-79e422e11e83\") " pod="openstack/dnsmasq-dns-55f844cf75-rtqpb" Dec 07 16:19:32 crc kubenswrapper[4716]: I1207 16:19:32.324295 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqjwl\" (UniqueName: \"kubernetes.io/projected/8b1a41e8-004e-4562-a8aa-79e422e11e83-kube-api-access-vqjwl\") pod \"dnsmasq-dns-55f844cf75-rtqpb\" (UID: \"8b1a41e8-004e-4562-a8aa-79e422e11e83\") " pod="openstack/dnsmasq-dns-55f844cf75-rtqpb" Dec 07 16:19:32 crc kubenswrapper[4716]: I1207 16:19:32.330698 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-rtqpb" Dec 07 16:19:32 crc kubenswrapper[4716]: I1207 16:19:32.394695 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/8b125bd6-f08f-442c-96a8-b903160be13a-httpd-config\") pod \"neutron-69d87fd46d-9k5rq\" (UID: \"8b125bd6-f08f-442c-96a8-b903160be13a\") " pod="openstack/neutron-69d87fd46d-9k5rq" Dec 07 16:19:32 crc kubenswrapper[4716]: I1207 16:19:32.394810 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jxj87\" (UniqueName: \"kubernetes.io/projected/8b125bd6-f08f-442c-96a8-b903160be13a-kube-api-access-jxj87\") pod \"neutron-69d87fd46d-9k5rq\" (UID: \"8b125bd6-f08f-442c-96a8-b903160be13a\") " pod="openstack/neutron-69d87fd46d-9k5rq" Dec 07 16:19:32 crc kubenswrapper[4716]: I1207 16:19:32.394841 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b125bd6-f08f-442c-96a8-b903160be13a-combined-ca-bundle\") pod \"neutron-69d87fd46d-9k5rq\" (UID: \"8b125bd6-f08f-442c-96a8-b903160be13a\") " pod="openstack/neutron-69d87fd46d-9k5rq" Dec 07 16:19:32 crc kubenswrapper[4716]: I1207 16:19:32.395004 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8b125bd6-f08f-442c-96a8-b903160be13a-config\") pod \"neutron-69d87fd46d-9k5rq\" (UID: \"8b125bd6-f08f-442c-96a8-b903160be13a\") " pod="openstack/neutron-69d87fd46d-9k5rq" Dec 07 16:19:32 crc kubenswrapper[4716]: I1207 16:19:32.395039 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b125bd6-f08f-442c-96a8-b903160be13a-ovndb-tls-certs\") pod \"neutron-69d87fd46d-9k5rq\" (UID: \"8b125bd6-f08f-442c-96a8-b903160be13a\") " pod="openstack/neutron-69d87fd46d-9k5rq" Dec 07 16:19:32 crc kubenswrapper[4716]: I1207 16:19:32.400888 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/8b125bd6-f08f-442c-96a8-b903160be13a-config\") pod \"neutron-69d87fd46d-9k5rq\" (UID: \"8b125bd6-f08f-442c-96a8-b903160be13a\") " pod="openstack/neutron-69d87fd46d-9k5rq" Dec 07 16:19:32 crc kubenswrapper[4716]: I1207 16:19:32.402405 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/8b125bd6-f08f-442c-96a8-b903160be13a-httpd-config\") pod \"neutron-69d87fd46d-9k5rq\" (UID: \"8b125bd6-f08f-442c-96a8-b903160be13a\") " pod="openstack/neutron-69d87fd46d-9k5rq" Dec 07 16:19:32 crc kubenswrapper[4716]: I1207 16:19:32.404817 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b125bd6-f08f-442c-96a8-b903160be13a-ovndb-tls-certs\") pod \"neutron-69d87fd46d-9k5rq\" (UID: \"8b125bd6-f08f-442c-96a8-b903160be13a\") " pod="openstack/neutron-69d87fd46d-9k5rq" Dec 07 16:19:32 crc kubenswrapper[4716]: I1207 16:19:32.414956 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jxj87\" (UniqueName: \"kubernetes.io/projected/8b125bd6-f08f-442c-96a8-b903160be13a-kube-api-access-jxj87\") pod \"neutron-69d87fd46d-9k5rq\" (UID: \"8b125bd6-f08f-442c-96a8-b903160be13a\") " pod="openstack/neutron-69d87fd46d-9k5rq" Dec 07 16:19:32 crc kubenswrapper[4716]: I1207 16:19:32.427391 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b125bd6-f08f-442c-96a8-b903160be13a-combined-ca-bundle\") pod \"neutron-69d87fd46d-9k5rq\" (UID: \"8b125bd6-f08f-442c-96a8-b903160be13a\") " pod="openstack/neutron-69d87fd46d-9k5rq" Dec 07 16:19:32 crc kubenswrapper[4716]: I1207 16:19:32.636496 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-69d87fd46d-9k5rq" Dec 07 16:19:32 crc kubenswrapper[4716]: I1207 16:19:32.725886 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 07 16:19:32 crc kubenswrapper[4716]: I1207 16:19:32.902636 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"afba3ceb-86bf-4aaf-8ac3-b58ab411850c","Type":"ContainerStarted","Data":"c28bdf8eceedbc243af87786b8abfd72c50c2a43117daebec6acb9d7c89977a0"} Dec 07 16:19:32 crc kubenswrapper[4716]: I1207 16:19:32.909359 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5eca408c-c86b-4dd6-a1f9-81ef16e12da4","Type":"ContainerStarted","Data":"74f894d324ce5d8ac07b7d951146a82fe1f7fc35ef5ec12b146942ecc20cd47a"} Dec 07 16:19:32 crc kubenswrapper[4716]: I1207 16:19:32.920256 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-45d84" event={"ID":"d103f298-33f8-49df-894b-50b813660371","Type":"ContainerStarted","Data":"3a713685c633539179ae87831baab5d3dfb90504cd091959ed09697ea02e7224"} Dec 07 16:19:32 crc kubenswrapper[4716]: I1207 16:19:32.960603 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7d76599897-4mhqt" podUID="5c76ced8-9170-4ccc-a497-87388e8546dd" containerName="horizon-log" containerID="cri-o://008406d13a6ea7c73401539a525b2aa9a7d98be0c715633e8d8eac24e5c7ccf6" gracePeriod=30 Dec 07 16:19:32 crc kubenswrapper[4716]: I1207 16:19:32.960696 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d76599897-4mhqt" event={"ID":"5c76ced8-9170-4ccc-a497-87388e8546dd","Type":"ContainerStarted","Data":"c372e8f6fe59dccc9147f9a8695d09a6bc9fc695005f54aabb932a3697e49115"} Dec 07 16:19:32 crc kubenswrapper[4716]: I1207 16:19:32.962869 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7d76599897-4mhqt" podUID="5c76ced8-9170-4ccc-a497-87388e8546dd" containerName="horizon" containerID="cri-o://c372e8f6fe59dccc9147f9a8695d09a6bc9fc695005f54aabb932a3697e49115" gracePeriod=30 Dec 07 16:19:33 crc kubenswrapper[4716]: I1207 16:19:32.999570 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f9b78866d-9g9kv" event={"ID":"d845ae0b-5d45-4021-a1e9-4b124298b65b","Type":"ContainerStarted","Data":"6c3bf64cb5cff454a91c99fe59fb4c6f4ca762f0c6e1b35aaf0ed4390d1b22de"} Dec 07 16:19:33 crc kubenswrapper[4716]: I1207 16:19:32.999638 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f9b78866d-9g9kv" event={"ID":"d845ae0b-5d45-4021-a1e9-4b124298b65b","Type":"ContainerStarted","Data":"cf541c59f23a5666cf2aab25e0a847674c886ca515e78700df3a4760e8774504"} Dec 07 16:19:33 crc kubenswrapper[4716]: I1207 16:19:33.050487 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7d76599897-4mhqt" podStartSLOduration=3.809968697 podStartE2EDuration="30.050471326s" podCreationTimestamp="2025-12-07 16:19:03 +0000 UTC" firstStartedPulling="2025-12-07 16:19:04.927823699 +0000 UTC m=+1007.618108631" lastFinishedPulling="2025-12-07 16:19:31.168326318 +0000 UTC m=+1033.858611260" observedRunningTime="2025-12-07 16:19:33.044858544 +0000 UTC m=+1035.735143456" watchObservedRunningTime="2025-12-07 16:19:33.050471326 +0000 UTC m=+1035.740756238" Dec 07 16:19:33 crc kubenswrapper[4716]: I1207 16:19:33.092309 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-86c7567d4-99rx9" event={"ID":"c291e114-7940-46bc-91d3-a8fa256549ff","Type":"ContainerStarted","Data":"00893cd59537178a447a7d7ca931b0f9922afd0ddd98cb384ea1435f3b09a2f9"} Dec 07 16:19:33 crc kubenswrapper[4716]: I1207 16:19:33.098142 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d8b488bdf-fvlzm" event={"ID":"d7db6aff-6831-49ea-9d51-24ec9ac40a88","Type":"ContainerStarted","Data":"e49f645f546db04a604348b06b91c87e2a6645528ec78ab475e94e276e8c3f6b"} Dec 07 16:19:33 crc kubenswrapper[4716]: I1207 16:19:33.098293 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7d8b488bdf-fvlzm" podUID="d7db6aff-6831-49ea-9d51-24ec9ac40a88" containerName="horizon-log" containerID="cri-o://d9c99d90e0f76afdd5481a03c49c1a155a22721530d081f1e90302e0f9d9fc9d" gracePeriod=30 Dec 07 16:19:33 crc kubenswrapper[4716]: I1207 16:19:33.098378 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7d8b488bdf-fvlzm" podUID="d7db6aff-6831-49ea-9d51-24ec9ac40a88" containerName="horizon" containerID="cri-o://e49f645f546db04a604348b06b91c87e2a6645528ec78ab475e94e276e8c3f6b" gracePeriod=30 Dec 07 16:19:33 crc kubenswrapper[4716]: I1207 16:19:33.116569 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-rtqpb"] Dec 07 16:19:33 crc kubenswrapper[4716]: I1207 16:19:33.127635 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-86c7567d4-99rx9" podStartSLOduration=23.127611873 podStartE2EDuration="23.127611873s" podCreationTimestamp="2025-12-07 16:19:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:19:33.111428013 +0000 UTC m=+1035.801712915" watchObservedRunningTime="2025-12-07 16:19:33.127611873 +0000 UTC m=+1035.817896785" Dec 07 16:19:33 crc kubenswrapper[4716]: I1207 16:19:33.144016 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7d8b488bdf-fvlzm" podStartSLOduration=6.633912467 podStartE2EDuration="31.143999669s" podCreationTimestamp="2025-12-07 16:19:02 +0000 UTC" firstStartedPulling="2025-12-07 16:19:04.890423402 +0000 UTC m=+1007.580708314" lastFinishedPulling="2025-12-07 16:19:29.400510604 +0000 UTC m=+1032.090795516" observedRunningTime="2025-12-07 16:19:33.135097507 +0000 UTC m=+1035.825382419" watchObservedRunningTime="2025-12-07 16:19:33.143999669 +0000 UTC m=+1035.834284581" Dec 07 16:19:33 crc kubenswrapper[4716]: I1207 16:19:33.419022 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-69d87fd46d-9k5rq"] Dec 07 16:19:34 crc kubenswrapper[4716]: I1207 16:19:34.119920 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-86c7567d4-99rx9" event={"ID":"c291e114-7940-46bc-91d3-a8fa256549ff","Type":"ContainerStarted","Data":"bc89298b4c5f29666ca1176ee3a6613d8d2303c5623d2bec2695ed37ae7b6619"} Dec 07 16:19:34 crc kubenswrapper[4716]: I1207 16:19:34.123881 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"afba3ceb-86bf-4aaf-8ac3-b58ab411850c","Type":"ContainerStarted","Data":"7e62702f6546007c2a5d5db6138318f092dfcf6f6759db6ed04b36ce56729518"} Dec 07 16:19:34 crc kubenswrapper[4716]: I1207 16:19:34.127919 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5eca408c-c86b-4dd6-a1f9-81ef16e12da4","Type":"ContainerStarted","Data":"887f5cff319603ea4d4794148d1a5340881877c3bc68e5d25f280bb442fe7459"} Dec 07 16:19:34 crc kubenswrapper[4716]: I1207 16:19:34.134677 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-45d84" event={"ID":"d103f298-33f8-49df-894b-50b813660371","Type":"ContainerStarted","Data":"30ab4927e99fb057a7b46218e65f98171201ab712c6f11193b31bc9789eb773a"} Dec 07 16:19:34 crc kubenswrapper[4716]: I1207 16:19:34.140519 4716 generic.go:334] "Generic (PLEG): container finished" podID="8b1a41e8-004e-4562-a8aa-79e422e11e83" containerID="0c15aa0ddf082834e5d336b546d2ca5f250cd7d2f66dea9b488bd3cf06fc2cac" exitCode=0 Dec 07 16:19:34 crc kubenswrapper[4716]: I1207 16:19:34.140576 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-rtqpb" event={"ID":"8b1a41e8-004e-4562-a8aa-79e422e11e83","Type":"ContainerDied","Data":"0c15aa0ddf082834e5d336b546d2ca5f250cd7d2f66dea9b488bd3cf06fc2cac"} Dec 07 16:19:34 crc kubenswrapper[4716]: I1207 16:19:34.140594 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-rtqpb" event={"ID":"8b1a41e8-004e-4562-a8aa-79e422e11e83","Type":"ContainerStarted","Data":"b8e11dd521785c2f3af96ea7a7ead0c4288d890f27725fba910d532cfb3c6209"} Dec 07 16:19:34 crc kubenswrapper[4716]: I1207 16:19:34.144064 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-qqklx" event={"ID":"6354ab8c-f35e-450f-96f9-8e305f778a54","Type":"ContainerStarted","Data":"ee99ded3785807ca8c912daa4b281fb89f2e8b61332ef14490a3f3d6d8909f5f"} Dec 07 16:19:34 crc kubenswrapper[4716]: I1207 16:19:34.152679 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f9b78866d-9g9kv" event={"ID":"d845ae0b-5d45-4021-a1e9-4b124298b65b","Type":"ContainerStarted","Data":"4afd39acaf476650f9b0c7dda10ca8b1b81c2b3dc196dba3577567e11a863999"} Dec 07 16:19:34 crc kubenswrapper[4716]: I1207 16:19:34.161359 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-69d87fd46d-9k5rq" event={"ID":"8b125bd6-f08f-442c-96a8-b903160be13a","Type":"ContainerStarted","Data":"9e0f9b2ae46383d32e78b438d5bf9227170de493caed859e596aafea34071033"} Dec 07 16:19:34 crc kubenswrapper[4716]: I1207 16:19:34.161407 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-69d87fd46d-9k5rq" event={"ID":"8b125bd6-f08f-442c-96a8-b903160be13a","Type":"ContainerStarted","Data":"c00c5a8c50e26dad3fc07ff6ed5d0663b13bd6897c9a5e4c05fae5699523e678"} Dec 07 16:19:34 crc kubenswrapper[4716]: I1207 16:19:34.161416 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-69d87fd46d-9k5rq" event={"ID":"8b125bd6-f08f-442c-96a8-b903160be13a","Type":"ContainerStarted","Data":"b103baf8f60bf57fbf5072b72913996c5f5704ba2a74a093f2ea7f47dfced9cb"} Dec 07 16:19:34 crc kubenswrapper[4716]: I1207 16:19:34.162253 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-69d87fd46d-9k5rq" Dec 07 16:19:34 crc kubenswrapper[4716]: I1207 16:19:34.175220 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-45d84" podStartSLOduration=16.175202333 podStartE2EDuration="16.175202333s" podCreationTimestamp="2025-12-07 16:19:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:19:34.156511345 +0000 UTC m=+1036.846796257" watchObservedRunningTime="2025-12-07 16:19:34.175202333 +0000 UTC m=+1036.865487245" Dec 07 16:19:34 crc kubenswrapper[4716]: I1207 16:19:34.188405 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7d76599897-4mhqt" Dec 07 16:19:34 crc kubenswrapper[4716]: I1207 16:19:34.262422 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-qqklx" podStartSLOduration=5.80635288 podStartE2EDuration="33.262388672s" podCreationTimestamp="2025-12-07 16:19:01 +0000 UTC" firstStartedPulling="2025-12-07 16:19:04.868491976 +0000 UTC m=+1007.558776888" lastFinishedPulling="2025-12-07 16:19:32.324527768 +0000 UTC m=+1035.014812680" observedRunningTime="2025-12-07 16:19:34.253490081 +0000 UTC m=+1036.943774993" watchObservedRunningTime="2025-12-07 16:19:34.262388672 +0000 UTC m=+1036.952673574" Dec 07 16:19:34 crc kubenswrapper[4716]: I1207 16:19:34.265865 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7f9b78866d-9g9kv" podStartSLOduration=24.265854856 podStartE2EDuration="24.265854856s" podCreationTimestamp="2025-12-07 16:19:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:19:34.227681249 +0000 UTC m=+1036.917966161" watchObservedRunningTime="2025-12-07 16:19:34.265854856 +0000 UTC m=+1036.956139768" Dec 07 16:19:34 crc kubenswrapper[4716]: I1207 16:19:34.300408 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-69d87fd46d-9k5rq" podStartSLOduration=2.300385415 podStartE2EDuration="2.300385415s" podCreationTimestamp="2025-12-07 16:19:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:19:34.277642306 +0000 UTC m=+1036.967927218" watchObservedRunningTime="2025-12-07 16:19:34.300385415 +0000 UTC m=+1036.990670317" Dec 07 16:19:34 crc kubenswrapper[4716]: I1207 16:19:34.548640 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-bd7b79585-wwp6b"] Dec 07 16:19:34 crc kubenswrapper[4716]: I1207 16:19:34.550176 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-bd7b79585-wwp6b" Dec 07 16:19:34 crc kubenswrapper[4716]: I1207 16:19:34.552996 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Dec 07 16:19:34 crc kubenswrapper[4716]: I1207 16:19:34.553184 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Dec 07 16:19:34 crc kubenswrapper[4716]: I1207 16:19:34.559861 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-bd7b79585-wwp6b"] Dec 07 16:19:34 crc kubenswrapper[4716]: I1207 16:19:34.587958 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/dd279760-9178-4ad8-ae1e-ae4e6fab3f3c-config\") pod \"neutron-bd7b79585-wwp6b\" (UID: \"dd279760-9178-4ad8-ae1e-ae4e6fab3f3c\") " pod="openstack/neutron-bd7b79585-wwp6b" Dec 07 16:19:34 crc kubenswrapper[4716]: I1207 16:19:34.587996 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/dd279760-9178-4ad8-ae1e-ae4e6fab3f3c-httpd-config\") pod \"neutron-bd7b79585-wwp6b\" (UID: \"dd279760-9178-4ad8-ae1e-ae4e6fab3f3c\") " pod="openstack/neutron-bd7b79585-wwp6b" Dec 07 16:19:34 crc kubenswrapper[4716]: I1207 16:19:34.588033 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd279760-9178-4ad8-ae1e-ae4e6fab3f3c-ovndb-tls-certs\") pod \"neutron-bd7b79585-wwp6b\" (UID: \"dd279760-9178-4ad8-ae1e-ae4e6fab3f3c\") " pod="openstack/neutron-bd7b79585-wwp6b" Dec 07 16:19:34 crc kubenswrapper[4716]: I1207 16:19:34.588088 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd279760-9178-4ad8-ae1e-ae4e6fab3f3c-public-tls-certs\") pod \"neutron-bd7b79585-wwp6b\" (UID: \"dd279760-9178-4ad8-ae1e-ae4e6fab3f3c\") " pod="openstack/neutron-bd7b79585-wwp6b" Dec 07 16:19:34 crc kubenswrapper[4716]: I1207 16:19:34.588131 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jtk26\" (UniqueName: \"kubernetes.io/projected/dd279760-9178-4ad8-ae1e-ae4e6fab3f3c-kube-api-access-jtk26\") pod \"neutron-bd7b79585-wwp6b\" (UID: \"dd279760-9178-4ad8-ae1e-ae4e6fab3f3c\") " pod="openstack/neutron-bd7b79585-wwp6b" Dec 07 16:19:34 crc kubenswrapper[4716]: I1207 16:19:34.588151 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd279760-9178-4ad8-ae1e-ae4e6fab3f3c-internal-tls-certs\") pod \"neutron-bd7b79585-wwp6b\" (UID: \"dd279760-9178-4ad8-ae1e-ae4e6fab3f3c\") " pod="openstack/neutron-bd7b79585-wwp6b" Dec 07 16:19:34 crc kubenswrapper[4716]: I1207 16:19:34.588190 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd279760-9178-4ad8-ae1e-ae4e6fab3f3c-combined-ca-bundle\") pod \"neutron-bd7b79585-wwp6b\" (UID: \"dd279760-9178-4ad8-ae1e-ae4e6fab3f3c\") " pod="openstack/neutron-bd7b79585-wwp6b" Dec 07 16:19:34 crc kubenswrapper[4716]: I1207 16:19:34.690445 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/dd279760-9178-4ad8-ae1e-ae4e6fab3f3c-config\") pod \"neutron-bd7b79585-wwp6b\" (UID: \"dd279760-9178-4ad8-ae1e-ae4e6fab3f3c\") " pod="openstack/neutron-bd7b79585-wwp6b" Dec 07 16:19:34 crc kubenswrapper[4716]: I1207 16:19:34.690737 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/dd279760-9178-4ad8-ae1e-ae4e6fab3f3c-httpd-config\") pod \"neutron-bd7b79585-wwp6b\" (UID: \"dd279760-9178-4ad8-ae1e-ae4e6fab3f3c\") " pod="openstack/neutron-bd7b79585-wwp6b" Dec 07 16:19:34 crc kubenswrapper[4716]: I1207 16:19:34.690869 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd279760-9178-4ad8-ae1e-ae4e6fab3f3c-ovndb-tls-certs\") pod \"neutron-bd7b79585-wwp6b\" (UID: \"dd279760-9178-4ad8-ae1e-ae4e6fab3f3c\") " pod="openstack/neutron-bd7b79585-wwp6b" Dec 07 16:19:34 crc kubenswrapper[4716]: I1207 16:19:34.690986 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd279760-9178-4ad8-ae1e-ae4e6fab3f3c-public-tls-certs\") pod \"neutron-bd7b79585-wwp6b\" (UID: \"dd279760-9178-4ad8-ae1e-ae4e6fab3f3c\") " pod="openstack/neutron-bd7b79585-wwp6b" Dec 07 16:19:34 crc kubenswrapper[4716]: I1207 16:19:34.691116 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jtk26\" (UniqueName: \"kubernetes.io/projected/dd279760-9178-4ad8-ae1e-ae4e6fab3f3c-kube-api-access-jtk26\") pod \"neutron-bd7b79585-wwp6b\" (UID: \"dd279760-9178-4ad8-ae1e-ae4e6fab3f3c\") " pod="openstack/neutron-bd7b79585-wwp6b" Dec 07 16:19:34 crc kubenswrapper[4716]: I1207 16:19:34.691544 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd279760-9178-4ad8-ae1e-ae4e6fab3f3c-internal-tls-certs\") pod \"neutron-bd7b79585-wwp6b\" (UID: \"dd279760-9178-4ad8-ae1e-ae4e6fab3f3c\") " pod="openstack/neutron-bd7b79585-wwp6b" Dec 07 16:19:34 crc kubenswrapper[4716]: I1207 16:19:34.691663 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd279760-9178-4ad8-ae1e-ae4e6fab3f3c-combined-ca-bundle\") pod \"neutron-bd7b79585-wwp6b\" (UID: \"dd279760-9178-4ad8-ae1e-ae4e6fab3f3c\") " pod="openstack/neutron-bd7b79585-wwp6b" Dec 07 16:19:34 crc kubenswrapper[4716]: I1207 16:19:34.696494 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd279760-9178-4ad8-ae1e-ae4e6fab3f3c-ovndb-tls-certs\") pod \"neutron-bd7b79585-wwp6b\" (UID: \"dd279760-9178-4ad8-ae1e-ae4e6fab3f3c\") " pod="openstack/neutron-bd7b79585-wwp6b" Dec 07 16:19:34 crc kubenswrapper[4716]: I1207 16:19:34.697420 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/dd279760-9178-4ad8-ae1e-ae4e6fab3f3c-config\") pod \"neutron-bd7b79585-wwp6b\" (UID: \"dd279760-9178-4ad8-ae1e-ae4e6fab3f3c\") " pod="openstack/neutron-bd7b79585-wwp6b" Dec 07 16:19:34 crc kubenswrapper[4716]: I1207 16:19:34.697613 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd279760-9178-4ad8-ae1e-ae4e6fab3f3c-combined-ca-bundle\") pod \"neutron-bd7b79585-wwp6b\" (UID: \"dd279760-9178-4ad8-ae1e-ae4e6fab3f3c\") " pod="openstack/neutron-bd7b79585-wwp6b" Dec 07 16:19:34 crc kubenswrapper[4716]: I1207 16:19:34.697968 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd279760-9178-4ad8-ae1e-ae4e6fab3f3c-public-tls-certs\") pod \"neutron-bd7b79585-wwp6b\" (UID: \"dd279760-9178-4ad8-ae1e-ae4e6fab3f3c\") " pod="openstack/neutron-bd7b79585-wwp6b" Dec 07 16:19:34 crc kubenswrapper[4716]: I1207 16:19:34.698250 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/dd279760-9178-4ad8-ae1e-ae4e6fab3f3c-httpd-config\") pod \"neutron-bd7b79585-wwp6b\" (UID: \"dd279760-9178-4ad8-ae1e-ae4e6fab3f3c\") " pod="openstack/neutron-bd7b79585-wwp6b" Dec 07 16:19:34 crc kubenswrapper[4716]: I1207 16:19:34.701137 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd279760-9178-4ad8-ae1e-ae4e6fab3f3c-internal-tls-certs\") pod \"neutron-bd7b79585-wwp6b\" (UID: \"dd279760-9178-4ad8-ae1e-ae4e6fab3f3c\") " pod="openstack/neutron-bd7b79585-wwp6b" Dec 07 16:19:34 crc kubenswrapper[4716]: I1207 16:19:34.738066 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jtk26\" (UniqueName: \"kubernetes.io/projected/dd279760-9178-4ad8-ae1e-ae4e6fab3f3c-kube-api-access-jtk26\") pod \"neutron-bd7b79585-wwp6b\" (UID: \"dd279760-9178-4ad8-ae1e-ae4e6fab3f3c\") " pod="openstack/neutron-bd7b79585-wwp6b" Dec 07 16:19:34 crc kubenswrapper[4716]: I1207 16:19:34.868495 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-bd7b79585-wwp6b" Dec 07 16:19:35 crc kubenswrapper[4716]: I1207 16:19:35.174200 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"afba3ceb-86bf-4aaf-8ac3-b58ab411850c","Type":"ContainerStarted","Data":"9d342a91e7dc00c4eac8190839f9ea8a4d0e4c05fb04b03355f7cb519bf04840"} Dec 07 16:19:35 crc kubenswrapper[4716]: I1207 16:19:35.182568 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5eca408c-c86b-4dd6-a1f9-81ef16e12da4","Type":"ContainerStarted","Data":"7a9e2b9815207ee0080821a823fef456aa90e0a1cfa6b17abc1d59687be33655"} Dec 07 16:19:35 crc kubenswrapper[4716]: I1207 16:19:35.241944 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=18.241922342 podStartE2EDuration="18.241922342s" podCreationTimestamp="2025-12-07 16:19:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:19:35.203241581 +0000 UTC m=+1037.893526493" watchObservedRunningTime="2025-12-07 16:19:35.241922342 +0000 UTC m=+1037.932207254" Dec 07 16:19:35 crc kubenswrapper[4716]: I1207 16:19:35.247762 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=18.24773902 podStartE2EDuration="18.24773902s" podCreationTimestamp="2025-12-07 16:19:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:19:35.239889857 +0000 UTC m=+1037.930174769" watchObservedRunningTime="2025-12-07 16:19:35.24773902 +0000 UTC m=+1037.938023922" Dec 07 16:19:37 crc kubenswrapper[4716]: I1207 16:19:37.983800 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 07 16:19:37 crc kubenswrapper[4716]: I1207 16:19:37.984579 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 07 16:19:38 crc kubenswrapper[4716]: I1207 16:19:38.018325 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 07 16:19:38 crc kubenswrapper[4716]: I1207 16:19:38.018370 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 07 16:19:38 crc kubenswrapper[4716]: I1207 16:19:38.084428 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 07 16:19:38 crc kubenswrapper[4716]: I1207 16:19:38.084865 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 07 16:19:38 crc kubenswrapper[4716]: I1207 16:19:38.100549 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 07 16:19:38 crc kubenswrapper[4716]: I1207 16:19:38.116514 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 07 16:19:38 crc kubenswrapper[4716]: I1207 16:19:38.216646 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 07 16:19:38 crc kubenswrapper[4716]: I1207 16:19:38.216680 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 07 16:19:38 crc kubenswrapper[4716]: I1207 16:19:38.216691 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 07 16:19:38 crc kubenswrapper[4716]: I1207 16:19:38.216712 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 07 16:19:38 crc kubenswrapper[4716]: I1207 16:19:38.499824 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-bd7b79585-wwp6b"] Dec 07 16:19:38 crc kubenswrapper[4716]: W1207 16:19:38.507353 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddd279760_9178_4ad8_ae1e_ae4e6fab3f3c.slice/crio-59d55ca0706c4286bee3a0c088300e2cf564a4fd76462e27b44ddbef0c98f390 WatchSource:0}: Error finding container 59d55ca0706c4286bee3a0c088300e2cf564a4fd76462e27b44ddbef0c98f390: Status 404 returned error can't find the container with id 59d55ca0706c4286bee3a0c088300e2cf564a4fd76462e27b44ddbef0c98f390 Dec 07 16:19:39 crc kubenswrapper[4716]: I1207 16:19:39.223466 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-bd7b79585-wwp6b" event={"ID":"dd279760-9178-4ad8-ae1e-ae4e6fab3f3c","Type":"ContainerStarted","Data":"357d6908f09329f68f8e2f8e6a89d8e33077b1d70a85da5075baa16537a58fe6"} Dec 07 16:19:39 crc kubenswrapper[4716]: I1207 16:19:39.223898 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-bd7b79585-wwp6b" event={"ID":"dd279760-9178-4ad8-ae1e-ae4e6fab3f3c","Type":"ContainerStarted","Data":"8cd84d7b745beba64eead2e2e7d6ee613693018dd2e98ec54032c0ad334e25f5"} Dec 07 16:19:39 crc kubenswrapper[4716]: I1207 16:19:39.223909 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-bd7b79585-wwp6b" event={"ID":"dd279760-9178-4ad8-ae1e-ae4e6fab3f3c","Type":"ContainerStarted","Data":"59d55ca0706c4286bee3a0c088300e2cf564a4fd76462e27b44ddbef0c98f390"} Dec 07 16:19:39 crc kubenswrapper[4716]: I1207 16:19:39.223924 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-bd7b79585-wwp6b" Dec 07 16:19:39 crc kubenswrapper[4716]: I1207 16:19:39.226031 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-rtqpb" event={"ID":"8b1a41e8-004e-4562-a8aa-79e422e11e83","Type":"ContainerStarted","Data":"a37b3fb034928ee7bc75507f7ff97f1249c5e124632b04b9390037efa09b3d1d"} Dec 07 16:19:39 crc kubenswrapper[4716]: I1207 16:19:39.226124 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-55f844cf75-rtqpb" Dec 07 16:19:39 crc kubenswrapper[4716]: I1207 16:19:39.229145 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2ee00bf0-1412-4fca-a570-6600db549923","Type":"ContainerStarted","Data":"411ac652ee7365d2f2c78ae6705384e1dadf6242d65ea1cab9c0a5600032ff1d"} Dec 07 16:19:39 crc kubenswrapper[4716]: I1207 16:19:39.254200 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-bd7b79585-wwp6b" podStartSLOduration=5.254180066 podStartE2EDuration="5.254180066s" podCreationTimestamp="2025-12-07 16:19:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:19:39.248644856 +0000 UTC m=+1041.938929768" watchObservedRunningTime="2025-12-07 16:19:39.254180066 +0000 UTC m=+1041.944464978" Dec 07 16:19:39 crc kubenswrapper[4716]: I1207 16:19:39.279687 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-55f844cf75-rtqpb" podStartSLOduration=8.279670336 podStartE2EDuration="8.279670336s" podCreationTimestamp="2025-12-07 16:19:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:19:39.278033511 +0000 UTC m=+1041.968318423" watchObservedRunningTime="2025-12-07 16:19:39.279670336 +0000 UTC m=+1041.969955248" Dec 07 16:19:40 crc kubenswrapper[4716]: I1207 16:19:40.248693 4716 generic.go:334] "Generic (PLEG): container finished" podID="d103f298-33f8-49df-894b-50b813660371" containerID="30ab4927e99fb057a7b46218e65f98171201ab712c6f11193b31bc9789eb773a" exitCode=0 Dec 07 16:19:40 crc kubenswrapper[4716]: I1207 16:19:40.248802 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-45d84" event={"ID":"d103f298-33f8-49df-894b-50b813660371","Type":"ContainerDied","Data":"30ab4927e99fb057a7b46218e65f98171201ab712c6f11193b31bc9789eb773a"} Dec 07 16:19:40 crc kubenswrapper[4716]: I1207 16:19:40.251491 4716 generic.go:334] "Generic (PLEG): container finished" podID="6354ab8c-f35e-450f-96f9-8e305f778a54" containerID="ee99ded3785807ca8c912daa4b281fb89f2e8b61332ef14490a3f3d6d8909f5f" exitCode=0 Dec 07 16:19:40 crc kubenswrapper[4716]: I1207 16:19:40.251627 4716 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 07 16:19:40 crc kubenswrapper[4716]: I1207 16:19:40.251655 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-qqklx" event={"ID":"6354ab8c-f35e-450f-96f9-8e305f778a54","Type":"ContainerDied","Data":"ee99ded3785807ca8c912daa4b281fb89f2e8b61332ef14490a3f3d6d8909f5f"} Dec 07 16:19:40 crc kubenswrapper[4716]: I1207 16:19:40.717453 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-86c7567d4-99rx9" Dec 07 16:19:40 crc kubenswrapper[4716]: I1207 16:19:40.717521 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-86c7567d4-99rx9" Dec 07 16:19:40 crc kubenswrapper[4716]: I1207 16:19:40.787606 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7f9b78866d-9g9kv" Dec 07 16:19:40 crc kubenswrapper[4716]: I1207 16:19:40.788431 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7f9b78866d-9g9kv" Dec 07 16:19:41 crc kubenswrapper[4716]: I1207 16:19:41.365340 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 07 16:19:41 crc kubenswrapper[4716]: I1207 16:19:41.365903 4716 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 07 16:19:41 crc kubenswrapper[4716]: I1207 16:19:41.865493 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-qqklx" Dec 07 16:19:41 crc kubenswrapper[4716]: I1207 16:19:41.968904 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6354ab8c-f35e-450f-96f9-8e305f778a54-combined-ca-bundle\") pod \"6354ab8c-f35e-450f-96f9-8e305f778a54\" (UID: \"6354ab8c-f35e-450f-96f9-8e305f778a54\") " Dec 07 16:19:41 crc kubenswrapper[4716]: I1207 16:19:41.968952 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-72mp2\" (UniqueName: \"kubernetes.io/projected/6354ab8c-f35e-450f-96f9-8e305f778a54-kube-api-access-72mp2\") pod \"6354ab8c-f35e-450f-96f9-8e305f778a54\" (UID: \"6354ab8c-f35e-450f-96f9-8e305f778a54\") " Dec 07 16:19:41 crc kubenswrapper[4716]: I1207 16:19:41.969118 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6354ab8c-f35e-450f-96f9-8e305f778a54-scripts\") pod \"6354ab8c-f35e-450f-96f9-8e305f778a54\" (UID: \"6354ab8c-f35e-450f-96f9-8e305f778a54\") " Dec 07 16:19:41 crc kubenswrapper[4716]: I1207 16:19:41.969189 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6354ab8c-f35e-450f-96f9-8e305f778a54-config-data\") pod \"6354ab8c-f35e-450f-96f9-8e305f778a54\" (UID: \"6354ab8c-f35e-450f-96f9-8e305f778a54\") " Dec 07 16:19:41 crc kubenswrapper[4716]: I1207 16:19:41.969236 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6354ab8c-f35e-450f-96f9-8e305f778a54-logs\") pod \"6354ab8c-f35e-450f-96f9-8e305f778a54\" (UID: \"6354ab8c-f35e-450f-96f9-8e305f778a54\") " Dec 07 16:19:41 crc kubenswrapper[4716]: I1207 16:19:41.983559 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6354ab8c-f35e-450f-96f9-8e305f778a54-logs" (OuterVolumeSpecName: "logs") pod "6354ab8c-f35e-450f-96f9-8e305f778a54" (UID: "6354ab8c-f35e-450f-96f9-8e305f778a54"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.001966 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6354ab8c-f35e-450f-96f9-8e305f778a54-scripts" (OuterVolumeSpecName: "scripts") pod "6354ab8c-f35e-450f-96f9-8e305f778a54" (UID: "6354ab8c-f35e-450f-96f9-8e305f778a54"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.011917 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6354ab8c-f35e-450f-96f9-8e305f778a54-kube-api-access-72mp2" (OuterVolumeSpecName: "kube-api-access-72mp2") pod "6354ab8c-f35e-450f-96f9-8e305f778a54" (UID: "6354ab8c-f35e-450f-96f9-8e305f778a54"). InnerVolumeSpecName "kube-api-access-72mp2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.033315 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6354ab8c-f35e-450f-96f9-8e305f778a54-config-data" (OuterVolumeSpecName: "config-data") pod "6354ab8c-f35e-450f-96f9-8e305f778a54" (UID: "6354ab8c-f35e-450f-96f9-8e305f778a54"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.034996 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.057550 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6354ab8c-f35e-450f-96f9-8e305f778a54-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6354ab8c-f35e-450f-96f9-8e305f778a54" (UID: "6354ab8c-f35e-450f-96f9-8e305f778a54"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.073807 4716 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6354ab8c-f35e-450f-96f9-8e305f778a54-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.073839 4716 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6354ab8c-f35e-450f-96f9-8e305f778a54-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.073848 4716 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6354ab8c-f35e-450f-96f9-8e305f778a54-logs\") on node \"crc\" DevicePath \"\"" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.073861 4716 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6354ab8c-f35e-450f-96f9-8e305f778a54-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.073871 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-72mp2\" (UniqueName: \"kubernetes.io/projected/6354ab8c-f35e-450f-96f9-8e305f778a54-kube-api-access-72mp2\") on node \"crc\" DevicePath \"\"" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.078759 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-45d84" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.131065 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.174594 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d103f298-33f8-49df-894b-50b813660371-combined-ca-bundle\") pod \"d103f298-33f8-49df-894b-50b813660371\" (UID: \"d103f298-33f8-49df-894b-50b813660371\") " Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.174889 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d103f298-33f8-49df-894b-50b813660371-config-data\") pod \"d103f298-33f8-49df-894b-50b813660371\" (UID: \"d103f298-33f8-49df-894b-50b813660371\") " Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.175063 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d103f298-33f8-49df-894b-50b813660371-fernet-keys\") pod \"d103f298-33f8-49df-894b-50b813660371\" (UID: \"d103f298-33f8-49df-894b-50b813660371\") " Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.175220 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qbzs2\" (UniqueName: \"kubernetes.io/projected/d103f298-33f8-49df-894b-50b813660371-kube-api-access-qbzs2\") pod \"d103f298-33f8-49df-894b-50b813660371\" (UID: \"d103f298-33f8-49df-894b-50b813660371\") " Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.175298 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d103f298-33f8-49df-894b-50b813660371-credential-keys\") pod \"d103f298-33f8-49df-894b-50b813660371\" (UID: \"d103f298-33f8-49df-894b-50b813660371\") " Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.175421 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d103f298-33f8-49df-894b-50b813660371-scripts\") pod \"d103f298-33f8-49df-894b-50b813660371\" (UID: \"d103f298-33f8-49df-894b-50b813660371\") " Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.195221 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d103f298-33f8-49df-894b-50b813660371-scripts" (OuterVolumeSpecName: "scripts") pod "d103f298-33f8-49df-894b-50b813660371" (UID: "d103f298-33f8-49df-894b-50b813660371"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.207318 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d103f298-33f8-49df-894b-50b813660371-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "d103f298-33f8-49df-894b-50b813660371" (UID: "d103f298-33f8-49df-894b-50b813660371"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.207336 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d103f298-33f8-49df-894b-50b813660371-kube-api-access-qbzs2" (OuterVolumeSpecName: "kube-api-access-qbzs2") pod "d103f298-33f8-49df-894b-50b813660371" (UID: "d103f298-33f8-49df-894b-50b813660371"). InnerVolumeSpecName "kube-api-access-qbzs2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.214210 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d103f298-33f8-49df-894b-50b813660371-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "d103f298-33f8-49df-894b-50b813660371" (UID: "d103f298-33f8-49df-894b-50b813660371"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.227726 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d103f298-33f8-49df-894b-50b813660371-config-data" (OuterVolumeSpecName: "config-data") pod "d103f298-33f8-49df-894b-50b813660371" (UID: "d103f298-33f8-49df-894b-50b813660371"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.230266 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d103f298-33f8-49df-894b-50b813660371-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d103f298-33f8-49df-894b-50b813660371" (UID: "d103f298-33f8-49df-894b-50b813660371"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.277853 4716 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d103f298-33f8-49df-894b-50b813660371-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.278235 4716 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d103f298-33f8-49df-894b-50b813660371-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.278249 4716 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d103f298-33f8-49df-894b-50b813660371-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.278258 4716 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d103f298-33f8-49df-894b-50b813660371-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.278267 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qbzs2\" (UniqueName: \"kubernetes.io/projected/d103f298-33f8-49df-894b-50b813660371-kube-api-access-qbzs2\") on node \"crc\" DevicePath \"\"" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.278275 4716 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d103f298-33f8-49df-894b-50b813660371-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.288996 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-qqklx" event={"ID":"6354ab8c-f35e-450f-96f9-8e305f778a54","Type":"ContainerDied","Data":"a615881d89bc426d8e5e0ae90cf47c8e936a3781f1552654e42c25608b198950"} Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.289052 4716 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a615881d89bc426d8e5e0ae90cf47c8e936a3781f1552654e42c25608b198950" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.289177 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-qqklx" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.308967 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-45d84" event={"ID":"d103f298-33f8-49df-894b-50b813660371","Type":"ContainerDied","Data":"3a713685c633539179ae87831baab5d3dfb90504cd091959ed09697ea02e7224"} Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.309029 4716 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3a713685c633539179ae87831baab5d3dfb90504cd091959ed09697ea02e7224" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.309263 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-45d84" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.431672 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-6c4445878c-bnpkh"] Dec 07 16:19:42 crc kubenswrapper[4716]: E1207 16:19:42.432169 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d103f298-33f8-49df-894b-50b813660371" containerName="keystone-bootstrap" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.432185 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="d103f298-33f8-49df-894b-50b813660371" containerName="keystone-bootstrap" Dec 07 16:19:42 crc kubenswrapper[4716]: E1207 16:19:42.432203 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6354ab8c-f35e-450f-96f9-8e305f778a54" containerName="placement-db-sync" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.432210 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="6354ab8c-f35e-450f-96f9-8e305f778a54" containerName="placement-db-sync" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.432431 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="d103f298-33f8-49df-894b-50b813660371" containerName="keystone-bootstrap" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.432467 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="6354ab8c-f35e-450f-96f9-8e305f778a54" containerName="placement-db-sync" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.433296 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6c4445878c-bnpkh" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.443141 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-746cf47744-w9nm7"] Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.445500 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-746cf47744-w9nm7" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.453495 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.453717 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.453829 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.453934 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.454044 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.454129 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-746cf47744-w9nm7"] Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.465689 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-lhq7l" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.470242 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-6c4445878c-bnpkh"] Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.471058 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-dtr4s" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.471355 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.471566 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.471704 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.475018 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.594105 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ade9101f-5435-455f-807c-d277918cbb46-credential-keys\") pod \"keystone-6c4445878c-bnpkh\" (UID: \"ade9101f-5435-455f-807c-d277918cbb46\") " pod="openstack/keystone-6c4445878c-bnpkh" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.594156 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ade9101f-5435-455f-807c-d277918cbb46-scripts\") pod \"keystone-6c4445878c-bnpkh\" (UID: \"ade9101f-5435-455f-807c-d277918cbb46\") " pod="openstack/keystone-6c4445878c-bnpkh" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.594177 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ade9101f-5435-455f-807c-d277918cbb46-combined-ca-bundle\") pod \"keystone-6c4445878c-bnpkh\" (UID: \"ade9101f-5435-455f-807c-d277918cbb46\") " pod="openstack/keystone-6c4445878c-bnpkh" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.594206 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0320eb6a-f473-4c4c-ae36-f74080dcdaa5-logs\") pod \"placement-746cf47744-w9nm7\" (UID: \"0320eb6a-f473-4c4c-ae36-f74080dcdaa5\") " pod="openstack/placement-746cf47744-w9nm7" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.594251 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ade9101f-5435-455f-807c-d277918cbb46-fernet-keys\") pod \"keystone-6c4445878c-bnpkh\" (UID: \"ade9101f-5435-455f-807c-d277918cbb46\") " pod="openstack/keystone-6c4445878c-bnpkh" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.594268 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0320eb6a-f473-4c4c-ae36-f74080dcdaa5-public-tls-certs\") pod \"placement-746cf47744-w9nm7\" (UID: \"0320eb6a-f473-4c4c-ae36-f74080dcdaa5\") " pod="openstack/placement-746cf47744-w9nm7" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.594294 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ade9101f-5435-455f-807c-d277918cbb46-internal-tls-certs\") pod \"keystone-6c4445878c-bnpkh\" (UID: \"ade9101f-5435-455f-807c-d277918cbb46\") " pod="openstack/keystone-6c4445878c-bnpkh" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.594322 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0320eb6a-f473-4c4c-ae36-f74080dcdaa5-combined-ca-bundle\") pod \"placement-746cf47744-w9nm7\" (UID: \"0320eb6a-f473-4c4c-ae36-f74080dcdaa5\") " pod="openstack/placement-746cf47744-w9nm7" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.594350 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0320eb6a-f473-4c4c-ae36-f74080dcdaa5-scripts\") pod \"placement-746cf47744-w9nm7\" (UID: \"0320eb6a-f473-4c4c-ae36-f74080dcdaa5\") " pod="openstack/placement-746cf47744-w9nm7" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.594379 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bv7tz\" (UniqueName: \"kubernetes.io/projected/0320eb6a-f473-4c4c-ae36-f74080dcdaa5-kube-api-access-bv7tz\") pod \"placement-746cf47744-w9nm7\" (UID: \"0320eb6a-f473-4c4c-ae36-f74080dcdaa5\") " pod="openstack/placement-746cf47744-w9nm7" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.594398 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0320eb6a-f473-4c4c-ae36-f74080dcdaa5-config-data\") pod \"placement-746cf47744-w9nm7\" (UID: \"0320eb6a-f473-4c4c-ae36-f74080dcdaa5\") " pod="openstack/placement-746cf47744-w9nm7" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.594423 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ade9101f-5435-455f-807c-d277918cbb46-public-tls-certs\") pod \"keystone-6c4445878c-bnpkh\" (UID: \"ade9101f-5435-455f-807c-d277918cbb46\") " pod="openstack/keystone-6c4445878c-bnpkh" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.594444 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fbprg\" (UniqueName: \"kubernetes.io/projected/ade9101f-5435-455f-807c-d277918cbb46-kube-api-access-fbprg\") pod \"keystone-6c4445878c-bnpkh\" (UID: \"ade9101f-5435-455f-807c-d277918cbb46\") " pod="openstack/keystone-6c4445878c-bnpkh" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.594470 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0320eb6a-f473-4c4c-ae36-f74080dcdaa5-internal-tls-certs\") pod \"placement-746cf47744-w9nm7\" (UID: \"0320eb6a-f473-4c4c-ae36-f74080dcdaa5\") " pod="openstack/placement-746cf47744-w9nm7" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.594494 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ade9101f-5435-455f-807c-d277918cbb46-config-data\") pod \"keystone-6c4445878c-bnpkh\" (UID: \"ade9101f-5435-455f-807c-d277918cbb46\") " pod="openstack/keystone-6c4445878c-bnpkh" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.656817 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7d8b488bdf-fvlzm" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.696796 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0320eb6a-f473-4c4c-ae36-f74080dcdaa5-internal-tls-certs\") pod \"placement-746cf47744-w9nm7\" (UID: \"0320eb6a-f473-4c4c-ae36-f74080dcdaa5\") " pod="openstack/placement-746cf47744-w9nm7" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.696879 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ade9101f-5435-455f-807c-d277918cbb46-config-data\") pod \"keystone-6c4445878c-bnpkh\" (UID: \"ade9101f-5435-455f-807c-d277918cbb46\") " pod="openstack/keystone-6c4445878c-bnpkh" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.696934 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ade9101f-5435-455f-807c-d277918cbb46-credential-keys\") pod \"keystone-6c4445878c-bnpkh\" (UID: \"ade9101f-5435-455f-807c-d277918cbb46\") " pod="openstack/keystone-6c4445878c-bnpkh" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.696965 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ade9101f-5435-455f-807c-d277918cbb46-scripts\") pod \"keystone-6c4445878c-bnpkh\" (UID: \"ade9101f-5435-455f-807c-d277918cbb46\") " pod="openstack/keystone-6c4445878c-bnpkh" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.696992 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ade9101f-5435-455f-807c-d277918cbb46-combined-ca-bundle\") pod \"keystone-6c4445878c-bnpkh\" (UID: \"ade9101f-5435-455f-807c-d277918cbb46\") " pod="openstack/keystone-6c4445878c-bnpkh" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.697021 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0320eb6a-f473-4c4c-ae36-f74080dcdaa5-logs\") pod \"placement-746cf47744-w9nm7\" (UID: \"0320eb6a-f473-4c4c-ae36-f74080dcdaa5\") " pod="openstack/placement-746cf47744-w9nm7" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.697059 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ade9101f-5435-455f-807c-d277918cbb46-fernet-keys\") pod \"keystone-6c4445878c-bnpkh\" (UID: \"ade9101f-5435-455f-807c-d277918cbb46\") " pod="openstack/keystone-6c4445878c-bnpkh" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.697112 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0320eb6a-f473-4c4c-ae36-f74080dcdaa5-public-tls-certs\") pod \"placement-746cf47744-w9nm7\" (UID: \"0320eb6a-f473-4c4c-ae36-f74080dcdaa5\") " pod="openstack/placement-746cf47744-w9nm7" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.697145 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ade9101f-5435-455f-807c-d277918cbb46-internal-tls-certs\") pod \"keystone-6c4445878c-bnpkh\" (UID: \"ade9101f-5435-455f-807c-d277918cbb46\") " pod="openstack/keystone-6c4445878c-bnpkh" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.697194 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0320eb6a-f473-4c4c-ae36-f74080dcdaa5-combined-ca-bundle\") pod \"placement-746cf47744-w9nm7\" (UID: \"0320eb6a-f473-4c4c-ae36-f74080dcdaa5\") " pod="openstack/placement-746cf47744-w9nm7" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.697235 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0320eb6a-f473-4c4c-ae36-f74080dcdaa5-scripts\") pod \"placement-746cf47744-w9nm7\" (UID: \"0320eb6a-f473-4c4c-ae36-f74080dcdaa5\") " pod="openstack/placement-746cf47744-w9nm7" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.697272 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bv7tz\" (UniqueName: \"kubernetes.io/projected/0320eb6a-f473-4c4c-ae36-f74080dcdaa5-kube-api-access-bv7tz\") pod \"placement-746cf47744-w9nm7\" (UID: \"0320eb6a-f473-4c4c-ae36-f74080dcdaa5\") " pod="openstack/placement-746cf47744-w9nm7" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.697299 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0320eb6a-f473-4c4c-ae36-f74080dcdaa5-config-data\") pod \"placement-746cf47744-w9nm7\" (UID: \"0320eb6a-f473-4c4c-ae36-f74080dcdaa5\") " pod="openstack/placement-746cf47744-w9nm7" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.697329 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ade9101f-5435-455f-807c-d277918cbb46-public-tls-certs\") pod \"keystone-6c4445878c-bnpkh\" (UID: \"ade9101f-5435-455f-807c-d277918cbb46\") " pod="openstack/keystone-6c4445878c-bnpkh" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.697351 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fbprg\" (UniqueName: \"kubernetes.io/projected/ade9101f-5435-455f-807c-d277918cbb46-kube-api-access-fbprg\") pod \"keystone-6c4445878c-bnpkh\" (UID: \"ade9101f-5435-455f-807c-d277918cbb46\") " pod="openstack/keystone-6c4445878c-bnpkh" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.697600 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0320eb6a-f473-4c4c-ae36-f74080dcdaa5-logs\") pod \"placement-746cf47744-w9nm7\" (UID: \"0320eb6a-f473-4c4c-ae36-f74080dcdaa5\") " pod="openstack/placement-746cf47744-w9nm7" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.703201 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0320eb6a-f473-4c4c-ae36-f74080dcdaa5-internal-tls-certs\") pod \"placement-746cf47744-w9nm7\" (UID: \"0320eb6a-f473-4c4c-ae36-f74080dcdaa5\") " pod="openstack/placement-746cf47744-w9nm7" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.704704 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0320eb6a-f473-4c4c-ae36-f74080dcdaa5-config-data\") pod \"placement-746cf47744-w9nm7\" (UID: \"0320eb6a-f473-4c4c-ae36-f74080dcdaa5\") " pod="openstack/placement-746cf47744-w9nm7" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.704768 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ade9101f-5435-455f-807c-d277918cbb46-combined-ca-bundle\") pod \"keystone-6c4445878c-bnpkh\" (UID: \"ade9101f-5435-455f-807c-d277918cbb46\") " pod="openstack/keystone-6c4445878c-bnpkh" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.707271 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ade9101f-5435-455f-807c-d277918cbb46-config-data\") pod \"keystone-6c4445878c-bnpkh\" (UID: \"ade9101f-5435-455f-807c-d277918cbb46\") " pod="openstack/keystone-6c4445878c-bnpkh" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.707430 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ade9101f-5435-455f-807c-d277918cbb46-credential-keys\") pod \"keystone-6c4445878c-bnpkh\" (UID: \"ade9101f-5435-455f-807c-d277918cbb46\") " pod="openstack/keystone-6c4445878c-bnpkh" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.707862 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ade9101f-5435-455f-807c-d277918cbb46-fernet-keys\") pod \"keystone-6c4445878c-bnpkh\" (UID: \"ade9101f-5435-455f-807c-d277918cbb46\") " pod="openstack/keystone-6c4445878c-bnpkh" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.708613 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0320eb6a-f473-4c4c-ae36-f74080dcdaa5-public-tls-certs\") pod \"placement-746cf47744-w9nm7\" (UID: \"0320eb6a-f473-4c4c-ae36-f74080dcdaa5\") " pod="openstack/placement-746cf47744-w9nm7" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.712352 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ade9101f-5435-455f-807c-d277918cbb46-scripts\") pod \"keystone-6c4445878c-bnpkh\" (UID: \"ade9101f-5435-455f-807c-d277918cbb46\") " pod="openstack/keystone-6c4445878c-bnpkh" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.718658 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ade9101f-5435-455f-807c-d277918cbb46-internal-tls-certs\") pod \"keystone-6c4445878c-bnpkh\" (UID: \"ade9101f-5435-455f-807c-d277918cbb46\") " pod="openstack/keystone-6c4445878c-bnpkh" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.721702 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0320eb6a-f473-4c4c-ae36-f74080dcdaa5-scripts\") pod \"placement-746cf47744-w9nm7\" (UID: \"0320eb6a-f473-4c4c-ae36-f74080dcdaa5\") " pod="openstack/placement-746cf47744-w9nm7" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.725915 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bv7tz\" (UniqueName: \"kubernetes.io/projected/0320eb6a-f473-4c4c-ae36-f74080dcdaa5-kube-api-access-bv7tz\") pod \"placement-746cf47744-w9nm7\" (UID: \"0320eb6a-f473-4c4c-ae36-f74080dcdaa5\") " pod="openstack/placement-746cf47744-w9nm7" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.727582 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ade9101f-5435-455f-807c-d277918cbb46-public-tls-certs\") pod \"keystone-6c4445878c-bnpkh\" (UID: \"ade9101f-5435-455f-807c-d277918cbb46\") " pod="openstack/keystone-6c4445878c-bnpkh" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.727677 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0320eb6a-f473-4c4c-ae36-f74080dcdaa5-combined-ca-bundle\") pod \"placement-746cf47744-w9nm7\" (UID: \"0320eb6a-f473-4c4c-ae36-f74080dcdaa5\") " pod="openstack/placement-746cf47744-w9nm7" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.729776 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fbprg\" (UniqueName: \"kubernetes.io/projected/ade9101f-5435-455f-807c-d277918cbb46-kube-api-access-fbprg\") pod \"keystone-6c4445878c-bnpkh\" (UID: \"ade9101f-5435-455f-807c-d277918cbb46\") " pod="openstack/keystone-6c4445878c-bnpkh" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.816369 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6c4445878c-bnpkh" Dec 07 16:19:42 crc kubenswrapper[4716]: I1207 16:19:42.830415 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-746cf47744-w9nm7" Dec 07 16:19:43 crc kubenswrapper[4716]: I1207 16:19:43.515566 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-6c4445878c-bnpkh"] Dec 07 16:19:43 crc kubenswrapper[4716]: I1207 16:19:43.771802 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-746cf47744-w9nm7"] Dec 07 16:19:44 crc kubenswrapper[4716]: I1207 16:19:44.036422 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 07 16:19:44 crc kubenswrapper[4716]: I1207 16:19:44.336958 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6c4445878c-bnpkh" event={"ID":"ade9101f-5435-455f-807c-d277918cbb46","Type":"ContainerStarted","Data":"cf6ca1a33f2a781b003ae282666e506e36c80c19953dc60d92eddf1bbdb821ef"} Dec 07 16:19:44 crc kubenswrapper[4716]: I1207 16:19:44.337289 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6c4445878c-bnpkh" event={"ID":"ade9101f-5435-455f-807c-d277918cbb46","Type":"ContainerStarted","Data":"20edc9461225d8da3b710b6bcb1a432886f92a7d23f720a4def422c36015978a"} Dec 07 16:19:44 crc kubenswrapper[4716]: I1207 16:19:44.338569 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-6c4445878c-bnpkh" Dec 07 16:19:44 crc kubenswrapper[4716]: I1207 16:19:44.340062 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-746cf47744-w9nm7" event={"ID":"0320eb6a-f473-4c4c-ae36-f74080dcdaa5","Type":"ContainerStarted","Data":"d93ec5cbda35f844be16972b754a5a5d06a171972915b18801319bd628a1153a"} Dec 07 16:19:44 crc kubenswrapper[4716]: I1207 16:19:44.379369 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-6c4445878c-bnpkh" podStartSLOduration=2.379351068 podStartE2EDuration="2.379351068s" podCreationTimestamp="2025-12-07 16:19:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:19:44.379001848 +0000 UTC m=+1047.069286760" watchObservedRunningTime="2025-12-07 16:19:44.379351068 +0000 UTC m=+1047.069635980" Dec 07 16:19:47 crc kubenswrapper[4716]: I1207 16:19:47.332374 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-55f844cf75-rtqpb" Dec 07 16:19:47 crc kubenswrapper[4716]: I1207 16:19:47.383049 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-746cf47744-w9nm7" event={"ID":"0320eb6a-f473-4c4c-ae36-f74080dcdaa5","Type":"ContainerStarted","Data":"a40a3325c31b3e0d3b3cc9f7813f0949e6871f936a32857dde962e4995689c4d"} Dec 07 16:19:47 crc kubenswrapper[4716]: I1207 16:19:47.402177 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-cq5v8"] Dec 07 16:19:47 crc kubenswrapper[4716]: I1207 16:19:47.402483 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-785d8bcb8c-cq5v8" podUID="5ced3e75-e271-433b-ad77-fe9101c18be5" containerName="dnsmasq-dns" containerID="cri-o://8aaa9cd86647e78761528be99baca4b3a306d7e641067604de34f07fb03d3f7e" gracePeriod=10 Dec 07 16:19:47 crc kubenswrapper[4716]: I1207 16:19:47.657714 4716 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-785d8bcb8c-cq5v8" podUID="5ced3e75-e271-433b-ad77-fe9101c18be5" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.142:5353: connect: connection refused" Dec 07 16:19:48 crc kubenswrapper[4716]: I1207 16:19:48.397430 4716 generic.go:334] "Generic (PLEG): container finished" podID="5ced3e75-e271-433b-ad77-fe9101c18be5" containerID="8aaa9cd86647e78761528be99baca4b3a306d7e641067604de34f07fb03d3f7e" exitCode=0 Dec 07 16:19:48 crc kubenswrapper[4716]: I1207 16:19:48.397508 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-cq5v8" event={"ID":"5ced3e75-e271-433b-ad77-fe9101c18be5","Type":"ContainerDied","Data":"8aaa9cd86647e78761528be99baca4b3a306d7e641067604de34f07fb03d3f7e"} Dec 07 16:19:49 crc kubenswrapper[4716]: I1207 16:19:49.234605 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-cq5v8" Dec 07 16:19:49 crc kubenswrapper[4716]: I1207 16:19:49.287675 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5ced3e75-e271-433b-ad77-fe9101c18be5-dns-svc\") pod \"5ced3e75-e271-433b-ad77-fe9101c18be5\" (UID: \"5ced3e75-e271-433b-ad77-fe9101c18be5\") " Dec 07 16:19:49 crc kubenswrapper[4716]: I1207 16:19:49.290040 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzlqz\" (UniqueName: \"kubernetes.io/projected/5ced3e75-e271-433b-ad77-fe9101c18be5-kube-api-access-nzlqz\") pod \"5ced3e75-e271-433b-ad77-fe9101c18be5\" (UID: \"5ced3e75-e271-433b-ad77-fe9101c18be5\") " Dec 07 16:19:49 crc kubenswrapper[4716]: I1207 16:19:49.290184 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5ced3e75-e271-433b-ad77-fe9101c18be5-ovsdbserver-nb\") pod \"5ced3e75-e271-433b-ad77-fe9101c18be5\" (UID: \"5ced3e75-e271-433b-ad77-fe9101c18be5\") " Dec 07 16:19:49 crc kubenswrapper[4716]: I1207 16:19:49.290213 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5ced3e75-e271-433b-ad77-fe9101c18be5-ovsdbserver-sb\") pod \"5ced3e75-e271-433b-ad77-fe9101c18be5\" (UID: \"5ced3e75-e271-433b-ad77-fe9101c18be5\") " Dec 07 16:19:49 crc kubenswrapper[4716]: I1207 16:19:49.290275 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ced3e75-e271-433b-ad77-fe9101c18be5-config\") pod \"5ced3e75-e271-433b-ad77-fe9101c18be5\" (UID: \"5ced3e75-e271-433b-ad77-fe9101c18be5\") " Dec 07 16:19:49 crc kubenswrapper[4716]: I1207 16:19:49.290379 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5ced3e75-e271-433b-ad77-fe9101c18be5-dns-swift-storage-0\") pod \"5ced3e75-e271-433b-ad77-fe9101c18be5\" (UID: \"5ced3e75-e271-433b-ad77-fe9101c18be5\") " Dec 07 16:19:49 crc kubenswrapper[4716]: I1207 16:19:49.312393 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ced3e75-e271-433b-ad77-fe9101c18be5-kube-api-access-nzlqz" (OuterVolumeSpecName: "kube-api-access-nzlqz") pod "5ced3e75-e271-433b-ad77-fe9101c18be5" (UID: "5ced3e75-e271-433b-ad77-fe9101c18be5"). InnerVolumeSpecName "kube-api-access-nzlqz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:19:49 crc kubenswrapper[4716]: I1207 16:19:49.375014 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ced3e75-e271-433b-ad77-fe9101c18be5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5ced3e75-e271-433b-ad77-fe9101c18be5" (UID: "5ced3e75-e271-433b-ad77-fe9101c18be5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:19:49 crc kubenswrapper[4716]: I1207 16:19:49.393180 4716 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5ced3e75-e271-433b-ad77-fe9101c18be5-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 07 16:19:49 crc kubenswrapper[4716]: I1207 16:19:49.393209 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzlqz\" (UniqueName: \"kubernetes.io/projected/5ced3e75-e271-433b-ad77-fe9101c18be5-kube-api-access-nzlqz\") on node \"crc\" DevicePath \"\"" Dec 07 16:19:49 crc kubenswrapper[4716]: I1207 16:19:49.394464 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ced3e75-e271-433b-ad77-fe9101c18be5-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5ced3e75-e271-433b-ad77-fe9101c18be5" (UID: "5ced3e75-e271-433b-ad77-fe9101c18be5"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:19:49 crc kubenswrapper[4716]: I1207 16:19:49.411639 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ced3e75-e271-433b-ad77-fe9101c18be5-config" (OuterVolumeSpecName: "config") pod "5ced3e75-e271-433b-ad77-fe9101c18be5" (UID: "5ced3e75-e271-433b-ad77-fe9101c18be5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:19:49 crc kubenswrapper[4716]: I1207 16:19:49.419684 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ced3e75-e271-433b-ad77-fe9101c18be5-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "5ced3e75-e271-433b-ad77-fe9101c18be5" (UID: "5ced3e75-e271-433b-ad77-fe9101c18be5"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:19:49 crc kubenswrapper[4716]: I1207 16:19:49.430523 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-cq5v8" Dec 07 16:19:49 crc kubenswrapper[4716]: I1207 16:19:49.430626 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-cq5v8" event={"ID":"5ced3e75-e271-433b-ad77-fe9101c18be5","Type":"ContainerDied","Data":"1306ff3757f816247c4c49b1313e0ecbee33cc633d87c01a1da6f9bff3f01822"} Dec 07 16:19:49 crc kubenswrapper[4716]: I1207 16:19:49.430697 4716 scope.go:117] "RemoveContainer" containerID="8aaa9cd86647e78761528be99baca4b3a306d7e641067604de34f07fb03d3f7e" Dec 07 16:19:49 crc kubenswrapper[4716]: I1207 16:19:49.438477 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2ee00bf0-1412-4fca-a570-6600db549923","Type":"ContainerStarted","Data":"52d4c8b65fd2dddf05a3fa2b1d30919639544b392333691964350460d53b6375"} Dec 07 16:19:49 crc kubenswrapper[4716]: I1207 16:19:49.441277 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-746cf47744-w9nm7" event={"ID":"0320eb6a-f473-4c4c-ae36-f74080dcdaa5","Type":"ContainerStarted","Data":"2d6b03559b48c01ab3997f50d4c219409c20bf10a97221c70e93a11f43778c34"} Dec 07 16:19:49 crc kubenswrapper[4716]: I1207 16:19:49.442194 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-746cf47744-w9nm7" Dec 07 16:19:49 crc kubenswrapper[4716]: I1207 16:19:49.442276 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-746cf47744-w9nm7" Dec 07 16:19:49 crc kubenswrapper[4716]: I1207 16:19:49.446009 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-tk47q" event={"ID":"717a326f-b1ac-4fbe-b1c8-b669081dab80","Type":"ContainerStarted","Data":"a4bb3f5fae952b8c1002cb654179ae8ea80bb4dbe8649a05733636adb9e748f2"} Dec 07 16:19:49 crc kubenswrapper[4716]: I1207 16:19:49.452757 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ced3e75-e271-433b-ad77-fe9101c18be5-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5ced3e75-e271-433b-ad77-fe9101c18be5" (UID: "5ced3e75-e271-433b-ad77-fe9101c18be5"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:19:49 crc kubenswrapper[4716]: I1207 16:19:49.459566 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-746cf47744-w9nm7" podStartSLOduration=7.459532703 podStartE2EDuration="7.459532703s" podCreationTimestamp="2025-12-07 16:19:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:19:49.457568829 +0000 UTC m=+1052.147853761" watchObservedRunningTime="2025-12-07 16:19:49.459532703 +0000 UTC m=+1052.149817615" Dec 07 16:19:49 crc kubenswrapper[4716]: I1207 16:19:49.473369 4716 scope.go:117] "RemoveContainer" containerID="04594e235359bf223f892f9d9da526ca2df5e5e5ccdd1a9c7517fa66eb8ca47c" Dec 07 16:19:49 crc kubenswrapper[4716]: I1207 16:19:49.485947 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-tk47q" podStartSLOduration=4.311987918 podStartE2EDuration="48.485928047s" podCreationTimestamp="2025-12-07 16:19:01 +0000 UTC" firstStartedPulling="2025-12-07 16:19:04.824239174 +0000 UTC m=+1007.514524086" lastFinishedPulling="2025-12-07 16:19:48.998179303 +0000 UTC m=+1051.688464215" observedRunningTime="2025-12-07 16:19:49.479615606 +0000 UTC m=+1052.169900508" watchObservedRunningTime="2025-12-07 16:19:49.485928047 +0000 UTC m=+1052.176212949" Dec 07 16:19:49 crc kubenswrapper[4716]: I1207 16:19:49.494416 4716 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5ced3e75-e271-433b-ad77-fe9101c18be5-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 07 16:19:49 crc kubenswrapper[4716]: I1207 16:19:49.494439 4716 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5ced3e75-e271-433b-ad77-fe9101c18be5-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 07 16:19:49 crc kubenswrapper[4716]: I1207 16:19:49.494450 4716 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ced3e75-e271-433b-ad77-fe9101c18be5-config\") on node \"crc\" DevicePath \"\"" Dec 07 16:19:49 crc kubenswrapper[4716]: I1207 16:19:49.494460 4716 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5ced3e75-e271-433b-ad77-fe9101c18be5-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 07 16:19:50 crc kubenswrapper[4716]: I1207 16:19:50.467150 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-cq5v8"] Dec 07 16:19:50 crc kubenswrapper[4716]: I1207 16:19:50.474090 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-cq5v8"] Dec 07 16:19:50 crc kubenswrapper[4716]: I1207 16:19:50.720006 4716 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-86c7567d4-99rx9" podUID="c291e114-7940-46bc-91d3-a8fa256549ff" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Dec 07 16:19:50 crc kubenswrapper[4716]: I1207 16:19:50.790450 4716 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7f9b78866d-9g9kv" podUID="d845ae0b-5d45-4021-a1e9-4b124298b65b" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.148:8443: connect: connection refused" Dec 07 16:19:51 crc kubenswrapper[4716]: I1207 16:19:51.533348 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-gqg6b" event={"ID":"f9413953-056c-4d73-b534-12e6816e6fb4","Type":"ContainerStarted","Data":"ad8e4e125c3aa6c78a78cf693c30cd14ae8bb48186cfb3c839fde875ef8d4b8f"} Dec 07 16:19:51 crc kubenswrapper[4716]: I1207 16:19:51.559775 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-gqg6b" podStartSLOduration=6.425835329 podStartE2EDuration="50.559723577s" podCreationTimestamp="2025-12-07 16:19:01 +0000 UTC" firstStartedPulling="2025-12-07 16:19:04.892806067 +0000 UTC m=+1007.583090979" lastFinishedPulling="2025-12-07 16:19:49.026694315 +0000 UTC m=+1051.716979227" observedRunningTime="2025-12-07 16:19:51.557437584 +0000 UTC m=+1054.247722496" watchObservedRunningTime="2025-12-07 16:19:51.559723577 +0000 UTC m=+1054.250008489" Dec 07 16:19:51 crc kubenswrapper[4716]: I1207 16:19:51.677342 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ced3e75-e271-433b-ad77-fe9101c18be5" path="/var/lib/kubelet/pods/5ced3e75-e271-433b-ad77-fe9101c18be5/volumes" Dec 07 16:19:52 crc kubenswrapper[4716]: I1207 16:19:52.544614 4716 generic.go:334] "Generic (PLEG): container finished" podID="717a326f-b1ac-4fbe-b1c8-b669081dab80" containerID="a4bb3f5fae952b8c1002cb654179ae8ea80bb4dbe8649a05733636adb9e748f2" exitCode=0 Dec 07 16:19:52 crc kubenswrapper[4716]: I1207 16:19:52.544664 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-tk47q" event={"ID":"717a326f-b1ac-4fbe-b1c8-b669081dab80","Type":"ContainerDied","Data":"a4bb3f5fae952b8c1002cb654179ae8ea80bb4dbe8649a05733636adb9e748f2"} Dec 07 16:19:52 crc kubenswrapper[4716]: I1207 16:19:52.760984 4716 patch_prober.go:28] interesting pod/machine-config-daemon-zcxxp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 16:19:52 crc kubenswrapper[4716]: I1207 16:19:52.761036 4716 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 16:19:52 crc kubenswrapper[4716]: I1207 16:19:52.761114 4716 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" Dec 07 16:19:52 crc kubenswrapper[4716]: I1207 16:19:52.761688 4716 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"73e652154907931afdb244fb07e6dd85483e6b06afe1a76ebcc2b8bf8b2c7310"} pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 07 16:19:52 crc kubenswrapper[4716]: I1207 16:19:52.761739 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" containerName="machine-config-daemon" containerID="cri-o://73e652154907931afdb244fb07e6dd85483e6b06afe1a76ebcc2b8bf8b2c7310" gracePeriod=600 Dec 07 16:19:53 crc kubenswrapper[4716]: I1207 16:19:53.557120 4716 generic.go:334] "Generic (PLEG): container finished" podID="c15b59eb-565d-4556-a4ce-75afdf159dc8" containerID="73e652154907931afdb244fb07e6dd85483e6b06afe1a76ebcc2b8bf8b2c7310" exitCode=0 Dec 07 16:19:53 crc kubenswrapper[4716]: I1207 16:19:53.557239 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" event={"ID":"c15b59eb-565d-4556-a4ce-75afdf159dc8","Type":"ContainerDied","Data":"73e652154907931afdb244fb07e6dd85483e6b06afe1a76ebcc2b8bf8b2c7310"} Dec 07 16:19:53 crc kubenswrapper[4716]: I1207 16:19:53.557755 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" event={"ID":"c15b59eb-565d-4556-a4ce-75afdf159dc8","Type":"ContainerStarted","Data":"43e1cbd05079c8cdd0bd98789bfa5b92ef2c8c3e87845aeb18f274ab5529de34"} Dec 07 16:19:53 crc kubenswrapper[4716]: I1207 16:19:53.557784 4716 scope.go:117] "RemoveContainer" containerID="8ec0c9fb49e1eaea181d9e297f07686725c05f58dac3a79c7a2ee3f00f979908" Dec 07 16:19:54 crc kubenswrapper[4716]: I1207 16:19:54.569612 4716 generic.go:334] "Generic (PLEG): container finished" podID="f9413953-056c-4d73-b534-12e6816e6fb4" containerID="ad8e4e125c3aa6c78a78cf693c30cd14ae8bb48186cfb3c839fde875ef8d4b8f" exitCode=0 Dec 07 16:19:54 crc kubenswrapper[4716]: I1207 16:19:54.570648 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-gqg6b" event={"ID":"f9413953-056c-4d73-b534-12e6816e6fb4","Type":"ContainerDied","Data":"ad8e4e125c3aa6c78a78cf693c30cd14ae8bb48186cfb3c839fde875ef8d4b8f"} Dec 07 16:19:57 crc kubenswrapper[4716]: I1207 16:19:57.361237 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-tk47q" Dec 07 16:19:57 crc kubenswrapper[4716]: I1207 16:19:57.368207 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-gqg6b" Dec 07 16:19:57 crc kubenswrapper[4716]: I1207 16:19:57.517808 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nrnjl\" (UniqueName: \"kubernetes.io/projected/f9413953-056c-4d73-b534-12e6816e6fb4-kube-api-access-nrnjl\") pod \"f9413953-056c-4d73-b534-12e6816e6fb4\" (UID: \"f9413953-056c-4d73-b534-12e6816e6fb4\") " Dec 07 16:19:57 crc kubenswrapper[4716]: I1207 16:19:57.517924 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/717a326f-b1ac-4fbe-b1c8-b669081dab80-db-sync-config-data\") pod \"717a326f-b1ac-4fbe-b1c8-b669081dab80\" (UID: \"717a326f-b1ac-4fbe-b1c8-b669081dab80\") " Dec 07 16:19:57 crc kubenswrapper[4716]: I1207 16:19:57.518036 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f9413953-056c-4d73-b534-12e6816e6fb4-db-sync-config-data\") pod \"f9413953-056c-4d73-b534-12e6816e6fb4\" (UID: \"f9413953-056c-4d73-b534-12e6816e6fb4\") " Dec 07 16:19:57 crc kubenswrapper[4716]: I1207 16:19:57.518121 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9413953-056c-4d73-b534-12e6816e6fb4-config-data\") pod \"f9413953-056c-4d73-b534-12e6816e6fb4\" (UID: \"f9413953-056c-4d73-b534-12e6816e6fb4\") " Dec 07 16:19:57 crc kubenswrapper[4716]: I1207 16:19:57.522223 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9413953-056c-4d73-b534-12e6816e6fb4-scripts\") pod \"f9413953-056c-4d73-b534-12e6816e6fb4\" (UID: \"f9413953-056c-4d73-b534-12e6816e6fb4\") " Dec 07 16:19:57 crc kubenswrapper[4716]: I1207 16:19:57.522297 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9413953-056c-4d73-b534-12e6816e6fb4-combined-ca-bundle\") pod \"f9413953-056c-4d73-b534-12e6816e6fb4\" (UID: \"f9413953-056c-4d73-b534-12e6816e6fb4\") " Dec 07 16:19:57 crc kubenswrapper[4716]: I1207 16:19:57.522314 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qgdkv\" (UniqueName: \"kubernetes.io/projected/717a326f-b1ac-4fbe-b1c8-b669081dab80-kube-api-access-qgdkv\") pod \"717a326f-b1ac-4fbe-b1c8-b669081dab80\" (UID: \"717a326f-b1ac-4fbe-b1c8-b669081dab80\") " Dec 07 16:19:57 crc kubenswrapper[4716]: I1207 16:19:57.522335 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/717a326f-b1ac-4fbe-b1c8-b669081dab80-combined-ca-bundle\") pod \"717a326f-b1ac-4fbe-b1c8-b669081dab80\" (UID: \"717a326f-b1ac-4fbe-b1c8-b669081dab80\") " Dec 07 16:19:57 crc kubenswrapper[4716]: I1207 16:19:57.522362 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f9413953-056c-4d73-b534-12e6816e6fb4-etc-machine-id\") pod \"f9413953-056c-4d73-b534-12e6816e6fb4\" (UID: \"f9413953-056c-4d73-b534-12e6816e6fb4\") " Dec 07 16:19:57 crc kubenswrapper[4716]: I1207 16:19:57.524088 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/717a326f-b1ac-4fbe-b1c8-b669081dab80-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "717a326f-b1ac-4fbe-b1c8-b669081dab80" (UID: "717a326f-b1ac-4fbe-b1c8-b669081dab80"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:19:57 crc kubenswrapper[4716]: I1207 16:19:57.524666 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9413953-056c-4d73-b534-12e6816e6fb4-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "f9413953-056c-4d73-b534-12e6816e6fb4" (UID: "f9413953-056c-4d73-b534-12e6816e6fb4"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:19:57 crc kubenswrapper[4716]: I1207 16:19:57.524712 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f9413953-056c-4d73-b534-12e6816e6fb4-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "f9413953-056c-4d73-b534-12e6816e6fb4" (UID: "f9413953-056c-4d73-b534-12e6816e6fb4"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 16:19:57 crc kubenswrapper[4716]: I1207 16:19:57.525146 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9413953-056c-4d73-b534-12e6816e6fb4-kube-api-access-nrnjl" (OuterVolumeSpecName: "kube-api-access-nrnjl") pod "f9413953-056c-4d73-b534-12e6816e6fb4" (UID: "f9413953-056c-4d73-b534-12e6816e6fb4"). InnerVolumeSpecName "kube-api-access-nrnjl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:19:57 crc kubenswrapper[4716]: I1207 16:19:57.526320 4716 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f9413953-056c-4d73-b534-12e6816e6fb4-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 07 16:19:57 crc kubenswrapper[4716]: I1207 16:19:57.526349 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nrnjl\" (UniqueName: \"kubernetes.io/projected/f9413953-056c-4d73-b534-12e6816e6fb4-kube-api-access-nrnjl\") on node \"crc\" DevicePath \"\"" Dec 07 16:19:57 crc kubenswrapper[4716]: I1207 16:19:57.526362 4716 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/717a326f-b1ac-4fbe-b1c8-b669081dab80-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 16:19:57 crc kubenswrapper[4716]: I1207 16:19:57.526372 4716 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f9413953-056c-4d73-b534-12e6816e6fb4-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 16:19:57 crc kubenswrapper[4716]: I1207 16:19:57.527729 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9413953-056c-4d73-b534-12e6816e6fb4-scripts" (OuterVolumeSpecName: "scripts") pod "f9413953-056c-4d73-b534-12e6816e6fb4" (UID: "f9413953-056c-4d73-b534-12e6816e6fb4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:19:57 crc kubenswrapper[4716]: I1207 16:19:57.528215 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/717a326f-b1ac-4fbe-b1c8-b669081dab80-kube-api-access-qgdkv" (OuterVolumeSpecName: "kube-api-access-qgdkv") pod "717a326f-b1ac-4fbe-b1c8-b669081dab80" (UID: "717a326f-b1ac-4fbe-b1c8-b669081dab80"). InnerVolumeSpecName "kube-api-access-qgdkv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:19:57 crc kubenswrapper[4716]: I1207 16:19:57.552351 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9413953-056c-4d73-b534-12e6816e6fb4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f9413953-056c-4d73-b534-12e6816e6fb4" (UID: "f9413953-056c-4d73-b534-12e6816e6fb4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:19:57 crc kubenswrapper[4716]: I1207 16:19:57.561843 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/717a326f-b1ac-4fbe-b1c8-b669081dab80-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "717a326f-b1ac-4fbe-b1c8-b669081dab80" (UID: "717a326f-b1ac-4fbe-b1c8-b669081dab80"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:19:57 crc kubenswrapper[4716]: I1207 16:19:57.562782 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9413953-056c-4d73-b534-12e6816e6fb4-config-data" (OuterVolumeSpecName: "config-data") pod "f9413953-056c-4d73-b534-12e6816e6fb4" (UID: "f9413953-056c-4d73-b534-12e6816e6fb4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:19:57 crc kubenswrapper[4716]: I1207 16:19:57.607894 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-tk47q" event={"ID":"717a326f-b1ac-4fbe-b1c8-b669081dab80","Type":"ContainerDied","Data":"15f2b878a373b032d52f19fadc7e27b03a3417197ec85e6a2ffcfeb5203f21c8"} Dec 07 16:19:57 crc kubenswrapper[4716]: I1207 16:19:57.607938 4716 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="15f2b878a373b032d52f19fadc7e27b03a3417197ec85e6a2ffcfeb5203f21c8" Dec 07 16:19:57 crc kubenswrapper[4716]: I1207 16:19:57.607997 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-tk47q" Dec 07 16:19:57 crc kubenswrapper[4716]: I1207 16:19:57.610449 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-gqg6b" event={"ID":"f9413953-056c-4d73-b534-12e6816e6fb4","Type":"ContainerDied","Data":"73ef3cdbf94738d92af0605456c808535aff71aa960d87d8943550a902ca6331"} Dec 07 16:19:57 crc kubenswrapper[4716]: I1207 16:19:57.610488 4716 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="73ef3cdbf94738d92af0605456c808535aff71aa960d87d8943550a902ca6331" Dec 07 16:19:57 crc kubenswrapper[4716]: I1207 16:19:57.610547 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-gqg6b" Dec 07 16:19:57 crc kubenswrapper[4716]: I1207 16:19:57.631525 4716 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9413953-056c-4d73-b534-12e6816e6fb4-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 16:19:57 crc kubenswrapper[4716]: I1207 16:19:57.631620 4716 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9413953-056c-4d73-b534-12e6816e6fb4-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 16:19:57 crc kubenswrapper[4716]: I1207 16:19:57.631661 4716 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9413953-056c-4d73-b534-12e6816e6fb4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 16:19:57 crc kubenswrapper[4716]: I1207 16:19:57.631756 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qgdkv\" (UniqueName: \"kubernetes.io/projected/717a326f-b1ac-4fbe-b1c8-b669081dab80-kube-api-access-qgdkv\") on node \"crc\" DevicePath \"\"" Dec 07 16:19:57 crc kubenswrapper[4716]: I1207 16:19:57.631767 4716 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/717a326f-b1ac-4fbe-b1c8-b669081dab80-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.653882 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-587c8bb479-4d8m7"] Dec 07 16:19:58 crc kubenswrapper[4716]: E1207 16:19:58.656125 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9413953-056c-4d73-b534-12e6816e6fb4" containerName="cinder-db-sync" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.656144 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9413953-056c-4d73-b534-12e6816e6fb4" containerName="cinder-db-sync" Dec 07 16:19:58 crc kubenswrapper[4716]: E1207 16:19:58.656156 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ced3e75-e271-433b-ad77-fe9101c18be5" containerName="dnsmasq-dns" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.656162 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ced3e75-e271-433b-ad77-fe9101c18be5" containerName="dnsmasq-dns" Dec 07 16:19:58 crc kubenswrapper[4716]: E1207 16:19:58.656171 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ced3e75-e271-433b-ad77-fe9101c18be5" containerName="init" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.656179 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ced3e75-e271-433b-ad77-fe9101c18be5" containerName="init" Dec 07 16:19:58 crc kubenswrapper[4716]: E1207 16:19:58.656208 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="717a326f-b1ac-4fbe-b1c8-b669081dab80" containerName="barbican-db-sync" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.656214 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="717a326f-b1ac-4fbe-b1c8-b669081dab80" containerName="barbican-db-sync" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.656375 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9413953-056c-4d73-b534-12e6816e6fb4" containerName="cinder-db-sync" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.656388 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ced3e75-e271-433b-ad77-fe9101c18be5" containerName="dnsmasq-dns" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.656401 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="717a326f-b1ac-4fbe-b1c8-b669081dab80" containerName="barbican-db-sync" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.654737 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2ee00bf0-1412-4fca-a570-6600db549923" containerName="ceilometer-notification-agent" containerID="cri-o://411ac652ee7365d2f2c78ae6705384e1dadf6242d65ea1cab9c0a5600032ff1d" gracePeriod=30 Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.654174 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2ee00bf0-1412-4fca-a570-6600db549923" containerName="ceilometer-central-agent" containerID="cri-o://65d1e11cc55500273d099448bfdb88d42d28a5fe058e6da987f62dc02869707b" gracePeriod=30 Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.654714 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2ee00bf0-1412-4fca-a570-6600db549923" containerName="proxy-httpd" containerID="cri-o://bfb1a76412f365d957dd1c3a79ba7fd2b942a53d42354cb44f0eb4e1d899bebd" gracePeriod=30 Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.654727 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2ee00bf0-1412-4fca-a570-6600db549923" containerName="sg-core" containerID="cri-o://52d4c8b65fd2dddf05a3fa2b1d30919639544b392333691964350460d53b6375" gracePeriod=30 Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.657547 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2ee00bf0-1412-4fca-a570-6600db549923","Type":"ContainerStarted","Data":"bfb1a76412f365d957dd1c3a79ba7fd2b942a53d42354cb44f0eb4e1d899bebd"} Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.657582 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.657659 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-587c8bb479-4d8m7" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.663123 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.663241 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.663351 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-kxsqg" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.700346 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-75c8ccd7fb-bhzds"] Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.701985 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-75c8ccd7fb-bhzds" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.703985 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.716134 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-587c8bb479-4d8m7"] Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.728910 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.731516 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.733120 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.702695339 podStartE2EDuration="56.733108045s" podCreationTimestamp="2025-12-07 16:19:02 +0000 UTC" firstStartedPulling="2025-12-07 16:19:04.868612489 +0000 UTC m=+1007.558897421" lastFinishedPulling="2025-12-07 16:19:57.899025215 +0000 UTC m=+1060.589310127" observedRunningTime="2025-12-07 16:19:58.69559876 +0000 UTC m=+1061.385883672" watchObservedRunningTime="2025-12-07 16:19:58.733108045 +0000 UTC m=+1061.423392957" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.735518 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.735540 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-cmrq8" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.735670 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.735771 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.772464 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-75c8ccd7fb-bhzds"] Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.797225 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.838796 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-qsdmq"] Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.840397 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-qsdmq" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.859206 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/214231d4-11de-4827-afd3-e5169b138b7b-config-data-custom\") pod \"barbican-keystone-listener-75c8ccd7fb-bhzds\" (UID: \"214231d4-11de-4827-afd3-e5169b138b7b\") " pod="openstack/barbican-keystone-listener-75c8ccd7fb-bhzds" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.859273 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/214231d4-11de-4827-afd3-e5169b138b7b-combined-ca-bundle\") pod \"barbican-keystone-listener-75c8ccd7fb-bhzds\" (UID: \"214231d4-11de-4827-afd3-e5169b138b7b\") " pod="openstack/barbican-keystone-listener-75c8ccd7fb-bhzds" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.859308 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc929236-8385-4f05-8ca5-37315e852be6-config-data\") pod \"barbican-worker-587c8bb479-4d8m7\" (UID: \"cc929236-8385-4f05-8ca5-37315e852be6\") " pod="openstack/barbican-worker-587c8bb479-4d8m7" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.859345 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cc929236-8385-4f05-8ca5-37315e852be6-config-data-custom\") pod \"barbican-worker-587c8bb479-4d8m7\" (UID: \"cc929236-8385-4f05-8ca5-37315e852be6\") " pod="openstack/barbican-worker-587c8bb479-4d8m7" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.859363 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/03ba9a5b-efe2-4180-8110-cd8db3e65278-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"03ba9a5b-efe2-4180-8110-cd8db3e65278\") " pod="openstack/cinder-scheduler-0" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.859388 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03ba9a5b-efe2-4180-8110-cd8db3e65278-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"03ba9a5b-efe2-4180-8110-cd8db3e65278\") " pod="openstack/cinder-scheduler-0" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.859436 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/214231d4-11de-4827-afd3-e5169b138b7b-config-data\") pod \"barbican-keystone-listener-75c8ccd7fb-bhzds\" (UID: \"214231d4-11de-4827-afd3-e5169b138b7b\") " pod="openstack/barbican-keystone-listener-75c8ccd7fb-bhzds" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.859468 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc929236-8385-4f05-8ca5-37315e852be6-logs\") pod \"barbican-worker-587c8bb479-4d8m7\" (UID: \"cc929236-8385-4f05-8ca5-37315e852be6\") " pod="openstack/barbican-worker-587c8bb479-4d8m7" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.859489 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/03ba9a5b-efe2-4180-8110-cd8db3e65278-scripts\") pod \"cinder-scheduler-0\" (UID: \"03ba9a5b-efe2-4180-8110-cd8db3e65278\") " pod="openstack/cinder-scheduler-0" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.859505 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/03ba9a5b-efe2-4180-8110-cd8db3e65278-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"03ba9a5b-efe2-4180-8110-cd8db3e65278\") " pod="openstack/cinder-scheduler-0" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.859520 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03ba9a5b-efe2-4180-8110-cd8db3e65278-config-data\") pod \"cinder-scheduler-0\" (UID: \"03ba9a5b-efe2-4180-8110-cd8db3e65278\") " pod="openstack/cinder-scheduler-0" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.859584 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/214231d4-11de-4827-afd3-e5169b138b7b-logs\") pod \"barbican-keystone-listener-75c8ccd7fb-bhzds\" (UID: \"214231d4-11de-4827-afd3-e5169b138b7b\") " pod="openstack/barbican-keystone-listener-75c8ccd7fb-bhzds" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.859695 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5t2cp\" (UniqueName: \"kubernetes.io/projected/03ba9a5b-efe2-4180-8110-cd8db3e65278-kube-api-access-5t2cp\") pod \"cinder-scheduler-0\" (UID: \"03ba9a5b-efe2-4180-8110-cd8db3e65278\") " pod="openstack/cinder-scheduler-0" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.859773 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc929236-8385-4f05-8ca5-37315e852be6-combined-ca-bundle\") pod \"barbican-worker-587c8bb479-4d8m7\" (UID: \"cc929236-8385-4f05-8ca5-37315e852be6\") " pod="openstack/barbican-worker-587c8bb479-4d8m7" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.859789 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n9rkp\" (UniqueName: \"kubernetes.io/projected/cc929236-8385-4f05-8ca5-37315e852be6-kube-api-access-n9rkp\") pod \"barbican-worker-587c8bb479-4d8m7\" (UID: \"cc929236-8385-4f05-8ca5-37315e852be6\") " pod="openstack/barbican-worker-587c8bb479-4d8m7" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.859810 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s84v4\" (UniqueName: \"kubernetes.io/projected/214231d4-11de-4827-afd3-e5169b138b7b-kube-api-access-s84v4\") pod \"barbican-keystone-listener-75c8ccd7fb-bhzds\" (UID: \"214231d4-11de-4827-afd3-e5169b138b7b\") " pod="openstack/barbican-keystone-listener-75c8ccd7fb-bhzds" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.880151 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-qsdmq"] Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.898681 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-qsdmq"] Dec 07 16:19:58 crc kubenswrapper[4716]: E1207 16:19:58.902514 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc dns-swift-storage-0 kube-api-access-gw8vm ovsdbserver-nb ovsdbserver-sb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-85ff748b95-qsdmq" podUID="add8b3eb-3a12-4834-95cd-7839dc327eb7" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.912593 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-vb6n9"] Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.914645 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-vb6n9" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.926628 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-vb6n9"] Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.961141 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/add8b3eb-3a12-4834-95cd-7839dc327eb7-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-qsdmq\" (UID: \"add8b3eb-3a12-4834-95cd-7839dc327eb7\") " pod="openstack/dnsmasq-dns-85ff748b95-qsdmq" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.961181 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/add8b3eb-3a12-4834-95cd-7839dc327eb7-config\") pod \"dnsmasq-dns-85ff748b95-qsdmq\" (UID: \"add8b3eb-3a12-4834-95cd-7839dc327eb7\") " pod="openstack/dnsmasq-dns-85ff748b95-qsdmq" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.961219 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5t2cp\" (UniqueName: \"kubernetes.io/projected/03ba9a5b-efe2-4180-8110-cd8db3e65278-kube-api-access-5t2cp\") pod \"cinder-scheduler-0\" (UID: \"03ba9a5b-efe2-4180-8110-cd8db3e65278\") " pod="openstack/cinder-scheduler-0" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.961261 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc929236-8385-4f05-8ca5-37315e852be6-combined-ca-bundle\") pod \"barbican-worker-587c8bb479-4d8m7\" (UID: \"cc929236-8385-4f05-8ca5-37315e852be6\") " pod="openstack/barbican-worker-587c8bb479-4d8m7" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.961278 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n9rkp\" (UniqueName: \"kubernetes.io/projected/cc929236-8385-4f05-8ca5-37315e852be6-kube-api-access-n9rkp\") pod \"barbican-worker-587c8bb479-4d8m7\" (UID: \"cc929236-8385-4f05-8ca5-37315e852be6\") " pod="openstack/barbican-worker-587c8bb479-4d8m7" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.961299 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s84v4\" (UniqueName: \"kubernetes.io/projected/214231d4-11de-4827-afd3-e5169b138b7b-kube-api-access-s84v4\") pod \"barbican-keystone-listener-75c8ccd7fb-bhzds\" (UID: \"214231d4-11de-4827-afd3-e5169b138b7b\") " pod="openstack/barbican-keystone-listener-75c8ccd7fb-bhzds" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.961338 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/add8b3eb-3a12-4834-95cd-7839dc327eb7-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-qsdmq\" (UID: \"add8b3eb-3a12-4834-95cd-7839dc327eb7\") " pod="openstack/dnsmasq-dns-85ff748b95-qsdmq" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.961359 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/214231d4-11de-4827-afd3-e5169b138b7b-config-data-custom\") pod \"barbican-keystone-listener-75c8ccd7fb-bhzds\" (UID: \"214231d4-11de-4827-afd3-e5169b138b7b\") " pod="openstack/barbican-keystone-listener-75c8ccd7fb-bhzds" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.961379 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/214231d4-11de-4827-afd3-e5169b138b7b-combined-ca-bundle\") pod \"barbican-keystone-listener-75c8ccd7fb-bhzds\" (UID: \"214231d4-11de-4827-afd3-e5169b138b7b\") " pod="openstack/barbican-keystone-listener-75c8ccd7fb-bhzds" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.961396 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc929236-8385-4f05-8ca5-37315e852be6-config-data\") pod \"barbican-worker-587c8bb479-4d8m7\" (UID: \"cc929236-8385-4f05-8ca5-37315e852be6\") " pod="openstack/barbican-worker-587c8bb479-4d8m7" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.961415 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cc929236-8385-4f05-8ca5-37315e852be6-config-data-custom\") pod \"barbican-worker-587c8bb479-4d8m7\" (UID: \"cc929236-8385-4f05-8ca5-37315e852be6\") " pod="openstack/barbican-worker-587c8bb479-4d8m7" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.961430 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/03ba9a5b-efe2-4180-8110-cd8db3e65278-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"03ba9a5b-efe2-4180-8110-cd8db3e65278\") " pod="openstack/cinder-scheduler-0" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.961449 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03ba9a5b-efe2-4180-8110-cd8db3e65278-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"03ba9a5b-efe2-4180-8110-cd8db3e65278\") " pod="openstack/cinder-scheduler-0" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.961478 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/214231d4-11de-4827-afd3-e5169b138b7b-config-data\") pod \"barbican-keystone-listener-75c8ccd7fb-bhzds\" (UID: \"214231d4-11de-4827-afd3-e5169b138b7b\") " pod="openstack/barbican-keystone-listener-75c8ccd7fb-bhzds" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.961500 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/add8b3eb-3a12-4834-95cd-7839dc327eb7-dns-svc\") pod \"dnsmasq-dns-85ff748b95-qsdmq\" (UID: \"add8b3eb-3a12-4834-95cd-7839dc327eb7\") " pod="openstack/dnsmasq-dns-85ff748b95-qsdmq" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.961516 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc929236-8385-4f05-8ca5-37315e852be6-logs\") pod \"barbican-worker-587c8bb479-4d8m7\" (UID: \"cc929236-8385-4f05-8ca5-37315e852be6\") " pod="openstack/barbican-worker-587c8bb479-4d8m7" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.961530 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gw8vm\" (UniqueName: \"kubernetes.io/projected/add8b3eb-3a12-4834-95cd-7839dc327eb7-kube-api-access-gw8vm\") pod \"dnsmasq-dns-85ff748b95-qsdmq\" (UID: \"add8b3eb-3a12-4834-95cd-7839dc327eb7\") " pod="openstack/dnsmasq-dns-85ff748b95-qsdmq" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.961550 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/03ba9a5b-efe2-4180-8110-cd8db3e65278-scripts\") pod \"cinder-scheduler-0\" (UID: \"03ba9a5b-efe2-4180-8110-cd8db3e65278\") " pod="openstack/cinder-scheduler-0" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.961569 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/03ba9a5b-efe2-4180-8110-cd8db3e65278-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"03ba9a5b-efe2-4180-8110-cd8db3e65278\") " pod="openstack/cinder-scheduler-0" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.961583 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03ba9a5b-efe2-4180-8110-cd8db3e65278-config-data\") pod \"cinder-scheduler-0\" (UID: \"03ba9a5b-efe2-4180-8110-cd8db3e65278\") " pod="openstack/cinder-scheduler-0" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.961604 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/add8b3eb-3a12-4834-95cd-7839dc327eb7-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-qsdmq\" (UID: \"add8b3eb-3a12-4834-95cd-7839dc327eb7\") " pod="openstack/dnsmasq-dns-85ff748b95-qsdmq" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.961632 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/214231d4-11de-4827-afd3-e5169b138b7b-logs\") pod \"barbican-keystone-listener-75c8ccd7fb-bhzds\" (UID: \"214231d4-11de-4827-afd3-e5169b138b7b\") " pod="openstack/barbican-keystone-listener-75c8ccd7fb-bhzds" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.961999 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/214231d4-11de-4827-afd3-e5169b138b7b-logs\") pod \"barbican-keystone-listener-75c8ccd7fb-bhzds\" (UID: \"214231d4-11de-4827-afd3-e5169b138b7b\") " pod="openstack/barbican-keystone-listener-75c8ccd7fb-bhzds" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.962145 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/03ba9a5b-efe2-4180-8110-cd8db3e65278-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"03ba9a5b-efe2-4180-8110-cd8db3e65278\") " pod="openstack/cinder-scheduler-0" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.968025 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03ba9a5b-efe2-4180-8110-cd8db3e65278-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"03ba9a5b-efe2-4180-8110-cd8db3e65278\") " pod="openstack/cinder-scheduler-0" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.968298 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc929236-8385-4f05-8ca5-37315e852be6-combined-ca-bundle\") pod \"barbican-worker-587c8bb479-4d8m7\" (UID: \"cc929236-8385-4f05-8ca5-37315e852be6\") " pod="openstack/barbican-worker-587c8bb479-4d8m7" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.970394 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-65b7f6996b-q6xwd"] Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.971019 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc929236-8385-4f05-8ca5-37315e852be6-logs\") pod \"barbican-worker-587c8bb479-4d8m7\" (UID: \"cc929236-8385-4f05-8ca5-37315e852be6\") " pod="openstack/barbican-worker-587c8bb479-4d8m7" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.975630 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/214231d4-11de-4827-afd3-e5169b138b7b-combined-ca-bundle\") pod \"barbican-keystone-listener-75c8ccd7fb-bhzds\" (UID: \"214231d4-11de-4827-afd3-e5169b138b7b\") " pod="openstack/barbican-keystone-listener-75c8ccd7fb-bhzds" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.977801 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-65b7f6996b-q6xwd" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.979342 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cc929236-8385-4f05-8ca5-37315e852be6-config-data-custom\") pod \"barbican-worker-587c8bb479-4d8m7\" (UID: \"cc929236-8385-4f05-8ca5-37315e852be6\") " pod="openstack/barbican-worker-587c8bb479-4d8m7" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.979922 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.983967 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc929236-8385-4f05-8ca5-37315e852be6-config-data\") pod \"barbican-worker-587c8bb479-4d8m7\" (UID: \"cc929236-8385-4f05-8ca5-37315e852be6\") " pod="openstack/barbican-worker-587c8bb479-4d8m7" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.985359 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03ba9a5b-efe2-4180-8110-cd8db3e65278-config-data\") pod \"cinder-scheduler-0\" (UID: \"03ba9a5b-efe2-4180-8110-cd8db3e65278\") " pod="openstack/cinder-scheduler-0" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.986048 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/03ba9a5b-efe2-4180-8110-cd8db3e65278-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"03ba9a5b-efe2-4180-8110-cd8db3e65278\") " pod="openstack/cinder-scheduler-0" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.986146 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/214231d4-11de-4827-afd3-e5169b138b7b-config-data-custom\") pod \"barbican-keystone-listener-75c8ccd7fb-bhzds\" (UID: \"214231d4-11de-4827-afd3-e5169b138b7b\") " pod="openstack/barbican-keystone-listener-75c8ccd7fb-bhzds" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.987707 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/214231d4-11de-4827-afd3-e5169b138b7b-config-data\") pod \"barbican-keystone-listener-75c8ccd7fb-bhzds\" (UID: \"214231d4-11de-4827-afd3-e5169b138b7b\") " pod="openstack/barbican-keystone-listener-75c8ccd7fb-bhzds" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.993594 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/03ba9a5b-efe2-4180-8110-cd8db3e65278-scripts\") pod \"cinder-scheduler-0\" (UID: \"03ba9a5b-efe2-4180-8110-cd8db3e65278\") " pod="openstack/cinder-scheduler-0" Dec 07 16:19:58 crc kubenswrapper[4716]: I1207 16:19:58.996991 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5t2cp\" (UniqueName: \"kubernetes.io/projected/03ba9a5b-efe2-4180-8110-cd8db3e65278-kube-api-access-5t2cp\") pod \"cinder-scheduler-0\" (UID: \"03ba9a5b-efe2-4180-8110-cd8db3e65278\") " pod="openstack/cinder-scheduler-0" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.013560 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s84v4\" (UniqueName: \"kubernetes.io/projected/214231d4-11de-4827-afd3-e5169b138b7b-kube-api-access-s84v4\") pod \"barbican-keystone-listener-75c8ccd7fb-bhzds\" (UID: \"214231d4-11de-4827-afd3-e5169b138b7b\") " pod="openstack/barbican-keystone-listener-75c8ccd7fb-bhzds" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.019004 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-65b7f6996b-q6xwd"] Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.037823 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-75c8ccd7fb-bhzds" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.048362 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n9rkp\" (UniqueName: \"kubernetes.io/projected/cc929236-8385-4f05-8ca5-37315e852be6-kube-api-access-n9rkp\") pod \"barbican-worker-587c8bb479-4d8m7\" (UID: \"cc929236-8385-4f05-8ca5-37315e852be6\") " pod="openstack/barbican-worker-587c8bb479-4d8m7" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.057998 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.060047 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.062790 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f9195c58-0e28-495d-9b97-cce7c292b7ac-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-vb6n9\" (UID: \"f9195c58-0e28-495d-9b97-cce7c292b7ac\") " pod="openstack/dnsmasq-dns-5c9776ccc5-vb6n9" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.062954 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/65aaa058-a1be-4477-b808-d0ea1d8e15fe-config-data-custom\") pod \"barbican-api-65b7f6996b-q6xwd\" (UID: \"65aaa058-a1be-4477-b808-d0ea1d8e15fe\") " pod="openstack/barbican-api-65b7f6996b-q6xwd" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.063029 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f9195c58-0e28-495d-9b97-cce7c292b7ac-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-vb6n9\" (UID: \"f9195c58-0e28-495d-9b97-cce7c292b7ac\") " pod="openstack/dnsmasq-dns-5c9776ccc5-vb6n9" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.063124 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wjbk7\" (UniqueName: \"kubernetes.io/projected/65aaa058-a1be-4477-b808-d0ea1d8e15fe-kube-api-access-wjbk7\") pod \"barbican-api-65b7f6996b-q6xwd\" (UID: \"65aaa058-a1be-4477-b808-d0ea1d8e15fe\") " pod="openstack/barbican-api-65b7f6996b-q6xwd" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.063210 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k58zw\" (UniqueName: \"kubernetes.io/projected/f9195c58-0e28-495d-9b97-cce7c292b7ac-kube-api-access-k58zw\") pod \"dnsmasq-dns-5c9776ccc5-vb6n9\" (UID: \"f9195c58-0e28-495d-9b97-cce7c292b7ac\") " pod="openstack/dnsmasq-dns-5c9776ccc5-vb6n9" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.063310 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65aaa058-a1be-4477-b808-d0ea1d8e15fe-combined-ca-bundle\") pod \"barbican-api-65b7f6996b-q6xwd\" (UID: \"65aaa058-a1be-4477-b808-d0ea1d8e15fe\") " pod="openstack/barbican-api-65b7f6996b-q6xwd" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.063393 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/add8b3eb-3a12-4834-95cd-7839dc327eb7-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-qsdmq\" (UID: \"add8b3eb-3a12-4834-95cd-7839dc327eb7\") " pod="openstack/dnsmasq-dns-85ff748b95-qsdmq" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.063477 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f9195c58-0e28-495d-9b97-cce7c292b7ac-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-vb6n9\" (UID: \"f9195c58-0e28-495d-9b97-cce7c292b7ac\") " pod="openstack/dnsmasq-dns-5c9776ccc5-vb6n9" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.063591 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/add8b3eb-3a12-4834-95cd-7839dc327eb7-dns-svc\") pod \"dnsmasq-dns-85ff748b95-qsdmq\" (UID: \"add8b3eb-3a12-4834-95cd-7839dc327eb7\") " pod="openstack/dnsmasq-dns-85ff748b95-qsdmq" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.063661 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f9195c58-0e28-495d-9b97-cce7c292b7ac-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-vb6n9\" (UID: \"f9195c58-0e28-495d-9b97-cce7c292b7ac\") " pod="openstack/dnsmasq-dns-5c9776ccc5-vb6n9" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.063727 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gw8vm\" (UniqueName: \"kubernetes.io/projected/add8b3eb-3a12-4834-95cd-7839dc327eb7-kube-api-access-gw8vm\") pod \"dnsmasq-dns-85ff748b95-qsdmq\" (UID: \"add8b3eb-3a12-4834-95cd-7839dc327eb7\") " pod="openstack/dnsmasq-dns-85ff748b95-qsdmq" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.063805 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/add8b3eb-3a12-4834-95cd-7839dc327eb7-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-qsdmq\" (UID: \"add8b3eb-3a12-4834-95cd-7839dc327eb7\") " pod="openstack/dnsmasq-dns-85ff748b95-qsdmq" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.063903 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/65aaa058-a1be-4477-b808-d0ea1d8e15fe-logs\") pod \"barbican-api-65b7f6996b-q6xwd\" (UID: \"65aaa058-a1be-4477-b808-d0ea1d8e15fe\") " pod="openstack/barbican-api-65b7f6996b-q6xwd" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.065178 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/add8b3eb-3a12-4834-95cd-7839dc327eb7-dns-svc\") pod \"dnsmasq-dns-85ff748b95-qsdmq\" (UID: \"add8b3eb-3a12-4834-95cd-7839dc327eb7\") " pod="openstack/dnsmasq-dns-85ff748b95-qsdmq" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.064674 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/add8b3eb-3a12-4834-95cd-7839dc327eb7-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-qsdmq\" (UID: \"add8b3eb-3a12-4834-95cd-7839dc327eb7\") " pod="openstack/dnsmasq-dns-85ff748b95-qsdmq" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.065197 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65aaa058-a1be-4477-b808-d0ea1d8e15fe-config-data\") pod \"barbican-api-65b7f6996b-q6xwd\" (UID: \"65aaa058-a1be-4477-b808-d0ea1d8e15fe\") " pod="openstack/barbican-api-65b7f6996b-q6xwd" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.065445 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/add8b3eb-3a12-4834-95cd-7839dc327eb7-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-qsdmq\" (UID: \"add8b3eb-3a12-4834-95cd-7839dc327eb7\") " pod="openstack/dnsmasq-dns-85ff748b95-qsdmq" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.065575 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/add8b3eb-3a12-4834-95cd-7839dc327eb7-config\") pod \"dnsmasq-dns-85ff748b95-qsdmq\" (UID: \"add8b3eb-3a12-4834-95cd-7839dc327eb7\") " pod="openstack/dnsmasq-dns-85ff748b95-qsdmq" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.065645 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9195c58-0e28-495d-9b97-cce7c292b7ac-config\") pod \"dnsmasq-dns-5c9776ccc5-vb6n9\" (UID: \"f9195c58-0e28-495d-9b97-cce7c292b7ac\") " pod="openstack/dnsmasq-dns-5c9776ccc5-vb6n9" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.066964 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/add8b3eb-3a12-4834-95cd-7839dc327eb7-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-qsdmq\" (UID: \"add8b3eb-3a12-4834-95cd-7839dc327eb7\") " pod="openstack/dnsmasq-dns-85ff748b95-qsdmq" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.072745 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/add8b3eb-3a12-4834-95cd-7839dc327eb7-config\") pod \"dnsmasq-dns-85ff748b95-qsdmq\" (UID: \"add8b3eb-3a12-4834-95cd-7839dc327eb7\") " pod="openstack/dnsmasq-dns-85ff748b95-qsdmq" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.073709 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.076481 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.076635 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.081375 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/add8b3eb-3a12-4834-95cd-7839dc327eb7-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-qsdmq\" (UID: \"add8b3eb-3a12-4834-95cd-7839dc327eb7\") " pod="openstack/dnsmasq-dns-85ff748b95-qsdmq" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.088796 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gw8vm\" (UniqueName: \"kubernetes.io/projected/add8b3eb-3a12-4834-95cd-7839dc327eb7-kube-api-access-gw8vm\") pod \"dnsmasq-dns-85ff748b95-qsdmq\" (UID: \"add8b3eb-3a12-4834-95cd-7839dc327eb7\") " pod="openstack/dnsmasq-dns-85ff748b95-qsdmq" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.167046 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f9195c58-0e28-495d-9b97-cce7c292b7ac-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-vb6n9\" (UID: \"f9195c58-0e28-495d-9b97-cce7c292b7ac\") " pod="openstack/dnsmasq-dns-5c9776ccc5-vb6n9" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.167116 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e1dd089c-abb1-4317-aa2f-d301b7ae730d-scripts\") pod \"cinder-api-0\" (UID: \"e1dd089c-abb1-4317-aa2f-d301b7ae730d\") " pod="openstack/cinder-api-0" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.167154 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e1dd089c-abb1-4317-aa2f-d301b7ae730d-config-data-custom\") pod \"cinder-api-0\" (UID: \"e1dd089c-abb1-4317-aa2f-d301b7ae730d\") " pod="openstack/cinder-api-0" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.167184 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvvr4\" (UniqueName: \"kubernetes.io/projected/e1dd089c-abb1-4317-aa2f-d301b7ae730d-kube-api-access-nvvr4\") pod \"cinder-api-0\" (UID: \"e1dd089c-abb1-4317-aa2f-d301b7ae730d\") " pod="openstack/cinder-api-0" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.167210 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f9195c58-0e28-495d-9b97-cce7c292b7ac-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-vb6n9\" (UID: \"f9195c58-0e28-495d-9b97-cce7c292b7ac\") " pod="openstack/dnsmasq-dns-5c9776ccc5-vb6n9" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.167267 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e1dd089c-abb1-4317-aa2f-d301b7ae730d-etc-machine-id\") pod \"cinder-api-0\" (UID: \"e1dd089c-abb1-4317-aa2f-d301b7ae730d\") " pod="openstack/cinder-api-0" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.167319 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/65aaa058-a1be-4477-b808-d0ea1d8e15fe-logs\") pod \"barbican-api-65b7f6996b-q6xwd\" (UID: \"65aaa058-a1be-4477-b808-d0ea1d8e15fe\") " pod="openstack/barbican-api-65b7f6996b-q6xwd" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.167339 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65aaa058-a1be-4477-b808-d0ea1d8e15fe-config-data\") pod \"barbican-api-65b7f6996b-q6xwd\" (UID: \"65aaa058-a1be-4477-b808-d0ea1d8e15fe\") " pod="openstack/barbican-api-65b7f6996b-q6xwd" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.167364 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9195c58-0e28-495d-9b97-cce7c292b7ac-config\") pod \"dnsmasq-dns-5c9776ccc5-vb6n9\" (UID: \"f9195c58-0e28-495d-9b97-cce7c292b7ac\") " pod="openstack/dnsmasq-dns-5c9776ccc5-vb6n9" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.167390 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f9195c58-0e28-495d-9b97-cce7c292b7ac-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-vb6n9\" (UID: \"f9195c58-0e28-495d-9b97-cce7c292b7ac\") " pod="openstack/dnsmasq-dns-5c9776ccc5-vb6n9" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.167409 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/65aaa058-a1be-4477-b808-d0ea1d8e15fe-config-data-custom\") pod \"barbican-api-65b7f6996b-q6xwd\" (UID: \"65aaa058-a1be-4477-b808-d0ea1d8e15fe\") " pod="openstack/barbican-api-65b7f6996b-q6xwd" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.167427 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1dd089c-abb1-4317-aa2f-d301b7ae730d-config-data\") pod \"cinder-api-0\" (UID: \"e1dd089c-abb1-4317-aa2f-d301b7ae730d\") " pod="openstack/cinder-api-0" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.167446 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f9195c58-0e28-495d-9b97-cce7c292b7ac-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-vb6n9\" (UID: \"f9195c58-0e28-495d-9b97-cce7c292b7ac\") " pod="openstack/dnsmasq-dns-5c9776ccc5-vb6n9" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.167461 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e1dd089c-abb1-4317-aa2f-d301b7ae730d-logs\") pod \"cinder-api-0\" (UID: \"e1dd089c-abb1-4317-aa2f-d301b7ae730d\") " pod="openstack/cinder-api-0" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.167487 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wjbk7\" (UniqueName: \"kubernetes.io/projected/65aaa058-a1be-4477-b808-d0ea1d8e15fe-kube-api-access-wjbk7\") pod \"barbican-api-65b7f6996b-q6xwd\" (UID: \"65aaa058-a1be-4477-b808-d0ea1d8e15fe\") " pod="openstack/barbican-api-65b7f6996b-q6xwd" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.167512 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1dd089c-abb1-4317-aa2f-d301b7ae730d-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"e1dd089c-abb1-4317-aa2f-d301b7ae730d\") " pod="openstack/cinder-api-0" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.167531 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k58zw\" (UniqueName: \"kubernetes.io/projected/f9195c58-0e28-495d-9b97-cce7c292b7ac-kube-api-access-k58zw\") pod \"dnsmasq-dns-5c9776ccc5-vb6n9\" (UID: \"f9195c58-0e28-495d-9b97-cce7c292b7ac\") " pod="openstack/dnsmasq-dns-5c9776ccc5-vb6n9" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.167556 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65aaa058-a1be-4477-b808-d0ea1d8e15fe-combined-ca-bundle\") pod \"barbican-api-65b7f6996b-q6xwd\" (UID: \"65aaa058-a1be-4477-b808-d0ea1d8e15fe\") " pod="openstack/barbican-api-65b7f6996b-q6xwd" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.174113 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9195c58-0e28-495d-9b97-cce7c292b7ac-config\") pod \"dnsmasq-dns-5c9776ccc5-vb6n9\" (UID: \"f9195c58-0e28-495d-9b97-cce7c292b7ac\") " pod="openstack/dnsmasq-dns-5c9776ccc5-vb6n9" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.174686 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f9195c58-0e28-495d-9b97-cce7c292b7ac-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-vb6n9\" (UID: \"f9195c58-0e28-495d-9b97-cce7c292b7ac\") " pod="openstack/dnsmasq-dns-5c9776ccc5-vb6n9" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.175240 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f9195c58-0e28-495d-9b97-cce7c292b7ac-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-vb6n9\" (UID: \"f9195c58-0e28-495d-9b97-cce7c292b7ac\") " pod="openstack/dnsmasq-dns-5c9776ccc5-vb6n9" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.175575 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/65aaa058-a1be-4477-b808-d0ea1d8e15fe-logs\") pod \"barbican-api-65b7f6996b-q6xwd\" (UID: \"65aaa058-a1be-4477-b808-d0ea1d8e15fe\") " pod="openstack/barbican-api-65b7f6996b-q6xwd" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.176734 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f9195c58-0e28-495d-9b97-cce7c292b7ac-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-vb6n9\" (UID: \"f9195c58-0e28-495d-9b97-cce7c292b7ac\") " pod="openstack/dnsmasq-dns-5c9776ccc5-vb6n9" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.177335 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f9195c58-0e28-495d-9b97-cce7c292b7ac-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-vb6n9\" (UID: \"f9195c58-0e28-495d-9b97-cce7c292b7ac\") " pod="openstack/dnsmasq-dns-5c9776ccc5-vb6n9" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.206318 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65aaa058-a1be-4477-b808-d0ea1d8e15fe-config-data\") pod \"barbican-api-65b7f6996b-q6xwd\" (UID: \"65aaa058-a1be-4477-b808-d0ea1d8e15fe\") " pod="openstack/barbican-api-65b7f6996b-q6xwd" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.207770 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/65aaa058-a1be-4477-b808-d0ea1d8e15fe-config-data-custom\") pod \"barbican-api-65b7f6996b-q6xwd\" (UID: \"65aaa058-a1be-4477-b808-d0ea1d8e15fe\") " pod="openstack/barbican-api-65b7f6996b-q6xwd" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.208186 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65aaa058-a1be-4477-b808-d0ea1d8e15fe-combined-ca-bundle\") pod \"barbican-api-65b7f6996b-q6xwd\" (UID: \"65aaa058-a1be-4477-b808-d0ea1d8e15fe\") " pod="openstack/barbican-api-65b7f6996b-q6xwd" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.215912 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k58zw\" (UniqueName: \"kubernetes.io/projected/f9195c58-0e28-495d-9b97-cce7c292b7ac-kube-api-access-k58zw\") pod \"dnsmasq-dns-5c9776ccc5-vb6n9\" (UID: \"f9195c58-0e28-495d-9b97-cce7c292b7ac\") " pod="openstack/dnsmasq-dns-5c9776ccc5-vb6n9" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.219792 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wjbk7\" (UniqueName: \"kubernetes.io/projected/65aaa058-a1be-4477-b808-d0ea1d8e15fe-kube-api-access-wjbk7\") pod \"barbican-api-65b7f6996b-q6xwd\" (UID: \"65aaa058-a1be-4477-b808-d0ea1d8e15fe\") " pod="openstack/barbican-api-65b7f6996b-q6xwd" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.246543 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-vb6n9" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.270134 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e1dd089c-abb1-4317-aa2f-d301b7ae730d-scripts\") pod \"cinder-api-0\" (UID: \"e1dd089c-abb1-4317-aa2f-d301b7ae730d\") " pod="openstack/cinder-api-0" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.270344 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e1dd089c-abb1-4317-aa2f-d301b7ae730d-config-data-custom\") pod \"cinder-api-0\" (UID: \"e1dd089c-abb1-4317-aa2f-d301b7ae730d\") " pod="openstack/cinder-api-0" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.270446 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvvr4\" (UniqueName: \"kubernetes.io/projected/e1dd089c-abb1-4317-aa2f-d301b7ae730d-kube-api-access-nvvr4\") pod \"cinder-api-0\" (UID: \"e1dd089c-abb1-4317-aa2f-d301b7ae730d\") " pod="openstack/cinder-api-0" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.270550 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e1dd089c-abb1-4317-aa2f-d301b7ae730d-etc-machine-id\") pod \"cinder-api-0\" (UID: \"e1dd089c-abb1-4317-aa2f-d301b7ae730d\") " pod="openstack/cinder-api-0" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.270651 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1dd089c-abb1-4317-aa2f-d301b7ae730d-config-data\") pod \"cinder-api-0\" (UID: \"e1dd089c-abb1-4317-aa2f-d301b7ae730d\") " pod="openstack/cinder-api-0" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.270718 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e1dd089c-abb1-4317-aa2f-d301b7ae730d-logs\") pod \"cinder-api-0\" (UID: \"e1dd089c-abb1-4317-aa2f-d301b7ae730d\") " pod="openstack/cinder-api-0" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.270794 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1dd089c-abb1-4317-aa2f-d301b7ae730d-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"e1dd089c-abb1-4317-aa2f-d301b7ae730d\") " pod="openstack/cinder-api-0" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.275144 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e1dd089c-abb1-4317-aa2f-d301b7ae730d-logs\") pod \"cinder-api-0\" (UID: \"e1dd089c-abb1-4317-aa2f-d301b7ae730d\") " pod="openstack/cinder-api-0" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.275210 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e1dd089c-abb1-4317-aa2f-d301b7ae730d-etc-machine-id\") pod \"cinder-api-0\" (UID: \"e1dd089c-abb1-4317-aa2f-d301b7ae730d\") " pod="openstack/cinder-api-0" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.292003 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e1dd089c-abb1-4317-aa2f-d301b7ae730d-scripts\") pod \"cinder-api-0\" (UID: \"e1dd089c-abb1-4317-aa2f-d301b7ae730d\") " pod="openstack/cinder-api-0" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.293413 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1dd089c-abb1-4317-aa2f-d301b7ae730d-config-data\") pod \"cinder-api-0\" (UID: \"e1dd089c-abb1-4317-aa2f-d301b7ae730d\") " pod="openstack/cinder-api-0" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.301727 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1dd089c-abb1-4317-aa2f-d301b7ae730d-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"e1dd089c-abb1-4317-aa2f-d301b7ae730d\") " pod="openstack/cinder-api-0" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.303191 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-587c8bb479-4d8m7" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.314528 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e1dd089c-abb1-4317-aa2f-d301b7ae730d-config-data-custom\") pod \"cinder-api-0\" (UID: \"e1dd089c-abb1-4317-aa2f-d301b7ae730d\") " pod="openstack/cinder-api-0" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.315362 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvvr4\" (UniqueName: \"kubernetes.io/projected/e1dd089c-abb1-4317-aa2f-d301b7ae730d-kube-api-access-nvvr4\") pod \"cinder-api-0\" (UID: \"e1dd089c-abb1-4317-aa2f-d301b7ae730d\") " pod="openstack/cinder-api-0" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.456278 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-65b7f6996b-q6xwd" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.486004 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.725256 4716 generic.go:334] "Generic (PLEG): container finished" podID="2ee00bf0-1412-4fca-a570-6600db549923" containerID="bfb1a76412f365d957dd1c3a79ba7fd2b942a53d42354cb44f0eb4e1d899bebd" exitCode=0 Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.725577 4716 generic.go:334] "Generic (PLEG): container finished" podID="2ee00bf0-1412-4fca-a570-6600db549923" containerID="52d4c8b65fd2dddf05a3fa2b1d30919639544b392333691964350460d53b6375" exitCode=2 Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.725589 4716 generic.go:334] "Generic (PLEG): container finished" podID="2ee00bf0-1412-4fca-a570-6600db549923" containerID="65d1e11cc55500273d099448bfdb88d42d28a5fe058e6da987f62dc02869707b" exitCode=0 Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.725683 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-qsdmq" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.726022 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2ee00bf0-1412-4fca-a570-6600db549923","Type":"ContainerDied","Data":"bfb1a76412f365d957dd1c3a79ba7fd2b942a53d42354cb44f0eb4e1d899bebd"} Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.726100 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2ee00bf0-1412-4fca-a570-6600db549923","Type":"ContainerDied","Data":"52d4c8b65fd2dddf05a3fa2b1d30919639544b392333691964350460d53b6375"} Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.726115 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2ee00bf0-1412-4fca-a570-6600db549923","Type":"ContainerDied","Data":"65d1e11cc55500273d099448bfdb88d42d28a5fe058e6da987f62dc02869707b"} Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.744157 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-qsdmq" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.830846 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-75c8ccd7fb-bhzds"] Dec 07 16:19:59 crc kubenswrapper[4716]: W1207 16:19:59.846288 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod214231d4_11de_4827_afd3_e5169b138b7b.slice/crio-a63236f4a2bf139fcef9a06bff5cf3e739976a42385d7abc8b0406bdf85ff24c WatchSource:0}: Error finding container a63236f4a2bf139fcef9a06bff5cf3e739976a42385d7abc8b0406bdf85ff24c: Status 404 returned error can't find the container with id a63236f4a2bf139fcef9a06bff5cf3e739976a42385d7abc8b0406bdf85ff24c Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.903573 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gw8vm\" (UniqueName: \"kubernetes.io/projected/add8b3eb-3a12-4834-95cd-7839dc327eb7-kube-api-access-gw8vm\") pod \"add8b3eb-3a12-4834-95cd-7839dc327eb7\" (UID: \"add8b3eb-3a12-4834-95cd-7839dc327eb7\") " Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.904347 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/add8b3eb-3a12-4834-95cd-7839dc327eb7-config\") pod \"add8b3eb-3a12-4834-95cd-7839dc327eb7\" (UID: \"add8b3eb-3a12-4834-95cd-7839dc327eb7\") " Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.904923 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/add8b3eb-3a12-4834-95cd-7839dc327eb7-ovsdbserver-sb\") pod \"add8b3eb-3a12-4834-95cd-7839dc327eb7\" (UID: \"add8b3eb-3a12-4834-95cd-7839dc327eb7\") " Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.905033 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/add8b3eb-3a12-4834-95cd-7839dc327eb7-dns-swift-storage-0\") pod \"add8b3eb-3a12-4834-95cd-7839dc327eb7\" (UID: \"add8b3eb-3a12-4834-95cd-7839dc327eb7\") " Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.905062 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/add8b3eb-3a12-4834-95cd-7839dc327eb7-ovsdbserver-nb\") pod \"add8b3eb-3a12-4834-95cd-7839dc327eb7\" (UID: \"add8b3eb-3a12-4834-95cd-7839dc327eb7\") " Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.905614 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/add8b3eb-3a12-4834-95cd-7839dc327eb7-config" (OuterVolumeSpecName: "config") pod "add8b3eb-3a12-4834-95cd-7839dc327eb7" (UID: "add8b3eb-3a12-4834-95cd-7839dc327eb7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.908519 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/add8b3eb-3a12-4834-95cd-7839dc327eb7-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "add8b3eb-3a12-4834-95cd-7839dc327eb7" (UID: "add8b3eb-3a12-4834-95cd-7839dc327eb7"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.908648 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/add8b3eb-3a12-4834-95cd-7839dc327eb7-dns-svc\") pod \"add8b3eb-3a12-4834-95cd-7839dc327eb7\" (UID: \"add8b3eb-3a12-4834-95cd-7839dc327eb7\") " Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.908659 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/add8b3eb-3a12-4834-95cd-7839dc327eb7-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "add8b3eb-3a12-4834-95cd-7839dc327eb7" (UID: "add8b3eb-3a12-4834-95cd-7839dc327eb7"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.909017 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/add8b3eb-3a12-4834-95cd-7839dc327eb7-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "add8b3eb-3a12-4834-95cd-7839dc327eb7" (UID: "add8b3eb-3a12-4834-95cd-7839dc327eb7"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.909563 4716 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/add8b3eb-3a12-4834-95cd-7839dc327eb7-config\") on node \"crc\" DevicePath \"\"" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.909583 4716 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/add8b3eb-3a12-4834-95cd-7839dc327eb7-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.909594 4716 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/add8b3eb-3a12-4834-95cd-7839dc327eb7-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.909604 4716 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/add8b3eb-3a12-4834-95cd-7839dc327eb7-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.910193 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/add8b3eb-3a12-4834-95cd-7839dc327eb7-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "add8b3eb-3a12-4834-95cd-7839dc327eb7" (UID: "add8b3eb-3a12-4834-95cd-7839dc327eb7"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.917048 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/add8b3eb-3a12-4834-95cd-7839dc327eb7-kube-api-access-gw8vm" (OuterVolumeSpecName: "kube-api-access-gw8vm") pod "add8b3eb-3a12-4834-95cd-7839dc327eb7" (UID: "add8b3eb-3a12-4834-95cd-7839dc327eb7"). InnerVolumeSpecName "kube-api-access-gw8vm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:19:59 crc kubenswrapper[4716]: I1207 16:19:59.931831 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-vb6n9"] Dec 07 16:19:59 crc kubenswrapper[4716]: W1207 16:19:59.932667 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf9195c58_0e28_495d_9b97_cce7c292b7ac.slice/crio-02f0c3164652010b3ae9c510cbde541b845559b46876f389528a76f1c0e93832 WatchSource:0}: Error finding container 02f0c3164652010b3ae9c510cbde541b845559b46876f389528a76f1c0e93832: Status 404 returned error can't find the container with id 02f0c3164652010b3ae9c510cbde541b845559b46876f389528a76f1c0e93832 Dec 07 16:20:00 crc kubenswrapper[4716]: I1207 16:20:00.011990 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gw8vm\" (UniqueName: \"kubernetes.io/projected/add8b3eb-3a12-4834-95cd-7839dc327eb7-kube-api-access-gw8vm\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:00 crc kubenswrapper[4716]: I1207 16:20:00.012019 4716 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/add8b3eb-3a12-4834-95cd-7839dc327eb7-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:00 crc kubenswrapper[4716]: I1207 16:20:00.018169 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-587c8bb479-4d8m7"] Dec 07 16:20:00 crc kubenswrapper[4716]: W1207 16:20:00.020517 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcc929236_8385_4f05_8ca5_37315e852be6.slice/crio-f62b95079cc47b1812fb039ec28bf6602fe286cd2a8c1a701366319bb89e4233 WatchSource:0}: Error finding container f62b95079cc47b1812fb039ec28bf6602fe286cd2a8c1a701366319bb89e4233: Status 404 returned error can't find the container with id f62b95079cc47b1812fb039ec28bf6602fe286cd2a8c1a701366319bb89e4233 Dec 07 16:20:00 crc kubenswrapper[4716]: W1207 16:20:00.023391 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod03ba9a5b_efe2_4180_8110_cd8db3e65278.slice/crio-07eb774f0cfe53a228c73435dc6bfeab5cabcce9cf02b1271fe77ebd6c1a6044 WatchSource:0}: Error finding container 07eb774f0cfe53a228c73435dc6bfeab5cabcce9cf02b1271fe77ebd6c1a6044: Status 404 returned error can't find the container with id 07eb774f0cfe53a228c73435dc6bfeab5cabcce9cf02b1271fe77ebd6c1a6044 Dec 07 16:20:00 crc kubenswrapper[4716]: I1207 16:20:00.023872 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 07 16:20:00 crc kubenswrapper[4716]: I1207 16:20:00.134195 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-65b7f6996b-q6xwd"] Dec 07 16:20:00 crc kubenswrapper[4716]: I1207 16:20:00.140545 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 07 16:20:00 crc kubenswrapper[4716]: W1207 16:20:00.145105 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode1dd089c_abb1_4317_aa2f_d301b7ae730d.slice/crio-2021e1edd583374810e404464702da8eb2de94144082e419cd74ae66b0abbc18 WatchSource:0}: Error finding container 2021e1edd583374810e404464702da8eb2de94144082e419cd74ae66b0abbc18: Status 404 returned error can't find the container with id 2021e1edd583374810e404464702da8eb2de94144082e419cd74ae66b0abbc18 Dec 07 16:20:00 crc kubenswrapper[4716]: I1207 16:20:00.761420 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"03ba9a5b-efe2-4180-8110-cd8db3e65278","Type":"ContainerStarted","Data":"07eb774f0cfe53a228c73435dc6bfeab5cabcce9cf02b1271fe77ebd6c1a6044"} Dec 07 16:20:00 crc kubenswrapper[4716]: I1207 16:20:00.764717 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e1dd089c-abb1-4317-aa2f-d301b7ae730d","Type":"ContainerStarted","Data":"2021e1edd583374810e404464702da8eb2de94144082e419cd74ae66b0abbc18"} Dec 07 16:20:00 crc kubenswrapper[4716]: I1207 16:20:00.790489 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-587c8bb479-4d8m7" event={"ID":"cc929236-8385-4f05-8ca5-37315e852be6","Type":"ContainerStarted","Data":"f62b95079cc47b1812fb039ec28bf6602fe286cd2a8c1a701366319bb89e4233"} Dec 07 16:20:00 crc kubenswrapper[4716]: I1207 16:20:00.793619 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-75c8ccd7fb-bhzds" event={"ID":"214231d4-11de-4827-afd3-e5169b138b7b","Type":"ContainerStarted","Data":"a63236f4a2bf139fcef9a06bff5cf3e739976a42385d7abc8b0406bdf85ff24c"} Dec 07 16:20:00 crc kubenswrapper[4716]: I1207 16:20:00.800146 4716 generic.go:334] "Generic (PLEG): container finished" podID="f9195c58-0e28-495d-9b97-cce7c292b7ac" containerID="1d0c3f6589d32291dc7ce091aaf5272b16e51ae7368ca4e1dce1771c26b7b79c" exitCode=0 Dec 07 16:20:00 crc kubenswrapper[4716]: I1207 16:20:00.800258 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-vb6n9" event={"ID":"f9195c58-0e28-495d-9b97-cce7c292b7ac","Type":"ContainerDied","Data":"1d0c3f6589d32291dc7ce091aaf5272b16e51ae7368ca4e1dce1771c26b7b79c"} Dec 07 16:20:00 crc kubenswrapper[4716]: I1207 16:20:00.800291 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-vb6n9" event={"ID":"f9195c58-0e28-495d-9b97-cce7c292b7ac","Type":"ContainerStarted","Data":"02f0c3164652010b3ae9c510cbde541b845559b46876f389528a76f1c0e93832"} Dec 07 16:20:00 crc kubenswrapper[4716]: I1207 16:20:00.812751 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-qsdmq" Dec 07 16:20:00 crc kubenswrapper[4716]: I1207 16:20:00.813825 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-65b7f6996b-q6xwd" event={"ID":"65aaa058-a1be-4477-b808-d0ea1d8e15fe","Type":"ContainerStarted","Data":"e21b4da38239a0d684080040edbc4305151f056711ac673ac6b4cb5a5648470c"} Dec 07 16:20:00 crc kubenswrapper[4716]: I1207 16:20:00.813865 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-65b7f6996b-q6xwd" event={"ID":"65aaa058-a1be-4477-b808-d0ea1d8e15fe","Type":"ContainerStarted","Data":"110da32e8b7d13bd7f08d374547e0e1ba723851317ad3a5d5cff00362d5ca8b4"} Dec 07 16:20:00 crc kubenswrapper[4716]: I1207 16:20:00.813878 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-65b7f6996b-q6xwd" event={"ID":"65aaa058-a1be-4477-b808-d0ea1d8e15fe","Type":"ContainerStarted","Data":"984ec5665a5889d05f6cc8d4025b879e51827ad9c65a2b8b61ddd700caab4c68"} Dec 07 16:20:00 crc kubenswrapper[4716]: I1207 16:20:00.815480 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-65b7f6996b-q6xwd" Dec 07 16:20:00 crc kubenswrapper[4716]: I1207 16:20:00.815519 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-65b7f6996b-q6xwd" Dec 07 16:20:00 crc kubenswrapper[4716]: I1207 16:20:00.844038 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 07 16:20:00 crc kubenswrapper[4716]: I1207 16:20:00.861901 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-65b7f6996b-q6xwd" podStartSLOduration=2.861881342 podStartE2EDuration="2.861881342s" podCreationTimestamp="2025-12-07 16:19:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:20:00.86069205 +0000 UTC m=+1063.550976962" watchObservedRunningTime="2025-12-07 16:20:00.861881342 +0000 UTC m=+1063.552166254" Dec 07 16:20:00 crc kubenswrapper[4716]: I1207 16:20:00.983300 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-qsdmq"] Dec 07 16:20:00 crc kubenswrapper[4716]: I1207 16:20:00.989484 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-qsdmq"] Dec 07 16:20:01 crc kubenswrapper[4716]: I1207 16:20:01.669384 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="add8b3eb-3a12-4834-95cd-7839dc327eb7" path="/var/lib/kubelet/pods/add8b3eb-3a12-4834-95cd-7839dc327eb7/volumes" Dec 07 16:20:01 crc kubenswrapper[4716]: I1207 16:20:01.839719 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-vb6n9" event={"ID":"f9195c58-0e28-495d-9b97-cce7c292b7ac","Type":"ContainerStarted","Data":"d55ffbf42228f71084be488104e19559e1a1a664a13930078576be29b62e11c6"} Dec 07 16:20:01 crc kubenswrapper[4716]: I1207 16:20:01.839943 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c9776ccc5-vb6n9" Dec 07 16:20:01 crc kubenswrapper[4716]: I1207 16:20:01.841988 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"03ba9a5b-efe2-4180-8110-cd8db3e65278","Type":"ContainerStarted","Data":"d9f29be0066946f7c67371ca40478054a4ae180dd2ab01cef84d12d1dde9ec07"} Dec 07 16:20:01 crc kubenswrapper[4716]: I1207 16:20:01.844062 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e1dd089c-abb1-4317-aa2f-d301b7ae730d","Type":"ContainerStarted","Data":"53eb298d61c184858cf42d228a84ba335e5e977399e58be6ed004d8ed973b4ba"} Dec 07 16:20:01 crc kubenswrapper[4716]: I1207 16:20:01.844106 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e1dd089c-abb1-4317-aa2f-d301b7ae730d","Type":"ContainerStarted","Data":"2d303dd7a108d3836287cc7f009cc60f84e17a4753a55a815b95ca9a76d71969"} Dec 07 16:20:01 crc kubenswrapper[4716]: I1207 16:20:01.844258 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="e1dd089c-abb1-4317-aa2f-d301b7ae730d" containerName="cinder-api-log" containerID="cri-o://2d303dd7a108d3836287cc7f009cc60f84e17a4753a55a815b95ca9a76d71969" gracePeriod=30 Dec 07 16:20:01 crc kubenswrapper[4716]: I1207 16:20:01.844349 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="e1dd089c-abb1-4317-aa2f-d301b7ae730d" containerName="cinder-api" containerID="cri-o://53eb298d61c184858cf42d228a84ba335e5e977399e58be6ed004d8ed973b4ba" gracePeriod=30 Dec 07 16:20:01 crc kubenswrapper[4716]: I1207 16:20:01.899923 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c9776ccc5-vb6n9" podStartSLOduration=3.897024755 podStartE2EDuration="3.897024755s" podCreationTimestamp="2025-12-07 16:19:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:20:01.86617319 +0000 UTC m=+1064.556458122" watchObservedRunningTime="2025-12-07 16:20:01.897024755 +0000 UTC m=+1064.587309677" Dec 07 16:20:01 crc kubenswrapper[4716]: I1207 16:20:01.902035 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.902023629 podStartE2EDuration="3.902023629s" podCreationTimestamp="2025-12-07 16:19:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:20:01.890000734 +0000 UTC m=+1064.580285646" watchObservedRunningTime="2025-12-07 16:20:01.902023629 +0000 UTC m=+1064.592308551" Dec 07 16:20:02 crc kubenswrapper[4716]: I1207 16:20:02.654807 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-69d87fd46d-9k5rq" Dec 07 16:20:02 crc kubenswrapper[4716]: I1207 16:20:02.722592 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-86c7567d4-99rx9" Dec 07 16:20:02 crc kubenswrapper[4716]: I1207 16:20:02.868874 4716 generic.go:334] "Generic (PLEG): container finished" podID="e1dd089c-abb1-4317-aa2f-d301b7ae730d" containerID="2d303dd7a108d3836287cc7f009cc60f84e17a4753a55a815b95ca9a76d71969" exitCode=143 Dec 07 16:20:02 crc kubenswrapper[4716]: I1207 16:20:02.869271 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e1dd089c-abb1-4317-aa2f-d301b7ae730d","Type":"ContainerDied","Data":"2d303dd7a108d3836287cc7f009cc60f84e17a4753a55a815b95ca9a76d71969"} Dec 07 16:20:02 crc kubenswrapper[4716]: I1207 16:20:02.877105 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-587c8bb479-4d8m7" event={"ID":"cc929236-8385-4f05-8ca5-37315e852be6","Type":"ContainerStarted","Data":"f3a78c1ee388341cffad29e225b0591dddf66b83b07838e7c7a6bc879ca05ee0"} Dec 07 16:20:02 crc kubenswrapper[4716]: I1207 16:20:02.881906 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-75c8ccd7fb-bhzds" event={"ID":"214231d4-11de-4827-afd3-e5169b138b7b","Type":"ContainerStarted","Data":"ae81bdb9629958b3807ec7dfb903509826e6a0c85218683cc4274de3a3f10a2d"} Dec 07 16:20:02 crc kubenswrapper[4716]: I1207 16:20:02.882043 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 07 16:20:02 crc kubenswrapper[4716]: I1207 16:20:02.887635 4716 generic.go:334] "Generic (PLEG): container finished" podID="2ee00bf0-1412-4fca-a570-6600db549923" containerID="411ac652ee7365d2f2c78ae6705384e1dadf6242d65ea1cab9c0a5600032ff1d" exitCode=0 Dec 07 16:20:02 crc kubenswrapper[4716]: I1207 16:20:02.887976 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2ee00bf0-1412-4fca-a570-6600db549923","Type":"ContainerDied","Data":"411ac652ee7365d2f2c78ae6705384e1dadf6242d65ea1cab9c0a5600032ff1d"} Dec 07 16:20:02 crc kubenswrapper[4716]: I1207 16:20:02.888041 4716 scope.go:117] "RemoveContainer" containerID="bfb1a76412f365d957dd1c3a79ba7fd2b942a53d42354cb44f0eb4e1d899bebd" Dec 07 16:20:02 crc kubenswrapper[4716]: I1207 16:20:02.924054 4716 scope.go:117] "RemoveContainer" containerID="52d4c8b65fd2dddf05a3fa2b1d30919639544b392333691964350460d53b6375" Dec 07 16:20:02 crc kubenswrapper[4716]: I1207 16:20:02.967035 4716 scope.go:117] "RemoveContainer" containerID="411ac652ee7365d2f2c78ae6705384e1dadf6242d65ea1cab9c0a5600032ff1d" Dec 07 16:20:02 crc kubenswrapper[4716]: I1207 16:20:02.977793 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2ee00bf0-1412-4fca-a570-6600db549923-scripts\") pod \"2ee00bf0-1412-4fca-a570-6600db549923\" (UID: \"2ee00bf0-1412-4fca-a570-6600db549923\") " Dec 07 16:20:02 crc kubenswrapper[4716]: I1207 16:20:02.977860 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2ee00bf0-1412-4fca-a570-6600db549923-sg-core-conf-yaml\") pod \"2ee00bf0-1412-4fca-a570-6600db549923\" (UID: \"2ee00bf0-1412-4fca-a570-6600db549923\") " Dec 07 16:20:02 crc kubenswrapper[4716]: I1207 16:20:02.977918 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hq68h\" (UniqueName: \"kubernetes.io/projected/2ee00bf0-1412-4fca-a570-6600db549923-kube-api-access-hq68h\") pod \"2ee00bf0-1412-4fca-a570-6600db549923\" (UID: \"2ee00bf0-1412-4fca-a570-6600db549923\") " Dec 07 16:20:02 crc kubenswrapper[4716]: I1207 16:20:02.977965 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2ee00bf0-1412-4fca-a570-6600db549923-log-httpd\") pod \"2ee00bf0-1412-4fca-a570-6600db549923\" (UID: \"2ee00bf0-1412-4fca-a570-6600db549923\") " Dec 07 16:20:02 crc kubenswrapper[4716]: I1207 16:20:02.977997 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2ee00bf0-1412-4fca-a570-6600db549923-run-httpd\") pod \"2ee00bf0-1412-4fca-a570-6600db549923\" (UID: \"2ee00bf0-1412-4fca-a570-6600db549923\") " Dec 07 16:20:02 crc kubenswrapper[4716]: I1207 16:20:02.978013 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ee00bf0-1412-4fca-a570-6600db549923-config-data\") pod \"2ee00bf0-1412-4fca-a570-6600db549923\" (UID: \"2ee00bf0-1412-4fca-a570-6600db549923\") " Dec 07 16:20:02 crc kubenswrapper[4716]: I1207 16:20:02.978050 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ee00bf0-1412-4fca-a570-6600db549923-combined-ca-bundle\") pod \"2ee00bf0-1412-4fca-a570-6600db549923\" (UID: \"2ee00bf0-1412-4fca-a570-6600db549923\") " Dec 07 16:20:02 crc kubenswrapper[4716]: I1207 16:20:02.979310 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ee00bf0-1412-4fca-a570-6600db549923-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "2ee00bf0-1412-4fca-a570-6600db549923" (UID: "2ee00bf0-1412-4fca-a570-6600db549923"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:20:02 crc kubenswrapper[4716]: I1207 16:20:02.979591 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ee00bf0-1412-4fca-a570-6600db549923-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "2ee00bf0-1412-4fca-a570-6600db549923" (UID: "2ee00bf0-1412-4fca-a570-6600db549923"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:20:02 crc kubenswrapper[4716]: I1207 16:20:02.984361 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ee00bf0-1412-4fca-a570-6600db549923-kube-api-access-hq68h" (OuterVolumeSpecName: "kube-api-access-hq68h") pod "2ee00bf0-1412-4fca-a570-6600db549923" (UID: "2ee00bf0-1412-4fca-a570-6600db549923"). InnerVolumeSpecName "kube-api-access-hq68h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:20:02 crc kubenswrapper[4716]: I1207 16:20:02.993662 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ee00bf0-1412-4fca-a570-6600db549923-scripts" (OuterVolumeSpecName: "scripts") pod "2ee00bf0-1412-4fca-a570-6600db549923" (UID: "2ee00bf0-1412-4fca-a570-6600db549923"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:20:03 crc kubenswrapper[4716]: I1207 16:20:03.012115 4716 scope.go:117] "RemoveContainer" containerID="65d1e11cc55500273d099448bfdb88d42d28a5fe058e6da987f62dc02869707b" Dec 07 16:20:03 crc kubenswrapper[4716]: I1207 16:20:03.017578 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ee00bf0-1412-4fca-a570-6600db549923-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "2ee00bf0-1412-4fca-a570-6600db549923" (UID: "2ee00bf0-1412-4fca-a570-6600db549923"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:20:03 crc kubenswrapper[4716]: I1207 16:20:03.079188 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ee00bf0-1412-4fca-a570-6600db549923-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2ee00bf0-1412-4fca-a570-6600db549923" (UID: "2ee00bf0-1412-4fca-a570-6600db549923"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:20:03 crc kubenswrapper[4716]: I1207 16:20:03.080471 4716 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ee00bf0-1412-4fca-a570-6600db549923-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:03 crc kubenswrapper[4716]: I1207 16:20:03.080494 4716 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2ee00bf0-1412-4fca-a570-6600db549923-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:03 crc kubenswrapper[4716]: I1207 16:20:03.080503 4716 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2ee00bf0-1412-4fca-a570-6600db549923-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:03 crc kubenswrapper[4716]: I1207 16:20:03.080511 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hq68h\" (UniqueName: \"kubernetes.io/projected/2ee00bf0-1412-4fca-a570-6600db549923-kube-api-access-hq68h\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:03 crc kubenswrapper[4716]: I1207 16:20:03.080523 4716 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2ee00bf0-1412-4fca-a570-6600db549923-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:03 crc kubenswrapper[4716]: I1207 16:20:03.080533 4716 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2ee00bf0-1412-4fca-a570-6600db549923-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:03 crc kubenswrapper[4716]: I1207 16:20:03.119248 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ee00bf0-1412-4fca-a570-6600db549923-config-data" (OuterVolumeSpecName: "config-data") pod "2ee00bf0-1412-4fca-a570-6600db549923" (UID: "2ee00bf0-1412-4fca-a570-6600db549923"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:20:03 crc kubenswrapper[4716]: I1207 16:20:03.184470 4716 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ee00bf0-1412-4fca-a570-6600db549923-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:03 crc kubenswrapper[4716]: I1207 16:20:03.272505 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-7f9b78866d-9g9kv" Dec 07 16:20:03 crc kubenswrapper[4716]: I1207 16:20:03.484532 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7d76599897-4mhqt" Dec 07 16:20:03 crc kubenswrapper[4716]: I1207 16:20:03.593591 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5c76ced8-9170-4ccc-a497-87388e8546dd-config-data\") pod \"5c76ced8-9170-4ccc-a497-87388e8546dd\" (UID: \"5c76ced8-9170-4ccc-a497-87388e8546dd\") " Dec 07 16:20:03 crc kubenswrapper[4716]: I1207 16:20:03.593940 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5c76ced8-9170-4ccc-a497-87388e8546dd-horizon-secret-key\") pod \"5c76ced8-9170-4ccc-a497-87388e8546dd\" (UID: \"5c76ced8-9170-4ccc-a497-87388e8546dd\") " Dec 07 16:20:03 crc kubenswrapper[4716]: I1207 16:20:03.593975 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5c76ced8-9170-4ccc-a497-87388e8546dd-scripts\") pod \"5c76ced8-9170-4ccc-a497-87388e8546dd\" (UID: \"5c76ced8-9170-4ccc-a497-87388e8546dd\") " Dec 07 16:20:03 crc kubenswrapper[4716]: I1207 16:20:03.594063 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8w7nq\" (UniqueName: \"kubernetes.io/projected/5c76ced8-9170-4ccc-a497-87388e8546dd-kube-api-access-8w7nq\") pod \"5c76ced8-9170-4ccc-a497-87388e8546dd\" (UID: \"5c76ced8-9170-4ccc-a497-87388e8546dd\") " Dec 07 16:20:03 crc kubenswrapper[4716]: I1207 16:20:03.594129 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c76ced8-9170-4ccc-a497-87388e8546dd-logs\") pod \"5c76ced8-9170-4ccc-a497-87388e8546dd\" (UID: \"5c76ced8-9170-4ccc-a497-87388e8546dd\") " Dec 07 16:20:03 crc kubenswrapper[4716]: I1207 16:20:03.594952 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c76ced8-9170-4ccc-a497-87388e8546dd-logs" (OuterVolumeSpecName: "logs") pod "5c76ced8-9170-4ccc-a497-87388e8546dd" (UID: "5c76ced8-9170-4ccc-a497-87388e8546dd"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:20:03 crc kubenswrapper[4716]: I1207 16:20:03.600355 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c76ced8-9170-4ccc-a497-87388e8546dd-kube-api-access-8w7nq" (OuterVolumeSpecName: "kube-api-access-8w7nq") pod "5c76ced8-9170-4ccc-a497-87388e8546dd" (UID: "5c76ced8-9170-4ccc-a497-87388e8546dd"). InnerVolumeSpecName "kube-api-access-8w7nq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:20:03 crc kubenswrapper[4716]: I1207 16:20:03.602214 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c76ced8-9170-4ccc-a497-87388e8546dd-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "5c76ced8-9170-4ccc-a497-87388e8546dd" (UID: "5c76ced8-9170-4ccc-a497-87388e8546dd"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:20:03 crc kubenswrapper[4716]: I1207 16:20:03.617727 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c76ced8-9170-4ccc-a497-87388e8546dd-config-data" (OuterVolumeSpecName: "config-data") pod "5c76ced8-9170-4ccc-a497-87388e8546dd" (UID: "5c76ced8-9170-4ccc-a497-87388e8546dd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:20:03 crc kubenswrapper[4716]: I1207 16:20:03.638257 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c76ced8-9170-4ccc-a497-87388e8546dd-scripts" (OuterVolumeSpecName: "scripts") pod "5c76ced8-9170-4ccc-a497-87388e8546dd" (UID: "5c76ced8-9170-4ccc-a497-87388e8546dd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:20:03 crc kubenswrapper[4716]: I1207 16:20:03.696296 4716 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5c76ced8-9170-4ccc-a497-87388e8546dd-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:03 crc kubenswrapper[4716]: I1207 16:20:03.696327 4716 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5c76ced8-9170-4ccc-a497-87388e8546dd-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:03 crc kubenswrapper[4716]: I1207 16:20:03.696338 4716 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5c76ced8-9170-4ccc-a497-87388e8546dd-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:03 crc kubenswrapper[4716]: I1207 16:20:03.696348 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8w7nq\" (UniqueName: \"kubernetes.io/projected/5c76ced8-9170-4ccc-a497-87388e8546dd-kube-api-access-8w7nq\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:03 crc kubenswrapper[4716]: I1207 16:20:03.696357 4716 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c76ced8-9170-4ccc-a497-87388e8546dd-logs\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:03 crc kubenswrapper[4716]: I1207 16:20:03.724525 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7d8b488bdf-fvlzm" Dec 07 16:20:03 crc kubenswrapper[4716]: I1207 16:20:03.797857 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cnh2x\" (UniqueName: \"kubernetes.io/projected/d7db6aff-6831-49ea-9d51-24ec9ac40a88-kube-api-access-cnh2x\") pod \"d7db6aff-6831-49ea-9d51-24ec9ac40a88\" (UID: \"d7db6aff-6831-49ea-9d51-24ec9ac40a88\") " Dec 07 16:20:03 crc kubenswrapper[4716]: I1207 16:20:03.798022 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d7db6aff-6831-49ea-9d51-24ec9ac40a88-horizon-secret-key\") pod \"d7db6aff-6831-49ea-9d51-24ec9ac40a88\" (UID: \"d7db6aff-6831-49ea-9d51-24ec9ac40a88\") " Dec 07 16:20:03 crc kubenswrapper[4716]: I1207 16:20:03.798052 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d7db6aff-6831-49ea-9d51-24ec9ac40a88-config-data\") pod \"d7db6aff-6831-49ea-9d51-24ec9ac40a88\" (UID: \"d7db6aff-6831-49ea-9d51-24ec9ac40a88\") " Dec 07 16:20:03 crc kubenswrapper[4716]: I1207 16:20:03.798239 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d7db6aff-6831-49ea-9d51-24ec9ac40a88-scripts\") pod \"d7db6aff-6831-49ea-9d51-24ec9ac40a88\" (UID: \"d7db6aff-6831-49ea-9d51-24ec9ac40a88\") " Dec 07 16:20:03 crc kubenswrapper[4716]: I1207 16:20:03.798273 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d7db6aff-6831-49ea-9d51-24ec9ac40a88-logs\") pod \"d7db6aff-6831-49ea-9d51-24ec9ac40a88\" (UID: \"d7db6aff-6831-49ea-9d51-24ec9ac40a88\") " Dec 07 16:20:03 crc kubenswrapper[4716]: I1207 16:20:03.798934 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d7db6aff-6831-49ea-9d51-24ec9ac40a88-logs" (OuterVolumeSpecName: "logs") pod "d7db6aff-6831-49ea-9d51-24ec9ac40a88" (UID: "d7db6aff-6831-49ea-9d51-24ec9ac40a88"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:20:03 crc kubenswrapper[4716]: I1207 16:20:03.801423 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7db6aff-6831-49ea-9d51-24ec9ac40a88-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "d7db6aff-6831-49ea-9d51-24ec9ac40a88" (UID: "d7db6aff-6831-49ea-9d51-24ec9ac40a88"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:20:03 crc kubenswrapper[4716]: I1207 16:20:03.803020 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7db6aff-6831-49ea-9d51-24ec9ac40a88-kube-api-access-cnh2x" (OuterVolumeSpecName: "kube-api-access-cnh2x") pod "d7db6aff-6831-49ea-9d51-24ec9ac40a88" (UID: "d7db6aff-6831-49ea-9d51-24ec9ac40a88"). InnerVolumeSpecName "kube-api-access-cnh2x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:20:03 crc kubenswrapper[4716]: I1207 16:20:03.827017 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7db6aff-6831-49ea-9d51-24ec9ac40a88-config-data" (OuterVolumeSpecName: "config-data") pod "d7db6aff-6831-49ea-9d51-24ec9ac40a88" (UID: "d7db6aff-6831-49ea-9d51-24ec9ac40a88"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:20:03 crc kubenswrapper[4716]: I1207 16:20:03.831449 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7db6aff-6831-49ea-9d51-24ec9ac40a88-scripts" (OuterVolumeSpecName: "scripts") pod "d7db6aff-6831-49ea-9d51-24ec9ac40a88" (UID: "d7db6aff-6831-49ea-9d51-24ec9ac40a88"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:20:03 crc kubenswrapper[4716]: I1207 16:20:03.898486 4716 generic.go:334] "Generic (PLEG): container finished" podID="5c76ced8-9170-4ccc-a497-87388e8546dd" containerID="c372e8f6fe59dccc9147f9a8695d09a6bc9fc695005f54aabb932a3697e49115" exitCode=137 Dec 07 16:20:03 crc kubenswrapper[4716]: I1207 16:20:03.898521 4716 generic.go:334] "Generic (PLEG): container finished" podID="5c76ced8-9170-4ccc-a497-87388e8546dd" containerID="008406d13a6ea7c73401539a525b2aa9a7d98be0c715633e8d8eac24e5c7ccf6" exitCode=137 Dec 07 16:20:03 crc kubenswrapper[4716]: I1207 16:20:03.898549 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7d76599897-4mhqt" Dec 07 16:20:03 crc kubenswrapper[4716]: I1207 16:20:03.898551 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d76599897-4mhqt" event={"ID":"5c76ced8-9170-4ccc-a497-87388e8546dd","Type":"ContainerDied","Data":"c372e8f6fe59dccc9147f9a8695d09a6bc9fc695005f54aabb932a3697e49115"} Dec 07 16:20:03 crc kubenswrapper[4716]: I1207 16:20:03.898648 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d76599897-4mhqt" event={"ID":"5c76ced8-9170-4ccc-a497-87388e8546dd","Type":"ContainerDied","Data":"008406d13a6ea7c73401539a525b2aa9a7d98be0c715633e8d8eac24e5c7ccf6"} Dec 07 16:20:03 crc kubenswrapper[4716]: I1207 16:20:03.898661 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d76599897-4mhqt" event={"ID":"5c76ced8-9170-4ccc-a497-87388e8546dd","Type":"ContainerDied","Data":"b23cf3815dfabcc10c50609de078972833d877d6fe2f266489c3f08fd714a42c"} Dec 07 16:20:03 crc kubenswrapper[4716]: I1207 16:20:03.898690 4716 scope.go:117] "RemoveContainer" containerID="c372e8f6fe59dccc9147f9a8695d09a6bc9fc695005f54aabb932a3697e49115" Dec 07 16:20:03 crc kubenswrapper[4716]: I1207 16:20:03.899995 4716 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d7db6aff-6831-49ea-9d51-24ec9ac40a88-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:03 crc kubenswrapper[4716]: I1207 16:20:03.900023 4716 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d7db6aff-6831-49ea-9d51-24ec9ac40a88-logs\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:03 crc kubenswrapper[4716]: I1207 16:20:03.900034 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cnh2x\" (UniqueName: \"kubernetes.io/projected/d7db6aff-6831-49ea-9d51-24ec9ac40a88-kube-api-access-cnh2x\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:03 crc kubenswrapper[4716]: I1207 16:20:03.900048 4716 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d7db6aff-6831-49ea-9d51-24ec9ac40a88-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:03 crc kubenswrapper[4716]: I1207 16:20:03.900058 4716 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d7db6aff-6831-49ea-9d51-24ec9ac40a88-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:03 crc kubenswrapper[4716]: I1207 16:20:03.900867 4716 generic.go:334] "Generic (PLEG): container finished" podID="d7db6aff-6831-49ea-9d51-24ec9ac40a88" containerID="e49f645f546db04a604348b06b91c87e2a6645528ec78ab475e94e276e8c3f6b" exitCode=137 Dec 07 16:20:03 crc kubenswrapper[4716]: I1207 16:20:03.900903 4716 generic.go:334] "Generic (PLEG): container finished" podID="d7db6aff-6831-49ea-9d51-24ec9ac40a88" containerID="d9c99d90e0f76afdd5481a03c49c1a155a22721530d081f1e90302e0f9d9fc9d" exitCode=137 Dec 07 16:20:03 crc kubenswrapper[4716]: I1207 16:20:03.900947 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7d8b488bdf-fvlzm" Dec 07 16:20:03 crc kubenswrapper[4716]: I1207 16:20:03.900897 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d8b488bdf-fvlzm" event={"ID":"d7db6aff-6831-49ea-9d51-24ec9ac40a88","Type":"ContainerDied","Data":"e49f645f546db04a604348b06b91c87e2a6645528ec78ab475e94e276e8c3f6b"} Dec 07 16:20:03 crc kubenswrapper[4716]: I1207 16:20:03.901101 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d8b488bdf-fvlzm" event={"ID":"d7db6aff-6831-49ea-9d51-24ec9ac40a88","Type":"ContainerDied","Data":"d9c99d90e0f76afdd5481a03c49c1a155a22721530d081f1e90302e0f9d9fc9d"} Dec 07 16:20:03 crc kubenswrapper[4716]: I1207 16:20:03.901116 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d8b488bdf-fvlzm" event={"ID":"d7db6aff-6831-49ea-9d51-24ec9ac40a88","Type":"ContainerDied","Data":"05f9d64947bebe40459dd781411ba7578a23432192b23ecd871939ba96d606d4"} Dec 07 16:20:03 crc kubenswrapper[4716]: I1207 16:20:03.904328 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-587c8bb479-4d8m7" event={"ID":"cc929236-8385-4f05-8ca5-37315e852be6","Type":"ContainerStarted","Data":"5cec17038640c075664e3b5f8861e649716bb95aa362e936e114bdbd6dd1333c"} Dec 07 16:20:03 crc kubenswrapper[4716]: I1207 16:20:03.911603 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-75c8ccd7fb-bhzds" event={"ID":"214231d4-11de-4827-afd3-e5169b138b7b","Type":"ContainerStarted","Data":"083087d73220e6a7ba480c60f521edf3dbc99a7bb980539e8f3fc60a311605c4"} Dec 07 16:20:03 crc kubenswrapper[4716]: I1207 16:20:03.913715 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2ee00bf0-1412-4fca-a570-6600db549923","Type":"ContainerDied","Data":"b0681f205b22e31369df55d644f1428738617f1f21c4866b0798f167f6aee150"} Dec 07 16:20:03 crc kubenswrapper[4716]: I1207 16:20:03.913828 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 07 16:20:03 crc kubenswrapper[4716]: I1207 16:20:03.933663 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"03ba9a5b-efe2-4180-8110-cd8db3e65278","Type":"ContainerStarted","Data":"c7da8acbb1527ff4c40a062486446356c869e70e1245faa5c70f2fc914a79388"} Dec 07 16:20:03 crc kubenswrapper[4716]: I1207 16:20:03.960041 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7d76599897-4mhqt"] Dec 07 16:20:03 crc kubenswrapper[4716]: I1207 16:20:03.978267 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7d76599897-4mhqt"] Dec 07 16:20:03 crc kubenswrapper[4716]: I1207 16:20:03.988515 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-587c8bb479-4d8m7" podStartSLOduration=3.7798062310000002 podStartE2EDuration="5.988490982s" podCreationTimestamp="2025-12-07 16:19:58 +0000 UTC" firstStartedPulling="2025-12-07 16:20:00.023476876 +0000 UTC m=+1062.713761788" lastFinishedPulling="2025-12-07 16:20:02.232161627 +0000 UTC m=+1064.922446539" observedRunningTime="2025-12-07 16:20:03.970820634 +0000 UTC m=+1066.661105546" watchObservedRunningTime="2025-12-07 16:20:03.988490982 +0000 UTC m=+1066.678775894" Dec 07 16:20:04 crc kubenswrapper[4716]: I1207 16:20:04.033268 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=5.098736744 podStartE2EDuration="6.033253093s" podCreationTimestamp="2025-12-07 16:19:58 +0000 UTC" firstStartedPulling="2025-12-07 16:20:00.027592207 +0000 UTC m=+1062.717877119" lastFinishedPulling="2025-12-07 16:20:00.962108556 +0000 UTC m=+1063.652393468" observedRunningTime="2025-12-07 16:20:04.030490349 +0000 UTC m=+1066.720775261" watchObservedRunningTime="2025-12-07 16:20:04.033253093 +0000 UTC m=+1066.723538005" Dec 07 16:20:04 crc kubenswrapper[4716]: I1207 16:20:04.035390 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-75c8ccd7fb-bhzds" podStartSLOduration=3.66380886 podStartE2EDuration="6.035384581s" podCreationTimestamp="2025-12-07 16:19:58 +0000 UTC" firstStartedPulling="2025-12-07 16:19:59.852863907 +0000 UTC m=+1062.543148829" lastFinishedPulling="2025-12-07 16:20:02.224439638 +0000 UTC m=+1064.914724550" observedRunningTime="2025-12-07 16:20:04.001209996 +0000 UTC m=+1066.691494908" watchObservedRunningTime="2025-12-07 16:20:04.035384581 +0000 UTC m=+1066.725669493" Dec 07 16:20:04 crc kubenswrapper[4716]: I1207 16:20:04.081101 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 07 16:20:04 crc kubenswrapper[4716]: I1207 16:20:04.082501 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 07 16:20:04 crc kubenswrapper[4716]: I1207 16:20:04.096755 4716 scope.go:117] "RemoveContainer" containerID="008406d13a6ea7c73401539a525b2aa9a7d98be0c715633e8d8eac24e5c7ccf6" Dec 07 16:20:04 crc kubenswrapper[4716]: I1207 16:20:04.114245 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 07 16:20:04 crc kubenswrapper[4716]: I1207 16:20:04.134150 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7d8b488bdf-fvlzm"] Dec 07 16:20:04 crc kubenswrapper[4716]: I1207 16:20:04.150290 4716 scope.go:117] "RemoveContainer" containerID="c372e8f6fe59dccc9147f9a8695d09a6bc9fc695005f54aabb932a3697e49115" Dec 07 16:20:04 crc kubenswrapper[4716]: E1207 16:20:04.150812 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c372e8f6fe59dccc9147f9a8695d09a6bc9fc695005f54aabb932a3697e49115\": container with ID starting with c372e8f6fe59dccc9147f9a8695d09a6bc9fc695005f54aabb932a3697e49115 not found: ID does not exist" containerID="c372e8f6fe59dccc9147f9a8695d09a6bc9fc695005f54aabb932a3697e49115" Dec 07 16:20:04 crc kubenswrapper[4716]: I1207 16:20:04.150854 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c372e8f6fe59dccc9147f9a8695d09a6bc9fc695005f54aabb932a3697e49115"} err="failed to get container status \"c372e8f6fe59dccc9147f9a8695d09a6bc9fc695005f54aabb932a3697e49115\": rpc error: code = NotFound desc = could not find container \"c372e8f6fe59dccc9147f9a8695d09a6bc9fc695005f54aabb932a3697e49115\": container with ID starting with c372e8f6fe59dccc9147f9a8695d09a6bc9fc695005f54aabb932a3697e49115 not found: ID does not exist" Dec 07 16:20:04 crc kubenswrapper[4716]: I1207 16:20:04.150883 4716 scope.go:117] "RemoveContainer" containerID="008406d13a6ea7c73401539a525b2aa9a7d98be0c715633e8d8eac24e5c7ccf6" Dec 07 16:20:04 crc kubenswrapper[4716]: E1207 16:20:04.154509 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"008406d13a6ea7c73401539a525b2aa9a7d98be0c715633e8d8eac24e5c7ccf6\": container with ID starting with 008406d13a6ea7c73401539a525b2aa9a7d98be0c715633e8d8eac24e5c7ccf6 not found: ID does not exist" containerID="008406d13a6ea7c73401539a525b2aa9a7d98be0c715633e8d8eac24e5c7ccf6" Dec 07 16:20:04 crc kubenswrapper[4716]: I1207 16:20:04.154558 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"008406d13a6ea7c73401539a525b2aa9a7d98be0c715633e8d8eac24e5c7ccf6"} err="failed to get container status \"008406d13a6ea7c73401539a525b2aa9a7d98be0c715633e8d8eac24e5c7ccf6\": rpc error: code = NotFound desc = could not find container \"008406d13a6ea7c73401539a525b2aa9a7d98be0c715633e8d8eac24e5c7ccf6\": container with ID starting with 008406d13a6ea7c73401539a525b2aa9a7d98be0c715633e8d8eac24e5c7ccf6 not found: ID does not exist" Dec 07 16:20:04 crc kubenswrapper[4716]: I1207 16:20:04.154592 4716 scope.go:117] "RemoveContainer" containerID="c372e8f6fe59dccc9147f9a8695d09a6bc9fc695005f54aabb932a3697e49115" Dec 07 16:20:04 crc kubenswrapper[4716]: I1207 16:20:04.155046 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c372e8f6fe59dccc9147f9a8695d09a6bc9fc695005f54aabb932a3697e49115"} err="failed to get container status \"c372e8f6fe59dccc9147f9a8695d09a6bc9fc695005f54aabb932a3697e49115\": rpc error: code = NotFound desc = could not find container \"c372e8f6fe59dccc9147f9a8695d09a6bc9fc695005f54aabb932a3697e49115\": container with ID starting with c372e8f6fe59dccc9147f9a8695d09a6bc9fc695005f54aabb932a3697e49115 not found: ID does not exist" Dec 07 16:20:04 crc kubenswrapper[4716]: I1207 16:20:04.155105 4716 scope.go:117] "RemoveContainer" containerID="008406d13a6ea7c73401539a525b2aa9a7d98be0c715633e8d8eac24e5c7ccf6" Dec 07 16:20:04 crc kubenswrapper[4716]: I1207 16:20:04.155488 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"008406d13a6ea7c73401539a525b2aa9a7d98be0c715633e8d8eac24e5c7ccf6"} err="failed to get container status \"008406d13a6ea7c73401539a525b2aa9a7d98be0c715633e8d8eac24e5c7ccf6\": rpc error: code = NotFound desc = could not find container \"008406d13a6ea7c73401539a525b2aa9a7d98be0c715633e8d8eac24e5c7ccf6\": container with ID starting with 008406d13a6ea7c73401539a525b2aa9a7d98be0c715633e8d8eac24e5c7ccf6 not found: ID does not exist" Dec 07 16:20:04 crc kubenswrapper[4716]: I1207 16:20:04.155526 4716 scope.go:117] "RemoveContainer" containerID="e49f645f546db04a604348b06b91c87e2a6645528ec78ab475e94e276e8c3f6b" Dec 07 16:20:04 crc kubenswrapper[4716]: I1207 16:20:04.159122 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7d8b488bdf-fvlzm"] Dec 07 16:20:04 crc kubenswrapper[4716]: I1207 16:20:04.170155 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 07 16:20:04 crc kubenswrapper[4716]: E1207 16:20:04.170561 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ee00bf0-1412-4fca-a570-6600db549923" containerName="ceilometer-central-agent" Dec 07 16:20:04 crc kubenswrapper[4716]: I1207 16:20:04.170578 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ee00bf0-1412-4fca-a570-6600db549923" containerName="ceilometer-central-agent" Dec 07 16:20:04 crc kubenswrapper[4716]: E1207 16:20:04.170595 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c76ced8-9170-4ccc-a497-87388e8546dd" containerName="horizon-log" Dec 07 16:20:04 crc kubenswrapper[4716]: I1207 16:20:04.170601 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c76ced8-9170-4ccc-a497-87388e8546dd" containerName="horizon-log" Dec 07 16:20:04 crc kubenswrapper[4716]: E1207 16:20:04.170608 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ee00bf0-1412-4fca-a570-6600db549923" containerName="sg-core" Dec 07 16:20:04 crc kubenswrapper[4716]: I1207 16:20:04.170615 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ee00bf0-1412-4fca-a570-6600db549923" containerName="sg-core" Dec 07 16:20:04 crc kubenswrapper[4716]: E1207 16:20:04.170629 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ee00bf0-1412-4fca-a570-6600db549923" containerName="ceilometer-notification-agent" Dec 07 16:20:04 crc kubenswrapper[4716]: I1207 16:20:04.170636 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ee00bf0-1412-4fca-a570-6600db549923" containerName="ceilometer-notification-agent" Dec 07 16:20:04 crc kubenswrapper[4716]: E1207 16:20:04.170650 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ee00bf0-1412-4fca-a570-6600db549923" containerName="proxy-httpd" Dec 07 16:20:04 crc kubenswrapper[4716]: I1207 16:20:04.170655 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ee00bf0-1412-4fca-a570-6600db549923" containerName="proxy-httpd" Dec 07 16:20:04 crc kubenswrapper[4716]: E1207 16:20:04.170669 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7db6aff-6831-49ea-9d51-24ec9ac40a88" containerName="horizon-log" Dec 07 16:20:04 crc kubenswrapper[4716]: I1207 16:20:04.170674 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7db6aff-6831-49ea-9d51-24ec9ac40a88" containerName="horizon-log" Dec 07 16:20:04 crc kubenswrapper[4716]: E1207 16:20:04.170687 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c76ced8-9170-4ccc-a497-87388e8546dd" containerName="horizon" Dec 07 16:20:04 crc kubenswrapper[4716]: I1207 16:20:04.170694 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c76ced8-9170-4ccc-a497-87388e8546dd" containerName="horizon" Dec 07 16:20:04 crc kubenswrapper[4716]: E1207 16:20:04.170708 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7db6aff-6831-49ea-9d51-24ec9ac40a88" containerName="horizon" Dec 07 16:20:04 crc kubenswrapper[4716]: I1207 16:20:04.170713 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7db6aff-6831-49ea-9d51-24ec9ac40a88" containerName="horizon" Dec 07 16:20:04 crc kubenswrapper[4716]: I1207 16:20:04.170892 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ee00bf0-1412-4fca-a570-6600db549923" containerName="proxy-httpd" Dec 07 16:20:04 crc kubenswrapper[4716]: I1207 16:20:04.170905 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ee00bf0-1412-4fca-a570-6600db549923" containerName="ceilometer-notification-agent" Dec 07 16:20:04 crc kubenswrapper[4716]: I1207 16:20:04.170911 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7db6aff-6831-49ea-9d51-24ec9ac40a88" containerName="horizon" Dec 07 16:20:04 crc kubenswrapper[4716]: I1207 16:20:04.170921 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ee00bf0-1412-4fca-a570-6600db549923" containerName="ceilometer-central-agent" Dec 07 16:20:04 crc kubenswrapper[4716]: I1207 16:20:04.170936 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7db6aff-6831-49ea-9d51-24ec9ac40a88" containerName="horizon-log" Dec 07 16:20:04 crc kubenswrapper[4716]: I1207 16:20:04.170946 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c76ced8-9170-4ccc-a497-87388e8546dd" containerName="horizon" Dec 07 16:20:04 crc kubenswrapper[4716]: I1207 16:20:04.170956 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c76ced8-9170-4ccc-a497-87388e8546dd" containerName="horizon-log" Dec 07 16:20:04 crc kubenswrapper[4716]: I1207 16:20:04.170963 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ee00bf0-1412-4fca-a570-6600db549923" containerName="sg-core" Dec 07 16:20:04 crc kubenswrapper[4716]: I1207 16:20:04.172653 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 07 16:20:04 crc kubenswrapper[4716]: I1207 16:20:04.175580 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 07 16:20:04 crc kubenswrapper[4716]: I1207 16:20:04.176622 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 07 16:20:04 crc kubenswrapper[4716]: I1207 16:20:04.179843 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 07 16:20:04 crc kubenswrapper[4716]: I1207 16:20:04.210925 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0480133a-7588-4a60-921e-e53de8998ca9-config-data\") pod \"ceilometer-0\" (UID: \"0480133a-7588-4a60-921e-e53de8998ca9\") " pod="openstack/ceilometer-0" Dec 07 16:20:04 crc kubenswrapper[4716]: I1207 16:20:04.211106 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltfdw\" (UniqueName: \"kubernetes.io/projected/0480133a-7588-4a60-921e-e53de8998ca9-kube-api-access-ltfdw\") pod \"ceilometer-0\" (UID: \"0480133a-7588-4a60-921e-e53de8998ca9\") " pod="openstack/ceilometer-0" Dec 07 16:20:04 crc kubenswrapper[4716]: I1207 16:20:04.211164 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0480133a-7588-4a60-921e-e53de8998ca9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0480133a-7588-4a60-921e-e53de8998ca9\") " pod="openstack/ceilometer-0" Dec 07 16:20:04 crc kubenswrapper[4716]: I1207 16:20:04.211191 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0480133a-7588-4a60-921e-e53de8998ca9-scripts\") pod \"ceilometer-0\" (UID: \"0480133a-7588-4a60-921e-e53de8998ca9\") " pod="openstack/ceilometer-0" Dec 07 16:20:04 crc kubenswrapper[4716]: I1207 16:20:04.211209 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0480133a-7588-4a60-921e-e53de8998ca9-run-httpd\") pod \"ceilometer-0\" (UID: \"0480133a-7588-4a60-921e-e53de8998ca9\") " pod="openstack/ceilometer-0" Dec 07 16:20:04 crc kubenswrapper[4716]: I1207 16:20:04.211269 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0480133a-7588-4a60-921e-e53de8998ca9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0480133a-7588-4a60-921e-e53de8998ca9\") " pod="openstack/ceilometer-0" Dec 07 16:20:04 crc kubenswrapper[4716]: I1207 16:20:04.211397 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0480133a-7588-4a60-921e-e53de8998ca9-log-httpd\") pod \"ceilometer-0\" (UID: \"0480133a-7588-4a60-921e-e53de8998ca9\") " pod="openstack/ceilometer-0" Dec 07 16:20:04 crc kubenswrapper[4716]: I1207 16:20:04.313328 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltfdw\" (UniqueName: \"kubernetes.io/projected/0480133a-7588-4a60-921e-e53de8998ca9-kube-api-access-ltfdw\") pod \"ceilometer-0\" (UID: \"0480133a-7588-4a60-921e-e53de8998ca9\") " pod="openstack/ceilometer-0" Dec 07 16:20:04 crc kubenswrapper[4716]: I1207 16:20:04.313382 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0480133a-7588-4a60-921e-e53de8998ca9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0480133a-7588-4a60-921e-e53de8998ca9\") " pod="openstack/ceilometer-0" Dec 07 16:20:04 crc kubenswrapper[4716]: I1207 16:20:04.313408 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0480133a-7588-4a60-921e-e53de8998ca9-scripts\") pod \"ceilometer-0\" (UID: \"0480133a-7588-4a60-921e-e53de8998ca9\") " pod="openstack/ceilometer-0" Dec 07 16:20:04 crc kubenswrapper[4716]: I1207 16:20:04.313426 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0480133a-7588-4a60-921e-e53de8998ca9-run-httpd\") pod \"ceilometer-0\" (UID: \"0480133a-7588-4a60-921e-e53de8998ca9\") " pod="openstack/ceilometer-0" Dec 07 16:20:04 crc kubenswrapper[4716]: I1207 16:20:04.313452 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0480133a-7588-4a60-921e-e53de8998ca9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0480133a-7588-4a60-921e-e53de8998ca9\") " pod="openstack/ceilometer-0" Dec 07 16:20:04 crc kubenswrapper[4716]: I1207 16:20:04.313503 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0480133a-7588-4a60-921e-e53de8998ca9-log-httpd\") pod \"ceilometer-0\" (UID: \"0480133a-7588-4a60-921e-e53de8998ca9\") " pod="openstack/ceilometer-0" Dec 07 16:20:04 crc kubenswrapper[4716]: I1207 16:20:04.313553 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0480133a-7588-4a60-921e-e53de8998ca9-config-data\") pod \"ceilometer-0\" (UID: \"0480133a-7588-4a60-921e-e53de8998ca9\") " pod="openstack/ceilometer-0" Dec 07 16:20:04 crc kubenswrapper[4716]: I1207 16:20:04.314366 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0480133a-7588-4a60-921e-e53de8998ca9-run-httpd\") pod \"ceilometer-0\" (UID: \"0480133a-7588-4a60-921e-e53de8998ca9\") " pod="openstack/ceilometer-0" Dec 07 16:20:04 crc kubenswrapper[4716]: I1207 16:20:04.315561 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0480133a-7588-4a60-921e-e53de8998ca9-log-httpd\") pod \"ceilometer-0\" (UID: \"0480133a-7588-4a60-921e-e53de8998ca9\") " pod="openstack/ceilometer-0" Dec 07 16:20:04 crc kubenswrapper[4716]: I1207 16:20:04.325957 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0480133a-7588-4a60-921e-e53de8998ca9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0480133a-7588-4a60-921e-e53de8998ca9\") " pod="openstack/ceilometer-0" Dec 07 16:20:04 crc kubenswrapper[4716]: I1207 16:20:04.326172 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0480133a-7588-4a60-921e-e53de8998ca9-config-data\") pod \"ceilometer-0\" (UID: \"0480133a-7588-4a60-921e-e53de8998ca9\") " pod="openstack/ceilometer-0" Dec 07 16:20:04 crc kubenswrapper[4716]: I1207 16:20:04.328616 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0480133a-7588-4a60-921e-e53de8998ca9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0480133a-7588-4a60-921e-e53de8998ca9\") " pod="openstack/ceilometer-0" Dec 07 16:20:04 crc kubenswrapper[4716]: I1207 16:20:04.329968 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0480133a-7588-4a60-921e-e53de8998ca9-scripts\") pod \"ceilometer-0\" (UID: \"0480133a-7588-4a60-921e-e53de8998ca9\") " pod="openstack/ceilometer-0" Dec 07 16:20:04 crc kubenswrapper[4716]: I1207 16:20:04.334863 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltfdw\" (UniqueName: \"kubernetes.io/projected/0480133a-7588-4a60-921e-e53de8998ca9-kube-api-access-ltfdw\") pod \"ceilometer-0\" (UID: \"0480133a-7588-4a60-921e-e53de8998ca9\") " pod="openstack/ceilometer-0" Dec 07 16:20:04 crc kubenswrapper[4716]: I1207 16:20:04.339169 4716 scope.go:117] "RemoveContainer" containerID="d9c99d90e0f76afdd5481a03c49c1a155a22721530d081f1e90302e0f9d9fc9d" Dec 07 16:20:04 crc kubenswrapper[4716]: I1207 16:20:04.430870 4716 scope.go:117] "RemoveContainer" containerID="e49f645f546db04a604348b06b91c87e2a6645528ec78ab475e94e276e8c3f6b" Dec 07 16:20:04 crc kubenswrapper[4716]: E1207 16:20:04.431429 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e49f645f546db04a604348b06b91c87e2a6645528ec78ab475e94e276e8c3f6b\": container with ID starting with e49f645f546db04a604348b06b91c87e2a6645528ec78ab475e94e276e8c3f6b not found: ID does not exist" containerID="e49f645f546db04a604348b06b91c87e2a6645528ec78ab475e94e276e8c3f6b" Dec 07 16:20:04 crc kubenswrapper[4716]: I1207 16:20:04.431483 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e49f645f546db04a604348b06b91c87e2a6645528ec78ab475e94e276e8c3f6b"} err="failed to get container status \"e49f645f546db04a604348b06b91c87e2a6645528ec78ab475e94e276e8c3f6b\": rpc error: code = NotFound desc = could not find container \"e49f645f546db04a604348b06b91c87e2a6645528ec78ab475e94e276e8c3f6b\": container with ID starting with e49f645f546db04a604348b06b91c87e2a6645528ec78ab475e94e276e8c3f6b not found: ID does not exist" Dec 07 16:20:04 crc kubenswrapper[4716]: I1207 16:20:04.431514 4716 scope.go:117] "RemoveContainer" containerID="d9c99d90e0f76afdd5481a03c49c1a155a22721530d081f1e90302e0f9d9fc9d" Dec 07 16:20:04 crc kubenswrapper[4716]: E1207 16:20:04.431941 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9c99d90e0f76afdd5481a03c49c1a155a22721530d081f1e90302e0f9d9fc9d\": container with ID starting with d9c99d90e0f76afdd5481a03c49c1a155a22721530d081f1e90302e0f9d9fc9d not found: ID does not exist" containerID="d9c99d90e0f76afdd5481a03c49c1a155a22721530d081f1e90302e0f9d9fc9d" Dec 07 16:20:04 crc kubenswrapper[4716]: I1207 16:20:04.431990 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9c99d90e0f76afdd5481a03c49c1a155a22721530d081f1e90302e0f9d9fc9d"} err="failed to get container status \"d9c99d90e0f76afdd5481a03c49c1a155a22721530d081f1e90302e0f9d9fc9d\": rpc error: code = NotFound desc = could not find container \"d9c99d90e0f76afdd5481a03c49c1a155a22721530d081f1e90302e0f9d9fc9d\": container with ID starting with d9c99d90e0f76afdd5481a03c49c1a155a22721530d081f1e90302e0f9d9fc9d not found: ID does not exist" Dec 07 16:20:04 crc kubenswrapper[4716]: I1207 16:20:04.432022 4716 scope.go:117] "RemoveContainer" containerID="e49f645f546db04a604348b06b91c87e2a6645528ec78ab475e94e276e8c3f6b" Dec 07 16:20:04 crc kubenswrapper[4716]: I1207 16:20:04.432293 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e49f645f546db04a604348b06b91c87e2a6645528ec78ab475e94e276e8c3f6b"} err="failed to get container status \"e49f645f546db04a604348b06b91c87e2a6645528ec78ab475e94e276e8c3f6b\": rpc error: code = NotFound desc = could not find container \"e49f645f546db04a604348b06b91c87e2a6645528ec78ab475e94e276e8c3f6b\": container with ID starting with e49f645f546db04a604348b06b91c87e2a6645528ec78ab475e94e276e8c3f6b not found: ID does not exist" Dec 07 16:20:04 crc kubenswrapper[4716]: I1207 16:20:04.432317 4716 scope.go:117] "RemoveContainer" containerID="d9c99d90e0f76afdd5481a03c49c1a155a22721530d081f1e90302e0f9d9fc9d" Dec 07 16:20:04 crc kubenswrapper[4716]: I1207 16:20:04.432518 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9c99d90e0f76afdd5481a03c49c1a155a22721530d081f1e90302e0f9d9fc9d"} err="failed to get container status \"d9c99d90e0f76afdd5481a03c49c1a155a22721530d081f1e90302e0f9d9fc9d\": rpc error: code = NotFound desc = could not find container \"d9c99d90e0f76afdd5481a03c49c1a155a22721530d081f1e90302e0f9d9fc9d\": container with ID starting with d9c99d90e0f76afdd5481a03c49c1a155a22721530d081f1e90302e0f9d9fc9d not found: ID does not exist" Dec 07 16:20:04 crc kubenswrapper[4716]: I1207 16:20:04.487127 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 07 16:20:04 crc kubenswrapper[4716]: I1207 16:20:04.495796 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 07 16:20:04 crc kubenswrapper[4716]: I1207 16:20:04.894678 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-bd7b79585-wwp6b" Dec 07 16:20:04 crc kubenswrapper[4716]: I1207 16:20:04.976145 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-69d87fd46d-9k5rq"] Dec 07 16:20:04 crc kubenswrapper[4716]: I1207 16:20:04.976457 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-69d87fd46d-9k5rq" podUID="8b125bd6-f08f-442c-96a8-b903160be13a" containerName="neutron-api" containerID="cri-o://c00c5a8c50e26dad3fc07ff6ed5d0663b13bd6897c9a5e4c05fae5699523e678" gracePeriod=30 Dec 07 16:20:04 crc kubenswrapper[4716]: I1207 16:20:04.976634 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-69d87fd46d-9k5rq" podUID="8b125bd6-f08f-442c-96a8-b903160be13a" containerName="neutron-httpd" containerID="cri-o://9e0f9b2ae46383d32e78b438d5bf9227170de493caed859e596aafea34071033" gracePeriod=30 Dec 07 16:20:04 crc kubenswrapper[4716]: I1207 16:20:04.988209 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 07 16:20:05 crc kubenswrapper[4716]: I1207 16:20:05.034811 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-86c7567d4-99rx9" Dec 07 16:20:05 crc kubenswrapper[4716]: I1207 16:20:05.259923 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-64c898bd6-wgw8m"] Dec 07 16:20:05 crc kubenswrapper[4716]: I1207 16:20:05.262199 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-64c898bd6-wgw8m" Dec 07 16:20:05 crc kubenswrapper[4716]: I1207 16:20:05.269700 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Dec 07 16:20:05 crc kubenswrapper[4716]: I1207 16:20:05.271726 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Dec 07 16:20:05 crc kubenswrapper[4716]: I1207 16:20:05.278288 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-64c898bd6-wgw8m"] Dec 07 16:20:05 crc kubenswrapper[4716]: I1207 16:20:05.355950 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/19efaa0a-3cda-4ef8-a845-bbcc78c747ec-logs\") pod \"barbican-api-64c898bd6-wgw8m\" (UID: \"19efaa0a-3cda-4ef8-a845-bbcc78c747ec\") " pod="openstack/barbican-api-64c898bd6-wgw8m" Dec 07 16:20:05 crc kubenswrapper[4716]: I1207 16:20:05.356327 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pdsd5\" (UniqueName: \"kubernetes.io/projected/19efaa0a-3cda-4ef8-a845-bbcc78c747ec-kube-api-access-pdsd5\") pod \"barbican-api-64c898bd6-wgw8m\" (UID: \"19efaa0a-3cda-4ef8-a845-bbcc78c747ec\") " pod="openstack/barbican-api-64c898bd6-wgw8m" Dec 07 16:20:05 crc kubenswrapper[4716]: I1207 16:20:05.356379 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19efaa0a-3cda-4ef8-a845-bbcc78c747ec-config-data\") pod \"barbican-api-64c898bd6-wgw8m\" (UID: \"19efaa0a-3cda-4ef8-a845-bbcc78c747ec\") " pod="openstack/barbican-api-64c898bd6-wgw8m" Dec 07 16:20:05 crc kubenswrapper[4716]: I1207 16:20:05.356414 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/19efaa0a-3cda-4ef8-a845-bbcc78c747ec-config-data-custom\") pod \"barbican-api-64c898bd6-wgw8m\" (UID: \"19efaa0a-3cda-4ef8-a845-bbcc78c747ec\") " pod="openstack/barbican-api-64c898bd6-wgw8m" Dec 07 16:20:05 crc kubenswrapper[4716]: I1207 16:20:05.356441 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/19efaa0a-3cda-4ef8-a845-bbcc78c747ec-public-tls-certs\") pod \"barbican-api-64c898bd6-wgw8m\" (UID: \"19efaa0a-3cda-4ef8-a845-bbcc78c747ec\") " pod="openstack/barbican-api-64c898bd6-wgw8m" Dec 07 16:20:05 crc kubenswrapper[4716]: I1207 16:20:05.356471 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19efaa0a-3cda-4ef8-a845-bbcc78c747ec-combined-ca-bundle\") pod \"barbican-api-64c898bd6-wgw8m\" (UID: \"19efaa0a-3cda-4ef8-a845-bbcc78c747ec\") " pod="openstack/barbican-api-64c898bd6-wgw8m" Dec 07 16:20:05 crc kubenswrapper[4716]: I1207 16:20:05.356487 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/19efaa0a-3cda-4ef8-a845-bbcc78c747ec-internal-tls-certs\") pod \"barbican-api-64c898bd6-wgw8m\" (UID: \"19efaa0a-3cda-4ef8-a845-bbcc78c747ec\") " pod="openstack/barbican-api-64c898bd6-wgw8m" Dec 07 16:20:05 crc kubenswrapper[4716]: I1207 16:20:05.460707 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19efaa0a-3cda-4ef8-a845-bbcc78c747ec-config-data\") pod \"barbican-api-64c898bd6-wgw8m\" (UID: \"19efaa0a-3cda-4ef8-a845-bbcc78c747ec\") " pod="openstack/barbican-api-64c898bd6-wgw8m" Dec 07 16:20:05 crc kubenswrapper[4716]: I1207 16:20:05.460812 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/19efaa0a-3cda-4ef8-a845-bbcc78c747ec-config-data-custom\") pod \"barbican-api-64c898bd6-wgw8m\" (UID: \"19efaa0a-3cda-4ef8-a845-bbcc78c747ec\") " pod="openstack/barbican-api-64c898bd6-wgw8m" Dec 07 16:20:05 crc kubenswrapper[4716]: I1207 16:20:05.460855 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/19efaa0a-3cda-4ef8-a845-bbcc78c747ec-public-tls-certs\") pod \"barbican-api-64c898bd6-wgw8m\" (UID: \"19efaa0a-3cda-4ef8-a845-bbcc78c747ec\") " pod="openstack/barbican-api-64c898bd6-wgw8m" Dec 07 16:20:05 crc kubenswrapper[4716]: I1207 16:20:05.460920 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19efaa0a-3cda-4ef8-a845-bbcc78c747ec-combined-ca-bundle\") pod \"barbican-api-64c898bd6-wgw8m\" (UID: \"19efaa0a-3cda-4ef8-a845-bbcc78c747ec\") " pod="openstack/barbican-api-64c898bd6-wgw8m" Dec 07 16:20:05 crc kubenswrapper[4716]: I1207 16:20:05.460945 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/19efaa0a-3cda-4ef8-a845-bbcc78c747ec-internal-tls-certs\") pod \"barbican-api-64c898bd6-wgw8m\" (UID: \"19efaa0a-3cda-4ef8-a845-bbcc78c747ec\") " pod="openstack/barbican-api-64c898bd6-wgw8m" Dec 07 16:20:05 crc kubenswrapper[4716]: I1207 16:20:05.461128 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/19efaa0a-3cda-4ef8-a845-bbcc78c747ec-logs\") pod \"barbican-api-64c898bd6-wgw8m\" (UID: \"19efaa0a-3cda-4ef8-a845-bbcc78c747ec\") " pod="openstack/barbican-api-64c898bd6-wgw8m" Dec 07 16:20:05 crc kubenswrapper[4716]: I1207 16:20:05.461178 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pdsd5\" (UniqueName: \"kubernetes.io/projected/19efaa0a-3cda-4ef8-a845-bbcc78c747ec-kube-api-access-pdsd5\") pod \"barbican-api-64c898bd6-wgw8m\" (UID: \"19efaa0a-3cda-4ef8-a845-bbcc78c747ec\") " pod="openstack/barbican-api-64c898bd6-wgw8m" Dec 07 16:20:05 crc kubenswrapper[4716]: I1207 16:20:05.464127 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/19efaa0a-3cda-4ef8-a845-bbcc78c747ec-logs\") pod \"barbican-api-64c898bd6-wgw8m\" (UID: \"19efaa0a-3cda-4ef8-a845-bbcc78c747ec\") " pod="openstack/barbican-api-64c898bd6-wgw8m" Dec 07 16:20:05 crc kubenswrapper[4716]: I1207 16:20:05.467534 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19efaa0a-3cda-4ef8-a845-bbcc78c747ec-config-data\") pod \"barbican-api-64c898bd6-wgw8m\" (UID: \"19efaa0a-3cda-4ef8-a845-bbcc78c747ec\") " pod="openstack/barbican-api-64c898bd6-wgw8m" Dec 07 16:20:05 crc kubenswrapper[4716]: I1207 16:20:05.467688 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/19efaa0a-3cda-4ef8-a845-bbcc78c747ec-internal-tls-certs\") pod \"barbican-api-64c898bd6-wgw8m\" (UID: \"19efaa0a-3cda-4ef8-a845-bbcc78c747ec\") " pod="openstack/barbican-api-64c898bd6-wgw8m" Dec 07 16:20:05 crc kubenswrapper[4716]: I1207 16:20:05.468732 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/19efaa0a-3cda-4ef8-a845-bbcc78c747ec-public-tls-certs\") pod \"barbican-api-64c898bd6-wgw8m\" (UID: \"19efaa0a-3cda-4ef8-a845-bbcc78c747ec\") " pod="openstack/barbican-api-64c898bd6-wgw8m" Dec 07 16:20:05 crc kubenswrapper[4716]: I1207 16:20:05.469565 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19efaa0a-3cda-4ef8-a845-bbcc78c747ec-combined-ca-bundle\") pod \"barbican-api-64c898bd6-wgw8m\" (UID: \"19efaa0a-3cda-4ef8-a845-bbcc78c747ec\") " pod="openstack/barbican-api-64c898bd6-wgw8m" Dec 07 16:20:05 crc kubenswrapper[4716]: I1207 16:20:05.470048 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/19efaa0a-3cda-4ef8-a845-bbcc78c747ec-config-data-custom\") pod \"barbican-api-64c898bd6-wgw8m\" (UID: \"19efaa0a-3cda-4ef8-a845-bbcc78c747ec\") " pod="openstack/barbican-api-64c898bd6-wgw8m" Dec 07 16:20:05 crc kubenswrapper[4716]: I1207 16:20:05.502774 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pdsd5\" (UniqueName: \"kubernetes.io/projected/19efaa0a-3cda-4ef8-a845-bbcc78c747ec-kube-api-access-pdsd5\") pod \"barbican-api-64c898bd6-wgw8m\" (UID: \"19efaa0a-3cda-4ef8-a845-bbcc78c747ec\") " pod="openstack/barbican-api-64c898bd6-wgw8m" Dec 07 16:20:05 crc kubenswrapper[4716]: I1207 16:20:05.588166 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-64c898bd6-wgw8m" Dec 07 16:20:05 crc kubenswrapper[4716]: I1207 16:20:05.670625 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ee00bf0-1412-4fca-a570-6600db549923" path="/var/lib/kubelet/pods/2ee00bf0-1412-4fca-a570-6600db549923/volumes" Dec 07 16:20:05 crc kubenswrapper[4716]: I1207 16:20:05.671460 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c76ced8-9170-4ccc-a497-87388e8546dd" path="/var/lib/kubelet/pods/5c76ced8-9170-4ccc-a497-87388e8546dd/volumes" Dec 07 16:20:05 crc kubenswrapper[4716]: I1207 16:20:05.672536 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7db6aff-6831-49ea-9d51-24ec9ac40a88" path="/var/lib/kubelet/pods/d7db6aff-6831-49ea-9d51-24ec9ac40a88/volumes" Dec 07 16:20:05 crc kubenswrapper[4716]: I1207 16:20:05.923726 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-7f9b78866d-9g9kv" Dec 07 16:20:05 crc kubenswrapper[4716]: I1207 16:20:05.982644 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-86c7567d4-99rx9"] Dec 07 16:20:05 crc kubenswrapper[4716]: I1207 16:20:05.987214 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0480133a-7588-4a60-921e-e53de8998ca9","Type":"ContainerStarted","Data":"610237798bb1fa0726f011172548d74003aeae66c492ec9ddc28726b612574eb"} Dec 07 16:20:05 crc kubenswrapper[4716]: I1207 16:20:05.987262 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0480133a-7588-4a60-921e-e53de8998ca9","Type":"ContainerStarted","Data":"e042f3a14b4b726e2937b7c1edb5cf38b368164ff2e4f333aa761c8edd18ee9b"} Dec 07 16:20:05 crc kubenswrapper[4716]: I1207 16:20:05.988585 4716 generic.go:334] "Generic (PLEG): container finished" podID="8b125bd6-f08f-442c-96a8-b903160be13a" containerID="9e0f9b2ae46383d32e78b438d5bf9227170de493caed859e596aafea34071033" exitCode=0 Dec 07 16:20:05 crc kubenswrapper[4716]: I1207 16:20:05.988885 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-86c7567d4-99rx9" podUID="c291e114-7940-46bc-91d3-a8fa256549ff" containerName="horizon-log" containerID="cri-o://00893cd59537178a447a7d7ca931b0f9922afd0ddd98cb384ea1435f3b09a2f9" gracePeriod=30 Dec 07 16:20:05 crc kubenswrapper[4716]: I1207 16:20:05.989271 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-69d87fd46d-9k5rq" event={"ID":"8b125bd6-f08f-442c-96a8-b903160be13a","Type":"ContainerDied","Data":"9e0f9b2ae46383d32e78b438d5bf9227170de493caed859e596aafea34071033"} Dec 07 16:20:05 crc kubenswrapper[4716]: I1207 16:20:05.990505 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-86c7567d4-99rx9" podUID="c291e114-7940-46bc-91d3-a8fa256549ff" containerName="horizon" containerID="cri-o://bc89298b4c5f29666ca1176ee3a6613d8d2303c5623d2bec2695ed37ae7b6619" gracePeriod=30 Dec 07 16:20:06 crc kubenswrapper[4716]: I1207 16:20:06.093686 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-64c898bd6-wgw8m"] Dec 07 16:20:07 crc kubenswrapper[4716]: I1207 16:20:07.013522 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0480133a-7588-4a60-921e-e53de8998ca9","Type":"ContainerStarted","Data":"9307786cb3b64f126e8b2107ac2b2c179b596785613c0be95dfc95f8b2b5f691"} Dec 07 16:20:07 crc kubenswrapper[4716]: I1207 16:20:07.017317 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-64c898bd6-wgw8m" event={"ID":"19efaa0a-3cda-4ef8-a845-bbcc78c747ec","Type":"ContainerStarted","Data":"51eea4bf53356dc1cb0771ab0e2b91be4fbd55193cbdb56c619b75374f8acd89"} Dec 07 16:20:07 crc kubenswrapper[4716]: I1207 16:20:07.017367 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-64c898bd6-wgw8m" event={"ID":"19efaa0a-3cda-4ef8-a845-bbcc78c747ec","Type":"ContainerStarted","Data":"4cb3ab9a236f3cfe50a99116636506caf341aaa86efac83bbf882acc32752f7a"} Dec 07 16:20:07 crc kubenswrapper[4716]: I1207 16:20:07.017380 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-64c898bd6-wgw8m" event={"ID":"19efaa0a-3cda-4ef8-a845-bbcc78c747ec","Type":"ContainerStarted","Data":"b8af6176c7e4932cf5230b789e4b33a9c4e74f92e6323238d656bdbbe05a4ed5"} Dec 07 16:20:07 crc kubenswrapper[4716]: I1207 16:20:07.017623 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-64c898bd6-wgw8m" Dec 07 16:20:07 crc kubenswrapper[4716]: I1207 16:20:07.017790 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-64c898bd6-wgw8m" Dec 07 16:20:07 crc kubenswrapper[4716]: I1207 16:20:07.044108 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-64c898bd6-wgw8m" podStartSLOduration=2.044085869 podStartE2EDuration="2.044085869s" podCreationTimestamp="2025-12-07 16:20:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:20:07.039208437 +0000 UTC m=+1069.729493349" watchObservedRunningTime="2025-12-07 16:20:07.044085869 +0000 UTC m=+1069.734370781" Dec 07 16:20:07 crc kubenswrapper[4716]: I1207 16:20:07.179692 4716 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-65b7f6996b-q6xwd" podUID="65aaa058-a1be-4477-b808-d0ea1d8e15fe" containerName="barbican-api" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 07 16:20:08 crc kubenswrapper[4716]: I1207 16:20:08.038467 4716 generic.go:334] "Generic (PLEG): container finished" podID="8b125bd6-f08f-442c-96a8-b903160be13a" containerID="c00c5a8c50e26dad3fc07ff6ed5d0663b13bd6897c9a5e4c05fae5699523e678" exitCode=0 Dec 07 16:20:08 crc kubenswrapper[4716]: I1207 16:20:08.038551 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-69d87fd46d-9k5rq" event={"ID":"8b125bd6-f08f-442c-96a8-b903160be13a","Type":"ContainerDied","Data":"c00c5a8c50e26dad3fc07ff6ed5d0663b13bd6897c9a5e4c05fae5699523e678"} Dec 07 16:20:08 crc kubenswrapper[4716]: I1207 16:20:08.057693 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0480133a-7588-4a60-921e-e53de8998ca9","Type":"ContainerStarted","Data":"6ccfca76fd46f01f21a671fb4b52ac11b7e1cfd9deceaa24b6f22fff4d8d1881"} Dec 07 16:20:08 crc kubenswrapper[4716]: I1207 16:20:08.439992 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-69d87fd46d-9k5rq" Dec 07 16:20:08 crc kubenswrapper[4716]: I1207 16:20:08.539581 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b125bd6-f08f-442c-96a8-b903160be13a-combined-ca-bundle\") pod \"8b125bd6-f08f-442c-96a8-b903160be13a\" (UID: \"8b125bd6-f08f-442c-96a8-b903160be13a\") " Dec 07 16:20:08 crc kubenswrapper[4716]: I1207 16:20:08.539699 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jxj87\" (UniqueName: \"kubernetes.io/projected/8b125bd6-f08f-442c-96a8-b903160be13a-kube-api-access-jxj87\") pod \"8b125bd6-f08f-442c-96a8-b903160be13a\" (UID: \"8b125bd6-f08f-442c-96a8-b903160be13a\") " Dec 07 16:20:08 crc kubenswrapper[4716]: I1207 16:20:08.539782 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/8b125bd6-f08f-442c-96a8-b903160be13a-httpd-config\") pod \"8b125bd6-f08f-442c-96a8-b903160be13a\" (UID: \"8b125bd6-f08f-442c-96a8-b903160be13a\") " Dec 07 16:20:08 crc kubenswrapper[4716]: I1207 16:20:08.539856 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8b125bd6-f08f-442c-96a8-b903160be13a-config\") pod \"8b125bd6-f08f-442c-96a8-b903160be13a\" (UID: \"8b125bd6-f08f-442c-96a8-b903160be13a\") " Dec 07 16:20:08 crc kubenswrapper[4716]: I1207 16:20:08.539952 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b125bd6-f08f-442c-96a8-b903160be13a-ovndb-tls-certs\") pod \"8b125bd6-f08f-442c-96a8-b903160be13a\" (UID: \"8b125bd6-f08f-442c-96a8-b903160be13a\") " Dec 07 16:20:08 crc kubenswrapper[4716]: I1207 16:20:08.549335 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b125bd6-f08f-442c-96a8-b903160be13a-kube-api-access-jxj87" (OuterVolumeSpecName: "kube-api-access-jxj87") pod "8b125bd6-f08f-442c-96a8-b903160be13a" (UID: "8b125bd6-f08f-442c-96a8-b903160be13a"). InnerVolumeSpecName "kube-api-access-jxj87". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:20:08 crc kubenswrapper[4716]: I1207 16:20:08.551344 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b125bd6-f08f-442c-96a8-b903160be13a-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "8b125bd6-f08f-442c-96a8-b903160be13a" (UID: "8b125bd6-f08f-442c-96a8-b903160be13a"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:20:08 crc kubenswrapper[4716]: I1207 16:20:08.590789 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b125bd6-f08f-442c-96a8-b903160be13a-config" (OuterVolumeSpecName: "config") pod "8b125bd6-f08f-442c-96a8-b903160be13a" (UID: "8b125bd6-f08f-442c-96a8-b903160be13a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:20:08 crc kubenswrapper[4716]: I1207 16:20:08.590846 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b125bd6-f08f-442c-96a8-b903160be13a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8b125bd6-f08f-442c-96a8-b903160be13a" (UID: "8b125bd6-f08f-442c-96a8-b903160be13a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:20:08 crc kubenswrapper[4716]: I1207 16:20:08.643066 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jxj87\" (UniqueName: \"kubernetes.io/projected/8b125bd6-f08f-442c-96a8-b903160be13a-kube-api-access-jxj87\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:08 crc kubenswrapper[4716]: I1207 16:20:08.643116 4716 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/8b125bd6-f08f-442c-96a8-b903160be13a-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:08 crc kubenswrapper[4716]: I1207 16:20:08.643128 4716 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/8b125bd6-f08f-442c-96a8-b903160be13a-config\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:08 crc kubenswrapper[4716]: I1207 16:20:08.643137 4716 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b125bd6-f08f-442c-96a8-b903160be13a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:08 crc kubenswrapper[4716]: I1207 16:20:08.651240 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b125bd6-f08f-442c-96a8-b903160be13a-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "8b125bd6-f08f-442c-96a8-b903160be13a" (UID: "8b125bd6-f08f-442c-96a8-b903160be13a"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:20:08 crc kubenswrapper[4716]: I1207 16:20:08.745272 4716 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b125bd6-f08f-442c-96a8-b903160be13a-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:09 crc kubenswrapper[4716]: I1207 16:20:09.070581 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0480133a-7588-4a60-921e-e53de8998ca9","Type":"ContainerStarted","Data":"bbbcc700531f528352244efa205cfeeb3355cef62d7bf14bab368ee8dbe7b12a"} Dec 07 16:20:09 crc kubenswrapper[4716]: I1207 16:20:09.070817 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 07 16:20:09 crc kubenswrapper[4716]: I1207 16:20:09.073021 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-69d87fd46d-9k5rq" event={"ID":"8b125bd6-f08f-442c-96a8-b903160be13a","Type":"ContainerDied","Data":"b103baf8f60bf57fbf5072b72913996c5f5704ba2a74a093f2ea7f47dfced9cb"} Dec 07 16:20:09 crc kubenswrapper[4716]: I1207 16:20:09.073094 4716 scope.go:117] "RemoveContainer" containerID="9e0f9b2ae46383d32e78b438d5bf9227170de493caed859e596aafea34071033" Dec 07 16:20:09 crc kubenswrapper[4716]: I1207 16:20:09.073219 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-69d87fd46d-9k5rq" Dec 07 16:20:09 crc kubenswrapper[4716]: I1207 16:20:09.101797 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.465769623 podStartE2EDuration="5.101778202s" podCreationTimestamp="2025-12-07 16:20:04 +0000 UTC" firstStartedPulling="2025-12-07 16:20:05.011535566 +0000 UTC m=+1067.701820478" lastFinishedPulling="2025-12-07 16:20:08.647544145 +0000 UTC m=+1071.337829057" observedRunningTime="2025-12-07 16:20:09.097649231 +0000 UTC m=+1071.787934153" watchObservedRunningTime="2025-12-07 16:20:09.101778202 +0000 UTC m=+1071.792063124" Dec 07 16:20:09 crc kubenswrapper[4716]: I1207 16:20:09.109318 4716 scope.go:117] "RemoveContainer" containerID="c00c5a8c50e26dad3fc07ff6ed5d0663b13bd6897c9a5e4c05fae5699523e678" Dec 07 16:20:09 crc kubenswrapper[4716]: I1207 16:20:09.133673 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-69d87fd46d-9k5rq"] Dec 07 16:20:09 crc kubenswrapper[4716]: I1207 16:20:09.146757 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-69d87fd46d-9k5rq"] Dec 07 16:20:09 crc kubenswrapper[4716]: I1207 16:20:09.249016 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c9776ccc5-vb6n9" Dec 07 16:20:09 crc kubenswrapper[4716]: I1207 16:20:09.318264 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-rtqpb"] Dec 07 16:20:09 crc kubenswrapper[4716]: I1207 16:20:09.318704 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-55f844cf75-rtqpb" podUID="8b1a41e8-004e-4562-a8aa-79e422e11e83" containerName="dnsmasq-dns" containerID="cri-o://a37b3fb034928ee7bc75507f7ff97f1249c5e124632b04b9390037efa09b3d1d" gracePeriod=10 Dec 07 16:20:09 crc kubenswrapper[4716]: I1207 16:20:09.357347 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 07 16:20:09 crc kubenswrapper[4716]: I1207 16:20:09.429748 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 07 16:20:09 crc kubenswrapper[4716]: I1207 16:20:09.671432 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8b125bd6-f08f-442c-96a8-b903160be13a" path="/var/lib/kubelet/pods/8b125bd6-f08f-442c-96a8-b903160be13a/volumes" Dec 07 16:20:09 crc kubenswrapper[4716]: I1207 16:20:09.900603 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-rtqpb" Dec 07 16:20:09 crc kubenswrapper[4716]: I1207 16:20:09.970554 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b1a41e8-004e-4562-a8aa-79e422e11e83-config\") pod \"8b1a41e8-004e-4562-a8aa-79e422e11e83\" (UID: \"8b1a41e8-004e-4562-a8aa-79e422e11e83\") " Dec 07 16:20:09 crc kubenswrapper[4716]: I1207 16:20:09.970713 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8b1a41e8-004e-4562-a8aa-79e422e11e83-dns-svc\") pod \"8b1a41e8-004e-4562-a8aa-79e422e11e83\" (UID: \"8b1a41e8-004e-4562-a8aa-79e422e11e83\") " Dec 07 16:20:09 crc kubenswrapper[4716]: I1207 16:20:09.970755 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8b1a41e8-004e-4562-a8aa-79e422e11e83-dns-swift-storage-0\") pod \"8b1a41e8-004e-4562-a8aa-79e422e11e83\" (UID: \"8b1a41e8-004e-4562-a8aa-79e422e11e83\") " Dec 07 16:20:09 crc kubenswrapper[4716]: I1207 16:20:09.970776 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vqjwl\" (UniqueName: \"kubernetes.io/projected/8b1a41e8-004e-4562-a8aa-79e422e11e83-kube-api-access-vqjwl\") pod \"8b1a41e8-004e-4562-a8aa-79e422e11e83\" (UID: \"8b1a41e8-004e-4562-a8aa-79e422e11e83\") " Dec 07 16:20:09 crc kubenswrapper[4716]: I1207 16:20:09.970867 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8b1a41e8-004e-4562-a8aa-79e422e11e83-ovsdbserver-nb\") pod \"8b1a41e8-004e-4562-a8aa-79e422e11e83\" (UID: \"8b1a41e8-004e-4562-a8aa-79e422e11e83\") " Dec 07 16:20:09 crc kubenswrapper[4716]: I1207 16:20:09.970902 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8b1a41e8-004e-4562-a8aa-79e422e11e83-ovsdbserver-sb\") pod \"8b1a41e8-004e-4562-a8aa-79e422e11e83\" (UID: \"8b1a41e8-004e-4562-a8aa-79e422e11e83\") " Dec 07 16:20:10 crc kubenswrapper[4716]: I1207 16:20:10.011733 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b1a41e8-004e-4562-a8aa-79e422e11e83-kube-api-access-vqjwl" (OuterVolumeSpecName: "kube-api-access-vqjwl") pod "8b1a41e8-004e-4562-a8aa-79e422e11e83" (UID: "8b1a41e8-004e-4562-a8aa-79e422e11e83"). InnerVolumeSpecName "kube-api-access-vqjwl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:20:10 crc kubenswrapper[4716]: I1207 16:20:10.044242 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8b1a41e8-004e-4562-a8aa-79e422e11e83-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "8b1a41e8-004e-4562-a8aa-79e422e11e83" (UID: "8b1a41e8-004e-4562-a8aa-79e422e11e83"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:20:10 crc kubenswrapper[4716]: I1207 16:20:10.058324 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8b1a41e8-004e-4562-a8aa-79e422e11e83-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8b1a41e8-004e-4562-a8aa-79e422e11e83" (UID: "8b1a41e8-004e-4562-a8aa-79e422e11e83"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:20:10 crc kubenswrapper[4716]: I1207 16:20:10.059036 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8b1a41e8-004e-4562-a8aa-79e422e11e83-config" (OuterVolumeSpecName: "config") pod "8b1a41e8-004e-4562-a8aa-79e422e11e83" (UID: "8b1a41e8-004e-4562-a8aa-79e422e11e83"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:20:10 crc kubenswrapper[4716]: I1207 16:20:10.059466 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8b1a41e8-004e-4562-a8aa-79e422e11e83-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8b1a41e8-004e-4562-a8aa-79e422e11e83" (UID: "8b1a41e8-004e-4562-a8aa-79e422e11e83"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:20:10 crc kubenswrapper[4716]: I1207 16:20:10.063844 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8b1a41e8-004e-4562-a8aa-79e422e11e83-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8b1a41e8-004e-4562-a8aa-79e422e11e83" (UID: "8b1a41e8-004e-4562-a8aa-79e422e11e83"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:20:10 crc kubenswrapper[4716]: I1207 16:20:10.072816 4716 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8b1a41e8-004e-4562-a8aa-79e422e11e83-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:10 crc kubenswrapper[4716]: I1207 16:20:10.072840 4716 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8b1a41e8-004e-4562-a8aa-79e422e11e83-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:10 crc kubenswrapper[4716]: I1207 16:20:10.072851 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vqjwl\" (UniqueName: \"kubernetes.io/projected/8b1a41e8-004e-4562-a8aa-79e422e11e83-kube-api-access-vqjwl\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:10 crc kubenswrapper[4716]: I1207 16:20:10.072860 4716 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8b1a41e8-004e-4562-a8aa-79e422e11e83-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:10 crc kubenswrapper[4716]: I1207 16:20:10.072869 4716 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8b1a41e8-004e-4562-a8aa-79e422e11e83-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:10 crc kubenswrapper[4716]: I1207 16:20:10.072887 4716 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b1a41e8-004e-4562-a8aa-79e422e11e83-config\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:10 crc kubenswrapper[4716]: I1207 16:20:10.089963 4716 generic.go:334] "Generic (PLEG): container finished" podID="8b1a41e8-004e-4562-a8aa-79e422e11e83" containerID="a37b3fb034928ee7bc75507f7ff97f1249c5e124632b04b9390037efa09b3d1d" exitCode=0 Dec 07 16:20:10 crc kubenswrapper[4716]: I1207 16:20:10.090071 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-rtqpb" Dec 07 16:20:10 crc kubenswrapper[4716]: I1207 16:20:10.090073 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-rtqpb" event={"ID":"8b1a41e8-004e-4562-a8aa-79e422e11e83","Type":"ContainerDied","Data":"a37b3fb034928ee7bc75507f7ff97f1249c5e124632b04b9390037efa09b3d1d"} Dec 07 16:20:10 crc kubenswrapper[4716]: I1207 16:20:10.090499 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-rtqpb" event={"ID":"8b1a41e8-004e-4562-a8aa-79e422e11e83","Type":"ContainerDied","Data":"b8e11dd521785c2f3af96ea7a7ead0c4288d890f27725fba910d532cfb3c6209"} Dec 07 16:20:10 crc kubenswrapper[4716]: I1207 16:20:10.090528 4716 scope.go:117] "RemoveContainer" containerID="a37b3fb034928ee7bc75507f7ff97f1249c5e124632b04b9390037efa09b3d1d" Dec 07 16:20:10 crc kubenswrapper[4716]: I1207 16:20:10.104792 4716 generic.go:334] "Generic (PLEG): container finished" podID="c291e114-7940-46bc-91d3-a8fa256549ff" containerID="bc89298b4c5f29666ca1176ee3a6613d8d2303c5623d2bec2695ed37ae7b6619" exitCode=0 Dec 07 16:20:10 crc kubenswrapper[4716]: I1207 16:20:10.104968 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-86c7567d4-99rx9" event={"ID":"c291e114-7940-46bc-91d3-a8fa256549ff","Type":"ContainerDied","Data":"bc89298b4c5f29666ca1176ee3a6613d8d2303c5623d2bec2695ed37ae7b6619"} Dec 07 16:20:10 crc kubenswrapper[4716]: I1207 16:20:10.105361 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="03ba9a5b-efe2-4180-8110-cd8db3e65278" containerName="cinder-scheduler" containerID="cri-o://d9f29be0066946f7c67371ca40478054a4ae180dd2ab01cef84d12d1dde9ec07" gracePeriod=30 Dec 07 16:20:10 crc kubenswrapper[4716]: I1207 16:20:10.105626 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="03ba9a5b-efe2-4180-8110-cd8db3e65278" containerName="probe" containerID="cri-o://c7da8acbb1527ff4c40a062486446356c869e70e1245faa5c70f2fc914a79388" gracePeriod=30 Dec 07 16:20:10 crc kubenswrapper[4716]: I1207 16:20:10.114218 4716 scope.go:117] "RemoveContainer" containerID="0c15aa0ddf082834e5d336b546d2ca5f250cd7d2f66dea9b488bd3cf06fc2cac" Dec 07 16:20:10 crc kubenswrapper[4716]: I1207 16:20:10.145182 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-rtqpb"] Dec 07 16:20:10 crc kubenswrapper[4716]: I1207 16:20:10.145225 4716 scope.go:117] "RemoveContainer" containerID="a37b3fb034928ee7bc75507f7ff97f1249c5e124632b04b9390037efa09b3d1d" Dec 07 16:20:10 crc kubenswrapper[4716]: E1207 16:20:10.148135 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a37b3fb034928ee7bc75507f7ff97f1249c5e124632b04b9390037efa09b3d1d\": container with ID starting with a37b3fb034928ee7bc75507f7ff97f1249c5e124632b04b9390037efa09b3d1d not found: ID does not exist" containerID="a37b3fb034928ee7bc75507f7ff97f1249c5e124632b04b9390037efa09b3d1d" Dec 07 16:20:10 crc kubenswrapper[4716]: I1207 16:20:10.148168 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a37b3fb034928ee7bc75507f7ff97f1249c5e124632b04b9390037efa09b3d1d"} err="failed to get container status \"a37b3fb034928ee7bc75507f7ff97f1249c5e124632b04b9390037efa09b3d1d\": rpc error: code = NotFound desc = could not find container \"a37b3fb034928ee7bc75507f7ff97f1249c5e124632b04b9390037efa09b3d1d\": container with ID starting with a37b3fb034928ee7bc75507f7ff97f1249c5e124632b04b9390037efa09b3d1d not found: ID does not exist" Dec 07 16:20:10 crc kubenswrapper[4716]: I1207 16:20:10.148189 4716 scope.go:117] "RemoveContainer" containerID="0c15aa0ddf082834e5d336b546d2ca5f250cd7d2f66dea9b488bd3cf06fc2cac" Dec 07 16:20:10 crc kubenswrapper[4716]: E1207 16:20:10.148527 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c15aa0ddf082834e5d336b546d2ca5f250cd7d2f66dea9b488bd3cf06fc2cac\": container with ID starting with 0c15aa0ddf082834e5d336b546d2ca5f250cd7d2f66dea9b488bd3cf06fc2cac not found: ID does not exist" containerID="0c15aa0ddf082834e5d336b546d2ca5f250cd7d2f66dea9b488bd3cf06fc2cac" Dec 07 16:20:10 crc kubenswrapper[4716]: I1207 16:20:10.148652 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c15aa0ddf082834e5d336b546d2ca5f250cd7d2f66dea9b488bd3cf06fc2cac"} err="failed to get container status \"0c15aa0ddf082834e5d336b546d2ca5f250cd7d2f66dea9b488bd3cf06fc2cac\": rpc error: code = NotFound desc = could not find container \"0c15aa0ddf082834e5d336b546d2ca5f250cd7d2f66dea9b488bd3cf06fc2cac\": container with ID starting with 0c15aa0ddf082834e5d336b546d2ca5f250cd7d2f66dea9b488bd3cf06fc2cac not found: ID does not exist" Dec 07 16:20:10 crc kubenswrapper[4716]: I1207 16:20:10.152458 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-rtqpb"] Dec 07 16:20:10 crc kubenswrapper[4716]: I1207 16:20:10.718393 4716 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-86c7567d4-99rx9" podUID="c291e114-7940-46bc-91d3-a8fa256549ff" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Dec 07 16:20:11 crc kubenswrapper[4716]: I1207 16:20:11.223246 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-65b7f6996b-q6xwd" Dec 07 16:20:11 crc kubenswrapper[4716]: I1207 16:20:11.227990 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-65b7f6996b-q6xwd" Dec 07 16:20:11 crc kubenswrapper[4716]: I1207 16:20:11.668706 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8b1a41e8-004e-4562-a8aa-79e422e11e83" path="/var/lib/kubelet/pods/8b1a41e8-004e-4562-a8aa-79e422e11e83/volumes" Dec 07 16:20:11 crc kubenswrapper[4716]: I1207 16:20:11.669539 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 07 16:20:12 crc kubenswrapper[4716]: I1207 16:20:12.130684 4716 generic.go:334] "Generic (PLEG): container finished" podID="03ba9a5b-efe2-4180-8110-cd8db3e65278" containerID="c7da8acbb1527ff4c40a062486446356c869e70e1245faa5c70f2fc914a79388" exitCode=0 Dec 07 16:20:12 crc kubenswrapper[4716]: I1207 16:20:12.130764 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"03ba9a5b-efe2-4180-8110-cd8db3e65278","Type":"ContainerDied","Data":"c7da8acbb1527ff4c40a062486446356c869e70e1245faa5c70f2fc914a79388"} Dec 07 16:20:14 crc kubenswrapper[4716]: I1207 16:20:14.071358 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-746cf47744-w9nm7" Dec 07 16:20:14 crc kubenswrapper[4716]: I1207 16:20:14.132716 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-746cf47744-w9nm7" Dec 07 16:20:14 crc kubenswrapper[4716]: I1207 16:20:14.867846 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-6c4445878c-bnpkh" Dec 07 16:20:15 crc kubenswrapper[4716]: I1207 16:20:15.092759 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 07 16:20:15 crc kubenswrapper[4716]: I1207 16:20:15.215406 4716 generic.go:334] "Generic (PLEG): container finished" podID="03ba9a5b-efe2-4180-8110-cd8db3e65278" containerID="d9f29be0066946f7c67371ca40478054a4ae180dd2ab01cef84d12d1dde9ec07" exitCode=0 Dec 07 16:20:15 crc kubenswrapper[4716]: I1207 16:20:15.215451 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"03ba9a5b-efe2-4180-8110-cd8db3e65278","Type":"ContainerDied","Data":"d9f29be0066946f7c67371ca40478054a4ae180dd2ab01cef84d12d1dde9ec07"} Dec 07 16:20:15 crc kubenswrapper[4716]: I1207 16:20:15.215483 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"03ba9a5b-efe2-4180-8110-cd8db3e65278","Type":"ContainerDied","Data":"07eb774f0cfe53a228c73435dc6bfeab5cabcce9cf02b1271fe77ebd6c1a6044"} Dec 07 16:20:15 crc kubenswrapper[4716]: I1207 16:20:15.215502 4716 scope.go:117] "RemoveContainer" containerID="c7da8acbb1527ff4c40a062486446356c869e70e1245faa5c70f2fc914a79388" Dec 07 16:20:15 crc kubenswrapper[4716]: I1207 16:20:15.215626 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 07 16:20:15 crc kubenswrapper[4716]: I1207 16:20:15.254859 4716 scope.go:117] "RemoveContainer" containerID="d9f29be0066946f7c67371ca40478054a4ae180dd2ab01cef84d12d1dde9ec07" Dec 07 16:20:15 crc kubenswrapper[4716]: I1207 16:20:15.280330 4716 scope.go:117] "RemoveContainer" containerID="c7da8acbb1527ff4c40a062486446356c869e70e1245faa5c70f2fc914a79388" Dec 07 16:20:15 crc kubenswrapper[4716]: E1207 16:20:15.280815 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c7da8acbb1527ff4c40a062486446356c869e70e1245faa5c70f2fc914a79388\": container with ID starting with c7da8acbb1527ff4c40a062486446356c869e70e1245faa5c70f2fc914a79388 not found: ID does not exist" containerID="c7da8acbb1527ff4c40a062486446356c869e70e1245faa5c70f2fc914a79388" Dec 07 16:20:15 crc kubenswrapper[4716]: I1207 16:20:15.280878 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7da8acbb1527ff4c40a062486446356c869e70e1245faa5c70f2fc914a79388"} err="failed to get container status \"c7da8acbb1527ff4c40a062486446356c869e70e1245faa5c70f2fc914a79388\": rpc error: code = NotFound desc = could not find container \"c7da8acbb1527ff4c40a062486446356c869e70e1245faa5c70f2fc914a79388\": container with ID starting with c7da8acbb1527ff4c40a062486446356c869e70e1245faa5c70f2fc914a79388 not found: ID does not exist" Dec 07 16:20:15 crc kubenswrapper[4716]: I1207 16:20:15.280913 4716 scope.go:117] "RemoveContainer" containerID="d9f29be0066946f7c67371ca40478054a4ae180dd2ab01cef84d12d1dde9ec07" Dec 07 16:20:15 crc kubenswrapper[4716]: E1207 16:20:15.281278 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9f29be0066946f7c67371ca40478054a4ae180dd2ab01cef84d12d1dde9ec07\": container with ID starting with d9f29be0066946f7c67371ca40478054a4ae180dd2ab01cef84d12d1dde9ec07 not found: ID does not exist" containerID="d9f29be0066946f7c67371ca40478054a4ae180dd2ab01cef84d12d1dde9ec07" Dec 07 16:20:15 crc kubenswrapper[4716]: I1207 16:20:15.281310 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9f29be0066946f7c67371ca40478054a4ae180dd2ab01cef84d12d1dde9ec07"} err="failed to get container status \"d9f29be0066946f7c67371ca40478054a4ae180dd2ab01cef84d12d1dde9ec07\": rpc error: code = NotFound desc = could not find container \"d9f29be0066946f7c67371ca40478054a4ae180dd2ab01cef84d12d1dde9ec07\": container with ID starting with d9f29be0066946f7c67371ca40478054a4ae180dd2ab01cef84d12d1dde9ec07 not found: ID does not exist" Dec 07 16:20:15 crc kubenswrapper[4716]: I1207 16:20:15.292951 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/03ba9a5b-efe2-4180-8110-cd8db3e65278-scripts\") pod \"03ba9a5b-efe2-4180-8110-cd8db3e65278\" (UID: \"03ba9a5b-efe2-4180-8110-cd8db3e65278\") " Dec 07 16:20:15 crc kubenswrapper[4716]: I1207 16:20:15.293020 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03ba9a5b-efe2-4180-8110-cd8db3e65278-combined-ca-bundle\") pod \"03ba9a5b-efe2-4180-8110-cd8db3e65278\" (UID: \"03ba9a5b-efe2-4180-8110-cd8db3e65278\") " Dec 07 16:20:15 crc kubenswrapper[4716]: I1207 16:20:15.293058 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5t2cp\" (UniqueName: \"kubernetes.io/projected/03ba9a5b-efe2-4180-8110-cd8db3e65278-kube-api-access-5t2cp\") pod \"03ba9a5b-efe2-4180-8110-cd8db3e65278\" (UID: \"03ba9a5b-efe2-4180-8110-cd8db3e65278\") " Dec 07 16:20:15 crc kubenswrapper[4716]: I1207 16:20:15.293112 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/03ba9a5b-efe2-4180-8110-cd8db3e65278-etc-machine-id\") pod \"03ba9a5b-efe2-4180-8110-cd8db3e65278\" (UID: \"03ba9a5b-efe2-4180-8110-cd8db3e65278\") " Dec 07 16:20:15 crc kubenswrapper[4716]: I1207 16:20:15.293187 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03ba9a5b-efe2-4180-8110-cd8db3e65278-config-data\") pod \"03ba9a5b-efe2-4180-8110-cd8db3e65278\" (UID: \"03ba9a5b-efe2-4180-8110-cd8db3e65278\") " Dec 07 16:20:15 crc kubenswrapper[4716]: I1207 16:20:15.293290 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/03ba9a5b-efe2-4180-8110-cd8db3e65278-config-data-custom\") pod \"03ba9a5b-efe2-4180-8110-cd8db3e65278\" (UID: \"03ba9a5b-efe2-4180-8110-cd8db3e65278\") " Dec 07 16:20:15 crc kubenswrapper[4716]: I1207 16:20:15.295192 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/03ba9a5b-efe2-4180-8110-cd8db3e65278-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "03ba9a5b-efe2-4180-8110-cd8db3e65278" (UID: "03ba9a5b-efe2-4180-8110-cd8db3e65278"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 16:20:15 crc kubenswrapper[4716]: I1207 16:20:15.300155 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03ba9a5b-efe2-4180-8110-cd8db3e65278-scripts" (OuterVolumeSpecName: "scripts") pod "03ba9a5b-efe2-4180-8110-cd8db3e65278" (UID: "03ba9a5b-efe2-4180-8110-cd8db3e65278"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:20:15 crc kubenswrapper[4716]: I1207 16:20:15.315874 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03ba9a5b-efe2-4180-8110-cd8db3e65278-kube-api-access-5t2cp" (OuterVolumeSpecName: "kube-api-access-5t2cp") pod "03ba9a5b-efe2-4180-8110-cd8db3e65278" (UID: "03ba9a5b-efe2-4180-8110-cd8db3e65278"). InnerVolumeSpecName "kube-api-access-5t2cp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:20:15 crc kubenswrapper[4716]: I1207 16:20:15.315984 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03ba9a5b-efe2-4180-8110-cd8db3e65278-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "03ba9a5b-efe2-4180-8110-cd8db3e65278" (UID: "03ba9a5b-efe2-4180-8110-cd8db3e65278"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:20:15 crc kubenswrapper[4716]: I1207 16:20:15.364413 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03ba9a5b-efe2-4180-8110-cd8db3e65278-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "03ba9a5b-efe2-4180-8110-cd8db3e65278" (UID: "03ba9a5b-efe2-4180-8110-cd8db3e65278"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:20:15 crc kubenswrapper[4716]: I1207 16:20:15.395833 4716 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/03ba9a5b-efe2-4180-8110-cd8db3e65278-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:15 crc kubenswrapper[4716]: I1207 16:20:15.396064 4716 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/03ba9a5b-efe2-4180-8110-cd8db3e65278-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:15 crc kubenswrapper[4716]: I1207 16:20:15.396190 4716 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03ba9a5b-efe2-4180-8110-cd8db3e65278-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:15 crc kubenswrapper[4716]: I1207 16:20:15.396258 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5t2cp\" (UniqueName: \"kubernetes.io/projected/03ba9a5b-efe2-4180-8110-cd8db3e65278-kube-api-access-5t2cp\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:15 crc kubenswrapper[4716]: I1207 16:20:15.396313 4716 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/03ba9a5b-efe2-4180-8110-cd8db3e65278-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:15 crc kubenswrapper[4716]: I1207 16:20:15.403548 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03ba9a5b-efe2-4180-8110-cd8db3e65278-config-data" (OuterVolumeSpecName: "config-data") pod "03ba9a5b-efe2-4180-8110-cd8db3e65278" (UID: "03ba9a5b-efe2-4180-8110-cd8db3e65278"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:20:15 crc kubenswrapper[4716]: I1207 16:20:15.498519 4716 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03ba9a5b-efe2-4180-8110-cd8db3e65278-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:15 crc kubenswrapper[4716]: I1207 16:20:15.546987 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 07 16:20:15 crc kubenswrapper[4716]: I1207 16:20:15.567736 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 07 16:20:15 crc kubenswrapper[4716]: I1207 16:20:15.581306 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 07 16:20:15 crc kubenswrapper[4716]: E1207 16:20:15.581685 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b125bd6-f08f-442c-96a8-b903160be13a" containerName="neutron-httpd" Dec 07 16:20:15 crc kubenswrapper[4716]: I1207 16:20:15.581706 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b125bd6-f08f-442c-96a8-b903160be13a" containerName="neutron-httpd" Dec 07 16:20:15 crc kubenswrapper[4716]: E1207 16:20:15.581721 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b1a41e8-004e-4562-a8aa-79e422e11e83" containerName="dnsmasq-dns" Dec 07 16:20:15 crc kubenswrapper[4716]: I1207 16:20:15.581728 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b1a41e8-004e-4562-a8aa-79e422e11e83" containerName="dnsmasq-dns" Dec 07 16:20:15 crc kubenswrapper[4716]: E1207 16:20:15.581739 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03ba9a5b-efe2-4180-8110-cd8db3e65278" containerName="cinder-scheduler" Dec 07 16:20:15 crc kubenswrapper[4716]: I1207 16:20:15.581744 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="03ba9a5b-efe2-4180-8110-cd8db3e65278" containerName="cinder-scheduler" Dec 07 16:20:15 crc kubenswrapper[4716]: E1207 16:20:15.581761 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b1a41e8-004e-4562-a8aa-79e422e11e83" containerName="init" Dec 07 16:20:15 crc kubenswrapper[4716]: I1207 16:20:15.581766 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b1a41e8-004e-4562-a8aa-79e422e11e83" containerName="init" Dec 07 16:20:15 crc kubenswrapper[4716]: E1207 16:20:15.581775 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03ba9a5b-efe2-4180-8110-cd8db3e65278" containerName="probe" Dec 07 16:20:15 crc kubenswrapper[4716]: I1207 16:20:15.581781 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="03ba9a5b-efe2-4180-8110-cd8db3e65278" containerName="probe" Dec 07 16:20:15 crc kubenswrapper[4716]: E1207 16:20:15.581800 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b125bd6-f08f-442c-96a8-b903160be13a" containerName="neutron-api" Dec 07 16:20:15 crc kubenswrapper[4716]: I1207 16:20:15.581806 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b125bd6-f08f-442c-96a8-b903160be13a" containerName="neutron-api" Dec 07 16:20:15 crc kubenswrapper[4716]: I1207 16:20:15.581969 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="03ba9a5b-efe2-4180-8110-cd8db3e65278" containerName="cinder-scheduler" Dec 07 16:20:15 crc kubenswrapper[4716]: I1207 16:20:15.581979 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="03ba9a5b-efe2-4180-8110-cd8db3e65278" containerName="probe" Dec 07 16:20:15 crc kubenswrapper[4716]: I1207 16:20:15.581994 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b125bd6-f08f-442c-96a8-b903160be13a" containerName="neutron-api" Dec 07 16:20:15 crc kubenswrapper[4716]: I1207 16:20:15.582011 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b1a41e8-004e-4562-a8aa-79e422e11e83" containerName="dnsmasq-dns" Dec 07 16:20:15 crc kubenswrapper[4716]: I1207 16:20:15.582025 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b125bd6-f08f-442c-96a8-b903160be13a" containerName="neutron-httpd" Dec 07 16:20:15 crc kubenswrapper[4716]: I1207 16:20:15.583022 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 07 16:20:15 crc kubenswrapper[4716]: I1207 16:20:15.585672 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 07 16:20:15 crc kubenswrapper[4716]: I1207 16:20:15.600978 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6db216a8-54ce-4596-a4ab-bb24a787b027-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"6db216a8-54ce-4596-a4ab-bb24a787b027\") " pod="openstack/cinder-scheduler-0" Dec 07 16:20:15 crc kubenswrapper[4716]: I1207 16:20:15.601145 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6db216a8-54ce-4596-a4ab-bb24a787b027-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"6db216a8-54ce-4596-a4ab-bb24a787b027\") " pod="openstack/cinder-scheduler-0" Dec 07 16:20:15 crc kubenswrapper[4716]: I1207 16:20:15.601426 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cfjcg\" (UniqueName: \"kubernetes.io/projected/6db216a8-54ce-4596-a4ab-bb24a787b027-kube-api-access-cfjcg\") pod \"cinder-scheduler-0\" (UID: \"6db216a8-54ce-4596-a4ab-bb24a787b027\") " pod="openstack/cinder-scheduler-0" Dec 07 16:20:15 crc kubenswrapper[4716]: I1207 16:20:15.601539 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6db216a8-54ce-4596-a4ab-bb24a787b027-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"6db216a8-54ce-4596-a4ab-bb24a787b027\") " pod="openstack/cinder-scheduler-0" Dec 07 16:20:15 crc kubenswrapper[4716]: I1207 16:20:15.601591 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6db216a8-54ce-4596-a4ab-bb24a787b027-config-data\") pod \"cinder-scheduler-0\" (UID: \"6db216a8-54ce-4596-a4ab-bb24a787b027\") " pod="openstack/cinder-scheduler-0" Dec 07 16:20:15 crc kubenswrapper[4716]: I1207 16:20:15.601662 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6db216a8-54ce-4596-a4ab-bb24a787b027-scripts\") pod \"cinder-scheduler-0\" (UID: \"6db216a8-54ce-4596-a4ab-bb24a787b027\") " pod="openstack/cinder-scheduler-0" Dec 07 16:20:15 crc kubenswrapper[4716]: I1207 16:20:15.604200 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 07 16:20:15 crc kubenswrapper[4716]: I1207 16:20:15.669188 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03ba9a5b-efe2-4180-8110-cd8db3e65278" path="/var/lib/kubelet/pods/03ba9a5b-efe2-4180-8110-cd8db3e65278/volumes" Dec 07 16:20:15 crc kubenswrapper[4716]: I1207 16:20:15.703645 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6db216a8-54ce-4596-a4ab-bb24a787b027-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"6db216a8-54ce-4596-a4ab-bb24a787b027\") " pod="openstack/cinder-scheduler-0" Dec 07 16:20:15 crc kubenswrapper[4716]: I1207 16:20:15.703729 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6db216a8-54ce-4596-a4ab-bb24a787b027-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"6db216a8-54ce-4596-a4ab-bb24a787b027\") " pod="openstack/cinder-scheduler-0" Dec 07 16:20:15 crc kubenswrapper[4716]: I1207 16:20:15.703852 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cfjcg\" (UniqueName: \"kubernetes.io/projected/6db216a8-54ce-4596-a4ab-bb24a787b027-kube-api-access-cfjcg\") pod \"cinder-scheduler-0\" (UID: \"6db216a8-54ce-4596-a4ab-bb24a787b027\") " pod="openstack/cinder-scheduler-0" Dec 07 16:20:15 crc kubenswrapper[4716]: I1207 16:20:15.703887 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6db216a8-54ce-4596-a4ab-bb24a787b027-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"6db216a8-54ce-4596-a4ab-bb24a787b027\") " pod="openstack/cinder-scheduler-0" Dec 07 16:20:15 crc kubenswrapper[4716]: I1207 16:20:15.703918 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6db216a8-54ce-4596-a4ab-bb24a787b027-config-data\") pod \"cinder-scheduler-0\" (UID: \"6db216a8-54ce-4596-a4ab-bb24a787b027\") " pod="openstack/cinder-scheduler-0" Dec 07 16:20:15 crc kubenswrapper[4716]: I1207 16:20:15.703956 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6db216a8-54ce-4596-a4ab-bb24a787b027-scripts\") pod \"cinder-scheduler-0\" (UID: \"6db216a8-54ce-4596-a4ab-bb24a787b027\") " pod="openstack/cinder-scheduler-0" Dec 07 16:20:15 crc kubenswrapper[4716]: I1207 16:20:15.706804 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6db216a8-54ce-4596-a4ab-bb24a787b027-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"6db216a8-54ce-4596-a4ab-bb24a787b027\") " pod="openstack/cinder-scheduler-0" Dec 07 16:20:15 crc kubenswrapper[4716]: I1207 16:20:15.708543 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6db216a8-54ce-4596-a4ab-bb24a787b027-scripts\") pod \"cinder-scheduler-0\" (UID: \"6db216a8-54ce-4596-a4ab-bb24a787b027\") " pod="openstack/cinder-scheduler-0" Dec 07 16:20:15 crc kubenswrapper[4716]: I1207 16:20:15.708631 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6db216a8-54ce-4596-a4ab-bb24a787b027-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"6db216a8-54ce-4596-a4ab-bb24a787b027\") " pod="openstack/cinder-scheduler-0" Dec 07 16:20:15 crc kubenswrapper[4716]: I1207 16:20:15.709466 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6db216a8-54ce-4596-a4ab-bb24a787b027-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"6db216a8-54ce-4596-a4ab-bb24a787b027\") " pod="openstack/cinder-scheduler-0" Dec 07 16:20:15 crc kubenswrapper[4716]: I1207 16:20:15.710572 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6db216a8-54ce-4596-a4ab-bb24a787b027-config-data\") pod \"cinder-scheduler-0\" (UID: \"6db216a8-54ce-4596-a4ab-bb24a787b027\") " pod="openstack/cinder-scheduler-0" Dec 07 16:20:15 crc kubenswrapper[4716]: I1207 16:20:15.729443 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cfjcg\" (UniqueName: \"kubernetes.io/projected/6db216a8-54ce-4596-a4ab-bb24a787b027-kube-api-access-cfjcg\") pod \"cinder-scheduler-0\" (UID: \"6db216a8-54ce-4596-a4ab-bb24a787b027\") " pod="openstack/cinder-scheduler-0" Dec 07 16:20:15 crc kubenswrapper[4716]: I1207 16:20:15.904483 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 07 16:20:16 crc kubenswrapper[4716]: I1207 16:20:16.402625 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 07 16:20:16 crc kubenswrapper[4716]: I1207 16:20:16.929899 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 07 16:20:16 crc kubenswrapper[4716]: I1207 16:20:16.931331 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 07 16:20:16 crc kubenswrapper[4716]: I1207 16:20:16.935094 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Dec 07 16:20:16 crc kubenswrapper[4716]: I1207 16:20:16.935357 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Dec 07 16:20:16 crc kubenswrapper[4716]: I1207 16:20:16.936959 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-c9fkc" Dec 07 16:20:16 crc kubenswrapper[4716]: I1207 16:20:16.939738 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a6ecf77-dd60-497a-8670-919d23db95dd-combined-ca-bundle\") pod \"openstackclient\" (UID: \"6a6ecf77-dd60-497a-8670-919d23db95dd\") " pod="openstack/openstackclient" Dec 07 16:20:16 crc kubenswrapper[4716]: I1207 16:20:16.939778 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nkbmj\" (UniqueName: \"kubernetes.io/projected/6a6ecf77-dd60-497a-8670-919d23db95dd-kube-api-access-nkbmj\") pod \"openstackclient\" (UID: \"6a6ecf77-dd60-497a-8670-919d23db95dd\") " pod="openstack/openstackclient" Dec 07 16:20:16 crc kubenswrapper[4716]: I1207 16:20:16.939852 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/6a6ecf77-dd60-497a-8670-919d23db95dd-openstack-config\") pod \"openstackclient\" (UID: \"6a6ecf77-dd60-497a-8670-919d23db95dd\") " pod="openstack/openstackclient" Dec 07 16:20:16 crc kubenswrapper[4716]: I1207 16:20:16.939873 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/6a6ecf77-dd60-497a-8670-919d23db95dd-openstack-config-secret\") pod \"openstackclient\" (UID: \"6a6ecf77-dd60-497a-8670-919d23db95dd\") " pod="openstack/openstackclient" Dec 07 16:20:16 crc kubenswrapper[4716]: I1207 16:20:16.949295 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 07 16:20:17 crc kubenswrapper[4716]: I1207 16:20:17.041388 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/6a6ecf77-dd60-497a-8670-919d23db95dd-openstack-config\") pod \"openstackclient\" (UID: \"6a6ecf77-dd60-497a-8670-919d23db95dd\") " pod="openstack/openstackclient" Dec 07 16:20:17 crc kubenswrapper[4716]: I1207 16:20:17.041434 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/6a6ecf77-dd60-497a-8670-919d23db95dd-openstack-config-secret\") pod \"openstackclient\" (UID: \"6a6ecf77-dd60-497a-8670-919d23db95dd\") " pod="openstack/openstackclient" Dec 07 16:20:17 crc kubenswrapper[4716]: I1207 16:20:17.041521 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a6ecf77-dd60-497a-8670-919d23db95dd-combined-ca-bundle\") pod \"openstackclient\" (UID: \"6a6ecf77-dd60-497a-8670-919d23db95dd\") " pod="openstack/openstackclient" Dec 07 16:20:17 crc kubenswrapper[4716]: I1207 16:20:17.041542 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nkbmj\" (UniqueName: \"kubernetes.io/projected/6a6ecf77-dd60-497a-8670-919d23db95dd-kube-api-access-nkbmj\") pod \"openstackclient\" (UID: \"6a6ecf77-dd60-497a-8670-919d23db95dd\") " pod="openstack/openstackclient" Dec 07 16:20:17 crc kubenswrapper[4716]: I1207 16:20:17.042322 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/6a6ecf77-dd60-497a-8670-919d23db95dd-openstack-config\") pod \"openstackclient\" (UID: \"6a6ecf77-dd60-497a-8670-919d23db95dd\") " pod="openstack/openstackclient" Dec 07 16:20:17 crc kubenswrapper[4716]: I1207 16:20:17.045804 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a6ecf77-dd60-497a-8670-919d23db95dd-combined-ca-bundle\") pod \"openstackclient\" (UID: \"6a6ecf77-dd60-497a-8670-919d23db95dd\") " pod="openstack/openstackclient" Dec 07 16:20:17 crc kubenswrapper[4716]: I1207 16:20:17.046466 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/6a6ecf77-dd60-497a-8670-919d23db95dd-openstack-config-secret\") pod \"openstackclient\" (UID: \"6a6ecf77-dd60-497a-8670-919d23db95dd\") " pod="openstack/openstackclient" Dec 07 16:20:17 crc kubenswrapper[4716]: I1207 16:20:17.091209 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nkbmj\" (UniqueName: \"kubernetes.io/projected/6a6ecf77-dd60-497a-8670-919d23db95dd-kube-api-access-nkbmj\") pod \"openstackclient\" (UID: \"6a6ecf77-dd60-497a-8670-919d23db95dd\") " pod="openstack/openstackclient" Dec 07 16:20:17 crc kubenswrapper[4716]: I1207 16:20:17.127737 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-64c898bd6-wgw8m" Dec 07 16:20:17 crc kubenswrapper[4716]: I1207 16:20:17.247031 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6db216a8-54ce-4596-a4ab-bb24a787b027","Type":"ContainerStarted","Data":"ecfc7f7bd91ccbf2c72a24cedf1c807fbe52a0451ed6766a5b9f94e6a9f9955a"} Dec 07 16:20:17 crc kubenswrapper[4716]: I1207 16:20:17.247702 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6db216a8-54ce-4596-a4ab-bb24a787b027","Type":"ContainerStarted","Data":"28dc0a06f6f2dd8d33e3600e512d95eded95ea5b3d52ccabc47f537672a6653b"} Dec 07 16:20:17 crc kubenswrapper[4716]: I1207 16:20:17.247895 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 07 16:20:17 crc kubenswrapper[4716]: I1207 16:20:17.276939 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-64c898bd6-wgw8m" Dec 07 16:20:17 crc kubenswrapper[4716]: I1207 16:20:17.352398 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-65b7f6996b-q6xwd"] Dec 07 16:20:17 crc kubenswrapper[4716]: I1207 16:20:17.352696 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-65b7f6996b-q6xwd" podUID="65aaa058-a1be-4477-b808-d0ea1d8e15fe" containerName="barbican-api-log" containerID="cri-o://110da32e8b7d13bd7f08d374547e0e1ba723851317ad3a5d5cff00362d5ca8b4" gracePeriod=30 Dec 07 16:20:17 crc kubenswrapper[4716]: I1207 16:20:17.353271 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-65b7f6996b-q6xwd" podUID="65aaa058-a1be-4477-b808-d0ea1d8e15fe" containerName="barbican-api" containerID="cri-o://e21b4da38239a0d684080040edbc4305151f056711ac673ac6b4cb5a5648470c" gracePeriod=30 Dec 07 16:20:17 crc kubenswrapper[4716]: I1207 16:20:17.818989 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 07 16:20:18 crc kubenswrapper[4716]: I1207 16:20:18.264415 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"6a6ecf77-dd60-497a-8670-919d23db95dd","Type":"ContainerStarted","Data":"c32df4802f606c8caba245f540e424db83663ed2afc0e75f3269d5bd50eabfdd"} Dec 07 16:20:18 crc kubenswrapper[4716]: I1207 16:20:18.277559 4716 generic.go:334] "Generic (PLEG): container finished" podID="65aaa058-a1be-4477-b808-d0ea1d8e15fe" containerID="110da32e8b7d13bd7f08d374547e0e1ba723851317ad3a5d5cff00362d5ca8b4" exitCode=143 Dec 07 16:20:18 crc kubenswrapper[4716]: I1207 16:20:18.277630 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-65b7f6996b-q6xwd" event={"ID":"65aaa058-a1be-4477-b808-d0ea1d8e15fe","Type":"ContainerDied","Data":"110da32e8b7d13bd7f08d374547e0e1ba723851317ad3a5d5cff00362d5ca8b4"} Dec 07 16:20:18 crc kubenswrapper[4716]: I1207 16:20:18.281862 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6db216a8-54ce-4596-a4ab-bb24a787b027","Type":"ContainerStarted","Data":"1ee5321c8e91c38456b689f1e04d604271cc6137ace7a0a6eeb56148e2f57a66"} Dec 07 16:20:18 crc kubenswrapper[4716]: I1207 16:20:18.313185 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.313162512 podStartE2EDuration="3.313162512s" podCreationTimestamp="2025-12-07 16:20:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:20:18.301520316 +0000 UTC m=+1080.991805228" watchObservedRunningTime="2025-12-07 16:20:18.313162512 +0000 UTC m=+1081.003447424" Dec 07 16:20:20 crc kubenswrapper[4716]: I1207 16:20:20.526469 4716 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-65b7f6996b-q6xwd" podUID="65aaa058-a1be-4477-b808-d0ea1d8e15fe" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.162:9311/healthcheck\": read tcp 10.217.0.2:42534->10.217.0.162:9311: read: connection reset by peer" Dec 07 16:20:20 crc kubenswrapper[4716]: I1207 16:20:20.526502 4716 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-65b7f6996b-q6xwd" podUID="65aaa058-a1be-4477-b808-d0ea1d8e15fe" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.162:9311/healthcheck\": read tcp 10.217.0.2:42524->10.217.0.162:9311: read: connection reset by peer" Dec 07 16:20:20 crc kubenswrapper[4716]: I1207 16:20:20.718025 4716 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-86c7567d4-99rx9" podUID="c291e114-7940-46bc-91d3-a8fa256549ff" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Dec 07 16:20:20 crc kubenswrapper[4716]: I1207 16:20:20.904950 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 07 16:20:20 crc kubenswrapper[4716]: I1207 16:20:20.938566 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-65b7f6996b-q6xwd" Dec 07 16:20:21 crc kubenswrapper[4716]: I1207 16:20:21.037489 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wjbk7\" (UniqueName: \"kubernetes.io/projected/65aaa058-a1be-4477-b808-d0ea1d8e15fe-kube-api-access-wjbk7\") pod \"65aaa058-a1be-4477-b808-d0ea1d8e15fe\" (UID: \"65aaa058-a1be-4477-b808-d0ea1d8e15fe\") " Dec 07 16:20:21 crc kubenswrapper[4716]: I1207 16:20:21.037599 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/65aaa058-a1be-4477-b808-d0ea1d8e15fe-config-data-custom\") pod \"65aaa058-a1be-4477-b808-d0ea1d8e15fe\" (UID: \"65aaa058-a1be-4477-b808-d0ea1d8e15fe\") " Dec 07 16:20:21 crc kubenswrapper[4716]: I1207 16:20:21.037734 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65aaa058-a1be-4477-b808-d0ea1d8e15fe-combined-ca-bundle\") pod \"65aaa058-a1be-4477-b808-d0ea1d8e15fe\" (UID: \"65aaa058-a1be-4477-b808-d0ea1d8e15fe\") " Dec 07 16:20:21 crc kubenswrapper[4716]: I1207 16:20:21.037799 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65aaa058-a1be-4477-b808-d0ea1d8e15fe-config-data\") pod \"65aaa058-a1be-4477-b808-d0ea1d8e15fe\" (UID: \"65aaa058-a1be-4477-b808-d0ea1d8e15fe\") " Dec 07 16:20:21 crc kubenswrapper[4716]: I1207 16:20:21.037866 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/65aaa058-a1be-4477-b808-d0ea1d8e15fe-logs\") pod \"65aaa058-a1be-4477-b808-d0ea1d8e15fe\" (UID: \"65aaa058-a1be-4477-b808-d0ea1d8e15fe\") " Dec 07 16:20:21 crc kubenswrapper[4716]: I1207 16:20:21.039497 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/65aaa058-a1be-4477-b808-d0ea1d8e15fe-logs" (OuterVolumeSpecName: "logs") pod "65aaa058-a1be-4477-b808-d0ea1d8e15fe" (UID: "65aaa058-a1be-4477-b808-d0ea1d8e15fe"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:20:21 crc kubenswrapper[4716]: I1207 16:20:21.040225 4716 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/65aaa058-a1be-4477-b808-d0ea1d8e15fe-logs\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:21 crc kubenswrapper[4716]: I1207 16:20:21.043585 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65aaa058-a1be-4477-b808-d0ea1d8e15fe-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "65aaa058-a1be-4477-b808-d0ea1d8e15fe" (UID: "65aaa058-a1be-4477-b808-d0ea1d8e15fe"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:20:21 crc kubenswrapper[4716]: I1207 16:20:21.049901 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65aaa058-a1be-4477-b808-d0ea1d8e15fe-kube-api-access-wjbk7" (OuterVolumeSpecName: "kube-api-access-wjbk7") pod "65aaa058-a1be-4477-b808-d0ea1d8e15fe" (UID: "65aaa058-a1be-4477-b808-d0ea1d8e15fe"). InnerVolumeSpecName "kube-api-access-wjbk7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:20:21 crc kubenswrapper[4716]: I1207 16:20:21.070223 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65aaa058-a1be-4477-b808-d0ea1d8e15fe-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "65aaa058-a1be-4477-b808-d0ea1d8e15fe" (UID: "65aaa058-a1be-4477-b808-d0ea1d8e15fe"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:20:21 crc kubenswrapper[4716]: I1207 16:20:21.092098 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65aaa058-a1be-4477-b808-d0ea1d8e15fe-config-data" (OuterVolumeSpecName: "config-data") pod "65aaa058-a1be-4477-b808-d0ea1d8e15fe" (UID: "65aaa058-a1be-4477-b808-d0ea1d8e15fe"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:20:21 crc kubenswrapper[4716]: I1207 16:20:21.141692 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wjbk7\" (UniqueName: \"kubernetes.io/projected/65aaa058-a1be-4477-b808-d0ea1d8e15fe-kube-api-access-wjbk7\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:21 crc kubenswrapper[4716]: I1207 16:20:21.141729 4716 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/65aaa058-a1be-4477-b808-d0ea1d8e15fe-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:21 crc kubenswrapper[4716]: I1207 16:20:21.141740 4716 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65aaa058-a1be-4477-b808-d0ea1d8e15fe-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:21 crc kubenswrapper[4716]: I1207 16:20:21.141752 4716 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65aaa058-a1be-4477-b808-d0ea1d8e15fe-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:21 crc kubenswrapper[4716]: I1207 16:20:21.250575 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-5488d7fd99-pwmf5"] Dec 07 16:20:21 crc kubenswrapper[4716]: E1207 16:20:21.250956 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65aaa058-a1be-4477-b808-d0ea1d8e15fe" containerName="barbican-api" Dec 07 16:20:21 crc kubenswrapper[4716]: I1207 16:20:21.250975 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="65aaa058-a1be-4477-b808-d0ea1d8e15fe" containerName="barbican-api" Dec 07 16:20:21 crc kubenswrapper[4716]: E1207 16:20:21.250991 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65aaa058-a1be-4477-b808-d0ea1d8e15fe" containerName="barbican-api-log" Dec 07 16:20:21 crc kubenswrapper[4716]: I1207 16:20:21.250998 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="65aaa058-a1be-4477-b808-d0ea1d8e15fe" containerName="barbican-api-log" Dec 07 16:20:21 crc kubenswrapper[4716]: I1207 16:20:21.253246 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="65aaa058-a1be-4477-b808-d0ea1d8e15fe" containerName="barbican-api-log" Dec 07 16:20:21 crc kubenswrapper[4716]: I1207 16:20:21.253275 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="65aaa058-a1be-4477-b808-d0ea1d8e15fe" containerName="barbican-api" Dec 07 16:20:21 crc kubenswrapper[4716]: I1207 16:20:21.258456 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-5488d7fd99-pwmf5" Dec 07 16:20:21 crc kubenswrapper[4716]: I1207 16:20:21.260442 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Dec 07 16:20:21 crc kubenswrapper[4716]: I1207 16:20:21.260600 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 07 16:20:21 crc kubenswrapper[4716]: I1207 16:20:21.260610 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Dec 07 16:20:21 crc kubenswrapper[4716]: I1207 16:20:21.267367 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-5488d7fd99-pwmf5"] Dec 07 16:20:21 crc kubenswrapper[4716]: I1207 16:20:21.315053 4716 generic.go:334] "Generic (PLEG): container finished" podID="65aaa058-a1be-4477-b808-d0ea1d8e15fe" containerID="e21b4da38239a0d684080040edbc4305151f056711ac673ac6b4cb5a5648470c" exitCode=0 Dec 07 16:20:21 crc kubenswrapper[4716]: I1207 16:20:21.315110 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-65b7f6996b-q6xwd" event={"ID":"65aaa058-a1be-4477-b808-d0ea1d8e15fe","Type":"ContainerDied","Data":"e21b4da38239a0d684080040edbc4305151f056711ac673ac6b4cb5a5648470c"} Dec 07 16:20:21 crc kubenswrapper[4716]: I1207 16:20:21.315141 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-65b7f6996b-q6xwd" event={"ID":"65aaa058-a1be-4477-b808-d0ea1d8e15fe","Type":"ContainerDied","Data":"984ec5665a5889d05f6cc8d4025b879e51827ad9c65a2b8b61ddd700caab4c68"} Dec 07 16:20:21 crc kubenswrapper[4716]: I1207 16:20:21.315163 4716 scope.go:117] "RemoveContainer" containerID="e21b4da38239a0d684080040edbc4305151f056711ac673ac6b4cb5a5648470c" Dec 07 16:20:21 crc kubenswrapper[4716]: I1207 16:20:21.315290 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-65b7f6996b-q6xwd" Dec 07 16:20:21 crc kubenswrapper[4716]: I1207 16:20:21.357143 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-65b7f6996b-q6xwd"] Dec 07 16:20:21 crc kubenswrapper[4716]: I1207 16:20:21.370106 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-65b7f6996b-q6xwd"] Dec 07 16:20:21 crc kubenswrapper[4716]: I1207 16:20:21.381488 4716 scope.go:117] "RemoveContainer" containerID="110da32e8b7d13bd7f08d374547e0e1ba723851317ad3a5d5cff00362d5ca8b4" Dec 07 16:20:21 crc kubenswrapper[4716]: I1207 16:20:21.406635 4716 scope.go:117] "RemoveContainer" containerID="e21b4da38239a0d684080040edbc4305151f056711ac673ac6b4cb5a5648470c" Dec 07 16:20:21 crc kubenswrapper[4716]: E1207 16:20:21.407164 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e21b4da38239a0d684080040edbc4305151f056711ac673ac6b4cb5a5648470c\": container with ID starting with e21b4da38239a0d684080040edbc4305151f056711ac673ac6b4cb5a5648470c not found: ID does not exist" containerID="e21b4da38239a0d684080040edbc4305151f056711ac673ac6b4cb5a5648470c" Dec 07 16:20:21 crc kubenswrapper[4716]: I1207 16:20:21.407199 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e21b4da38239a0d684080040edbc4305151f056711ac673ac6b4cb5a5648470c"} err="failed to get container status \"e21b4da38239a0d684080040edbc4305151f056711ac673ac6b4cb5a5648470c\": rpc error: code = NotFound desc = could not find container \"e21b4da38239a0d684080040edbc4305151f056711ac673ac6b4cb5a5648470c\": container with ID starting with e21b4da38239a0d684080040edbc4305151f056711ac673ac6b4cb5a5648470c not found: ID does not exist" Dec 07 16:20:21 crc kubenswrapper[4716]: I1207 16:20:21.407227 4716 scope.go:117] "RemoveContainer" containerID="110da32e8b7d13bd7f08d374547e0e1ba723851317ad3a5d5cff00362d5ca8b4" Dec 07 16:20:21 crc kubenswrapper[4716]: E1207 16:20:21.407492 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"110da32e8b7d13bd7f08d374547e0e1ba723851317ad3a5d5cff00362d5ca8b4\": container with ID starting with 110da32e8b7d13bd7f08d374547e0e1ba723851317ad3a5d5cff00362d5ca8b4 not found: ID does not exist" containerID="110da32e8b7d13bd7f08d374547e0e1ba723851317ad3a5d5cff00362d5ca8b4" Dec 07 16:20:21 crc kubenswrapper[4716]: I1207 16:20:21.407526 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"110da32e8b7d13bd7f08d374547e0e1ba723851317ad3a5d5cff00362d5ca8b4"} err="failed to get container status \"110da32e8b7d13bd7f08d374547e0e1ba723851317ad3a5d5cff00362d5ca8b4\": rpc error: code = NotFound desc = could not find container \"110da32e8b7d13bd7f08d374547e0e1ba723851317ad3a5d5cff00362d5ca8b4\": container with ID starting with 110da32e8b7d13bd7f08d374547e0e1ba723851317ad3a5d5cff00362d5ca8b4 not found: ID does not exist" Dec 07 16:20:21 crc kubenswrapper[4716]: I1207 16:20:21.447191 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/826c0644-3046-453f-a139-5bd9c95216d0-run-httpd\") pod \"swift-proxy-5488d7fd99-pwmf5\" (UID: \"826c0644-3046-453f-a139-5bd9c95216d0\") " pod="openstack/swift-proxy-5488d7fd99-pwmf5" Dec 07 16:20:21 crc kubenswrapper[4716]: I1207 16:20:21.447278 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/826c0644-3046-453f-a139-5bd9c95216d0-log-httpd\") pod \"swift-proxy-5488d7fd99-pwmf5\" (UID: \"826c0644-3046-453f-a139-5bd9c95216d0\") " pod="openstack/swift-proxy-5488d7fd99-pwmf5" Dec 07 16:20:21 crc kubenswrapper[4716]: I1207 16:20:21.447350 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hq99\" (UniqueName: \"kubernetes.io/projected/826c0644-3046-453f-a139-5bd9c95216d0-kube-api-access-4hq99\") pod \"swift-proxy-5488d7fd99-pwmf5\" (UID: \"826c0644-3046-453f-a139-5bd9c95216d0\") " pod="openstack/swift-proxy-5488d7fd99-pwmf5" Dec 07 16:20:21 crc kubenswrapper[4716]: I1207 16:20:21.447372 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/826c0644-3046-453f-a139-5bd9c95216d0-internal-tls-certs\") pod \"swift-proxy-5488d7fd99-pwmf5\" (UID: \"826c0644-3046-453f-a139-5bd9c95216d0\") " pod="openstack/swift-proxy-5488d7fd99-pwmf5" Dec 07 16:20:21 crc kubenswrapper[4716]: I1207 16:20:21.447411 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/826c0644-3046-453f-a139-5bd9c95216d0-config-data\") pod \"swift-proxy-5488d7fd99-pwmf5\" (UID: \"826c0644-3046-453f-a139-5bd9c95216d0\") " pod="openstack/swift-proxy-5488d7fd99-pwmf5" Dec 07 16:20:21 crc kubenswrapper[4716]: I1207 16:20:21.447462 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/826c0644-3046-453f-a139-5bd9c95216d0-combined-ca-bundle\") pod \"swift-proxy-5488d7fd99-pwmf5\" (UID: \"826c0644-3046-453f-a139-5bd9c95216d0\") " pod="openstack/swift-proxy-5488d7fd99-pwmf5" Dec 07 16:20:21 crc kubenswrapper[4716]: I1207 16:20:21.447593 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/826c0644-3046-453f-a139-5bd9c95216d0-etc-swift\") pod \"swift-proxy-5488d7fd99-pwmf5\" (UID: \"826c0644-3046-453f-a139-5bd9c95216d0\") " pod="openstack/swift-proxy-5488d7fd99-pwmf5" Dec 07 16:20:21 crc kubenswrapper[4716]: I1207 16:20:21.447618 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/826c0644-3046-453f-a139-5bd9c95216d0-public-tls-certs\") pod \"swift-proxy-5488d7fd99-pwmf5\" (UID: \"826c0644-3046-453f-a139-5bd9c95216d0\") " pod="openstack/swift-proxy-5488d7fd99-pwmf5" Dec 07 16:20:21 crc kubenswrapper[4716]: I1207 16:20:21.548898 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/826c0644-3046-453f-a139-5bd9c95216d0-log-httpd\") pod \"swift-proxy-5488d7fd99-pwmf5\" (UID: \"826c0644-3046-453f-a139-5bd9c95216d0\") " pod="openstack/swift-proxy-5488d7fd99-pwmf5" Dec 07 16:20:21 crc kubenswrapper[4716]: I1207 16:20:21.549247 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hq99\" (UniqueName: \"kubernetes.io/projected/826c0644-3046-453f-a139-5bd9c95216d0-kube-api-access-4hq99\") pod \"swift-proxy-5488d7fd99-pwmf5\" (UID: \"826c0644-3046-453f-a139-5bd9c95216d0\") " pod="openstack/swift-proxy-5488d7fd99-pwmf5" Dec 07 16:20:21 crc kubenswrapper[4716]: I1207 16:20:21.549269 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/826c0644-3046-453f-a139-5bd9c95216d0-internal-tls-certs\") pod \"swift-proxy-5488d7fd99-pwmf5\" (UID: \"826c0644-3046-453f-a139-5bd9c95216d0\") " pod="openstack/swift-proxy-5488d7fd99-pwmf5" Dec 07 16:20:21 crc kubenswrapper[4716]: I1207 16:20:21.549294 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/826c0644-3046-453f-a139-5bd9c95216d0-config-data\") pod \"swift-proxy-5488d7fd99-pwmf5\" (UID: \"826c0644-3046-453f-a139-5bd9c95216d0\") " pod="openstack/swift-proxy-5488d7fd99-pwmf5" Dec 07 16:20:21 crc kubenswrapper[4716]: I1207 16:20:21.549343 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/826c0644-3046-453f-a139-5bd9c95216d0-combined-ca-bundle\") pod \"swift-proxy-5488d7fd99-pwmf5\" (UID: \"826c0644-3046-453f-a139-5bd9c95216d0\") " pod="openstack/swift-proxy-5488d7fd99-pwmf5" Dec 07 16:20:21 crc kubenswrapper[4716]: I1207 16:20:21.549375 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/826c0644-3046-453f-a139-5bd9c95216d0-etc-swift\") pod \"swift-proxy-5488d7fd99-pwmf5\" (UID: \"826c0644-3046-453f-a139-5bd9c95216d0\") " pod="openstack/swift-proxy-5488d7fd99-pwmf5" Dec 07 16:20:21 crc kubenswrapper[4716]: I1207 16:20:21.549394 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/826c0644-3046-453f-a139-5bd9c95216d0-public-tls-certs\") pod \"swift-proxy-5488d7fd99-pwmf5\" (UID: \"826c0644-3046-453f-a139-5bd9c95216d0\") " pod="openstack/swift-proxy-5488d7fd99-pwmf5" Dec 07 16:20:21 crc kubenswrapper[4716]: I1207 16:20:21.549439 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/826c0644-3046-453f-a139-5bd9c95216d0-run-httpd\") pod \"swift-proxy-5488d7fd99-pwmf5\" (UID: \"826c0644-3046-453f-a139-5bd9c95216d0\") " pod="openstack/swift-proxy-5488d7fd99-pwmf5" Dec 07 16:20:21 crc kubenswrapper[4716]: I1207 16:20:21.549954 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/826c0644-3046-453f-a139-5bd9c95216d0-run-httpd\") pod \"swift-proxy-5488d7fd99-pwmf5\" (UID: \"826c0644-3046-453f-a139-5bd9c95216d0\") " pod="openstack/swift-proxy-5488d7fd99-pwmf5" Dec 07 16:20:21 crc kubenswrapper[4716]: I1207 16:20:21.550519 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/826c0644-3046-453f-a139-5bd9c95216d0-log-httpd\") pod \"swift-proxy-5488d7fd99-pwmf5\" (UID: \"826c0644-3046-453f-a139-5bd9c95216d0\") " pod="openstack/swift-proxy-5488d7fd99-pwmf5" Dec 07 16:20:21 crc kubenswrapper[4716]: I1207 16:20:21.556375 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/826c0644-3046-453f-a139-5bd9c95216d0-combined-ca-bundle\") pod \"swift-proxy-5488d7fd99-pwmf5\" (UID: \"826c0644-3046-453f-a139-5bd9c95216d0\") " pod="openstack/swift-proxy-5488d7fd99-pwmf5" Dec 07 16:20:21 crc kubenswrapper[4716]: I1207 16:20:21.556538 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/826c0644-3046-453f-a139-5bd9c95216d0-config-data\") pod \"swift-proxy-5488d7fd99-pwmf5\" (UID: \"826c0644-3046-453f-a139-5bd9c95216d0\") " pod="openstack/swift-proxy-5488d7fd99-pwmf5" Dec 07 16:20:21 crc kubenswrapper[4716]: I1207 16:20:21.556715 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/826c0644-3046-453f-a139-5bd9c95216d0-public-tls-certs\") pod \"swift-proxy-5488d7fd99-pwmf5\" (UID: \"826c0644-3046-453f-a139-5bd9c95216d0\") " pod="openstack/swift-proxy-5488d7fd99-pwmf5" Dec 07 16:20:21 crc kubenswrapper[4716]: I1207 16:20:21.557271 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/826c0644-3046-453f-a139-5bd9c95216d0-etc-swift\") pod \"swift-proxy-5488d7fd99-pwmf5\" (UID: \"826c0644-3046-453f-a139-5bd9c95216d0\") " pod="openstack/swift-proxy-5488d7fd99-pwmf5" Dec 07 16:20:21 crc kubenswrapper[4716]: I1207 16:20:21.558341 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/826c0644-3046-453f-a139-5bd9c95216d0-internal-tls-certs\") pod \"swift-proxy-5488d7fd99-pwmf5\" (UID: \"826c0644-3046-453f-a139-5bd9c95216d0\") " pod="openstack/swift-proxy-5488d7fd99-pwmf5" Dec 07 16:20:21 crc kubenswrapper[4716]: I1207 16:20:21.567505 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hq99\" (UniqueName: \"kubernetes.io/projected/826c0644-3046-453f-a139-5bd9c95216d0-kube-api-access-4hq99\") pod \"swift-proxy-5488d7fd99-pwmf5\" (UID: \"826c0644-3046-453f-a139-5bd9c95216d0\") " pod="openstack/swift-proxy-5488d7fd99-pwmf5" Dec 07 16:20:21 crc kubenswrapper[4716]: I1207 16:20:21.602341 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-5488d7fd99-pwmf5" Dec 07 16:20:21 crc kubenswrapper[4716]: I1207 16:20:21.685066 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65aaa058-a1be-4477-b808-d0ea1d8e15fe" path="/var/lib/kubelet/pods/65aaa058-a1be-4477-b808-d0ea1d8e15fe/volumes" Dec 07 16:20:21 crc kubenswrapper[4716]: I1207 16:20:21.986757 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 07 16:20:21 crc kubenswrapper[4716]: I1207 16:20:21.987298 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0480133a-7588-4a60-921e-e53de8998ca9" containerName="ceilometer-central-agent" containerID="cri-o://610237798bb1fa0726f011172548d74003aeae66c492ec9ddc28726b612574eb" gracePeriod=30 Dec 07 16:20:21 crc kubenswrapper[4716]: I1207 16:20:21.989803 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0480133a-7588-4a60-921e-e53de8998ca9" containerName="proxy-httpd" containerID="cri-o://bbbcc700531f528352244efa205cfeeb3355cef62d7bf14bab368ee8dbe7b12a" gracePeriod=30 Dec 07 16:20:21 crc kubenswrapper[4716]: I1207 16:20:21.989852 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0480133a-7588-4a60-921e-e53de8998ca9" containerName="sg-core" containerID="cri-o://6ccfca76fd46f01f21a671fb4b52ac11b7e1cfd9deceaa24b6f22fff4d8d1881" gracePeriod=30 Dec 07 16:20:21 crc kubenswrapper[4716]: I1207 16:20:21.989945 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0480133a-7588-4a60-921e-e53de8998ca9" containerName="ceilometer-notification-agent" containerID="cri-o://9307786cb3b64f126e8b2107ac2b2c179b596785613c0be95dfc95f8b2b5f691" gracePeriod=30 Dec 07 16:20:21 crc kubenswrapper[4716]: I1207 16:20:21.993511 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 07 16:20:22 crc kubenswrapper[4716]: I1207 16:20:22.220673 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-5488d7fd99-pwmf5"] Dec 07 16:20:22 crc kubenswrapper[4716]: I1207 16:20:22.329147 4716 generic.go:334] "Generic (PLEG): container finished" podID="0480133a-7588-4a60-921e-e53de8998ca9" containerID="bbbcc700531f528352244efa205cfeeb3355cef62d7bf14bab368ee8dbe7b12a" exitCode=0 Dec 07 16:20:22 crc kubenswrapper[4716]: I1207 16:20:22.329482 4716 generic.go:334] "Generic (PLEG): container finished" podID="0480133a-7588-4a60-921e-e53de8998ca9" containerID="6ccfca76fd46f01f21a671fb4b52ac11b7e1cfd9deceaa24b6f22fff4d8d1881" exitCode=2 Dec 07 16:20:22 crc kubenswrapper[4716]: I1207 16:20:22.329232 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0480133a-7588-4a60-921e-e53de8998ca9","Type":"ContainerDied","Data":"bbbcc700531f528352244efa205cfeeb3355cef62d7bf14bab368ee8dbe7b12a"} Dec 07 16:20:22 crc kubenswrapper[4716]: I1207 16:20:22.329584 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0480133a-7588-4a60-921e-e53de8998ca9","Type":"ContainerDied","Data":"6ccfca76fd46f01f21a671fb4b52ac11b7e1cfd9deceaa24b6f22fff4d8d1881"} Dec 07 16:20:22 crc kubenswrapper[4716]: I1207 16:20:22.330934 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5488d7fd99-pwmf5" event={"ID":"826c0644-3046-453f-a139-5bd9c95216d0","Type":"ContainerStarted","Data":"242cb8253207361b79e7d1020fcf4ca9bd886f4d3430685cb0bfe93db3ded0e3"} Dec 07 16:20:23 crc kubenswrapper[4716]: I1207 16:20:23.348936 4716 generic.go:334] "Generic (PLEG): container finished" podID="0480133a-7588-4a60-921e-e53de8998ca9" containerID="9307786cb3b64f126e8b2107ac2b2c179b596785613c0be95dfc95f8b2b5f691" exitCode=0 Dec 07 16:20:23 crc kubenswrapper[4716]: I1207 16:20:23.348982 4716 generic.go:334] "Generic (PLEG): container finished" podID="0480133a-7588-4a60-921e-e53de8998ca9" containerID="610237798bb1fa0726f011172548d74003aeae66c492ec9ddc28726b612574eb" exitCode=0 Dec 07 16:20:23 crc kubenswrapper[4716]: I1207 16:20:23.349022 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0480133a-7588-4a60-921e-e53de8998ca9","Type":"ContainerDied","Data":"9307786cb3b64f126e8b2107ac2b2c179b596785613c0be95dfc95f8b2b5f691"} Dec 07 16:20:23 crc kubenswrapper[4716]: I1207 16:20:23.349091 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0480133a-7588-4a60-921e-e53de8998ca9","Type":"ContainerDied","Data":"610237798bb1fa0726f011172548d74003aeae66c492ec9ddc28726b612574eb"} Dec 07 16:20:23 crc kubenswrapper[4716]: I1207 16:20:23.350162 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5488d7fd99-pwmf5" event={"ID":"826c0644-3046-453f-a139-5bd9c95216d0","Type":"ContainerStarted","Data":"4fe329f83059af3a761dc94572da16e20d0c43659a749ea3e408dd02c7e3b051"} Dec 07 16:20:23 crc kubenswrapper[4716]: I1207 16:20:23.934206 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 07 16:20:23 crc kubenswrapper[4716]: I1207 16:20:23.934731 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="5eca408c-c86b-4dd6-a1f9-81ef16e12da4" containerName="glance-log" containerID="cri-o://887f5cff319603ea4d4794148d1a5340881877c3bc68e5d25f280bb442fe7459" gracePeriod=30 Dec 07 16:20:23 crc kubenswrapper[4716]: I1207 16:20:23.935168 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="5eca408c-c86b-4dd6-a1f9-81ef16e12da4" containerName="glance-httpd" containerID="cri-o://7a9e2b9815207ee0080821a823fef456aa90e0a1cfa6b17abc1d59687be33655" gracePeriod=30 Dec 07 16:20:24 crc kubenswrapper[4716]: I1207 16:20:24.359925 4716 generic.go:334] "Generic (PLEG): container finished" podID="5eca408c-c86b-4dd6-a1f9-81ef16e12da4" containerID="887f5cff319603ea4d4794148d1a5340881877c3bc68e5d25f280bb442fe7459" exitCode=143 Dec 07 16:20:24 crc kubenswrapper[4716]: I1207 16:20:24.359976 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5eca408c-c86b-4dd6-a1f9-81ef16e12da4","Type":"ContainerDied","Data":"887f5cff319603ea4d4794148d1a5340881877c3bc68e5d25f280bb442fe7459"} Dec 07 16:20:24 crc kubenswrapper[4716]: I1207 16:20:24.984908 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 07 16:20:24 crc kubenswrapper[4716]: I1207 16:20:24.985209 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="afba3ceb-86bf-4aaf-8ac3-b58ab411850c" containerName="glance-log" containerID="cri-o://7e62702f6546007c2a5d5db6138318f092dfcf6f6759db6ed04b36ce56729518" gracePeriod=30 Dec 07 16:20:24 crc kubenswrapper[4716]: I1207 16:20:24.985357 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="afba3ceb-86bf-4aaf-8ac3-b58ab411850c" containerName="glance-httpd" containerID="cri-o://9d342a91e7dc00c4eac8190839f9ea8a4d0e4c05fb04b03355f7cb519bf04840" gracePeriod=30 Dec 07 16:20:25 crc kubenswrapper[4716]: I1207 16:20:25.374470 4716 generic.go:334] "Generic (PLEG): container finished" podID="afba3ceb-86bf-4aaf-8ac3-b58ab411850c" containerID="7e62702f6546007c2a5d5db6138318f092dfcf6f6759db6ed04b36ce56729518" exitCode=143 Dec 07 16:20:25 crc kubenswrapper[4716]: I1207 16:20:25.374571 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"afba3ceb-86bf-4aaf-8ac3-b58ab411850c","Type":"ContainerDied","Data":"7e62702f6546007c2a5d5db6138318f092dfcf6f6759db6ed04b36ce56729518"} Dec 07 16:20:26 crc kubenswrapper[4716]: I1207 16:20:26.143970 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 07 16:20:27 crc kubenswrapper[4716]: I1207 16:20:27.399164 4716 generic.go:334] "Generic (PLEG): container finished" podID="5eca408c-c86b-4dd6-a1f9-81ef16e12da4" containerID="7a9e2b9815207ee0080821a823fef456aa90e0a1cfa6b17abc1d59687be33655" exitCode=0 Dec 07 16:20:27 crc kubenswrapper[4716]: I1207 16:20:27.399422 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5eca408c-c86b-4dd6-a1f9-81ef16e12da4","Type":"ContainerDied","Data":"7a9e2b9815207ee0080821a823fef456aa90e0a1cfa6b17abc1d59687be33655"} Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.312669 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.363253 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.412935 4716 generic.go:334] "Generic (PLEG): container finished" podID="afba3ceb-86bf-4aaf-8ac3-b58ab411850c" containerID="9d342a91e7dc00c4eac8190839f9ea8a4d0e4c05fb04b03355f7cb519bf04840" exitCode=0 Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.413012 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"afba3ceb-86bf-4aaf-8ac3-b58ab411850c","Type":"ContainerDied","Data":"9d342a91e7dc00c4eac8190839f9ea8a4d0e4c05fb04b03355f7cb519bf04840"} Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.415840 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5eca408c-c86b-4dd6-a1f9-81ef16e12da4","Type":"ContainerDied","Data":"74f894d324ce5d8ac07b7d951146a82fe1f7fc35ef5ec12b146942ecc20cd47a"} Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.415884 4716 scope.go:117] "RemoveContainer" containerID="7a9e2b9815207ee0080821a823fef456aa90e0a1cfa6b17abc1d59687be33655" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.415995 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.416496 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"5eca408c-c86b-4dd6-a1f9-81ef16e12da4\" (UID: \"5eca408c-c86b-4dd6-a1f9-81ef16e12da4\") " Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.416563 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0480133a-7588-4a60-921e-e53de8998ca9-scripts\") pod \"0480133a-7588-4a60-921e-e53de8998ca9\" (UID: \"0480133a-7588-4a60-921e-e53de8998ca9\") " Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.416605 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hnll7\" (UniqueName: \"kubernetes.io/projected/5eca408c-c86b-4dd6-a1f9-81ef16e12da4-kube-api-access-hnll7\") pod \"5eca408c-c86b-4dd6-a1f9-81ef16e12da4\" (UID: \"5eca408c-c86b-4dd6-a1f9-81ef16e12da4\") " Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.416670 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0480133a-7588-4a60-921e-e53de8998ca9-run-httpd\") pod \"0480133a-7588-4a60-921e-e53de8998ca9\" (UID: \"0480133a-7588-4a60-921e-e53de8998ca9\") " Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.416695 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5eca408c-c86b-4dd6-a1f9-81ef16e12da4-combined-ca-bundle\") pod \"5eca408c-c86b-4dd6-a1f9-81ef16e12da4\" (UID: \"5eca408c-c86b-4dd6-a1f9-81ef16e12da4\") " Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.416755 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5eca408c-c86b-4dd6-a1f9-81ef16e12da4-internal-tls-certs\") pod \"5eca408c-c86b-4dd6-a1f9-81ef16e12da4\" (UID: \"5eca408c-c86b-4dd6-a1f9-81ef16e12da4\") " Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.416773 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0480133a-7588-4a60-921e-e53de8998ca9-config-data\") pod \"0480133a-7588-4a60-921e-e53de8998ca9\" (UID: \"0480133a-7588-4a60-921e-e53de8998ca9\") " Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.416800 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5eca408c-c86b-4dd6-a1f9-81ef16e12da4-logs\") pod \"5eca408c-c86b-4dd6-a1f9-81ef16e12da4\" (UID: \"5eca408c-c86b-4dd6-a1f9-81ef16e12da4\") " Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.416861 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5eca408c-c86b-4dd6-a1f9-81ef16e12da4-scripts\") pod \"5eca408c-c86b-4dd6-a1f9-81ef16e12da4\" (UID: \"5eca408c-c86b-4dd6-a1f9-81ef16e12da4\") " Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.416899 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0480133a-7588-4a60-921e-e53de8998ca9-log-httpd\") pod \"0480133a-7588-4a60-921e-e53de8998ca9\" (UID: \"0480133a-7588-4a60-921e-e53de8998ca9\") " Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.416916 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0480133a-7588-4a60-921e-e53de8998ca9-sg-core-conf-yaml\") pod \"0480133a-7588-4a60-921e-e53de8998ca9\" (UID: \"0480133a-7588-4a60-921e-e53de8998ca9\") " Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.416951 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5eca408c-c86b-4dd6-a1f9-81ef16e12da4-config-data\") pod \"5eca408c-c86b-4dd6-a1f9-81ef16e12da4\" (UID: \"5eca408c-c86b-4dd6-a1f9-81ef16e12da4\") " Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.416973 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0480133a-7588-4a60-921e-e53de8998ca9-combined-ca-bundle\") pod \"0480133a-7588-4a60-921e-e53de8998ca9\" (UID: \"0480133a-7588-4a60-921e-e53de8998ca9\") " Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.416995 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ltfdw\" (UniqueName: \"kubernetes.io/projected/0480133a-7588-4a60-921e-e53de8998ca9-kube-api-access-ltfdw\") pod \"0480133a-7588-4a60-921e-e53de8998ca9\" (UID: \"0480133a-7588-4a60-921e-e53de8998ca9\") " Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.417032 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5eca408c-c86b-4dd6-a1f9-81ef16e12da4-httpd-run\") pod \"5eca408c-c86b-4dd6-a1f9-81ef16e12da4\" (UID: \"5eca408c-c86b-4dd6-a1f9-81ef16e12da4\") " Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.417130 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0480133a-7588-4a60-921e-e53de8998ca9-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "0480133a-7588-4a60-921e-e53de8998ca9" (UID: "0480133a-7588-4a60-921e-e53de8998ca9"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.417368 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5eca408c-c86b-4dd6-a1f9-81ef16e12da4-logs" (OuterVolumeSpecName: "logs") pod "5eca408c-c86b-4dd6-a1f9-81ef16e12da4" (UID: "5eca408c-c86b-4dd6-a1f9-81ef16e12da4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.417439 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0480133a-7588-4a60-921e-e53de8998ca9-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "0480133a-7588-4a60-921e-e53de8998ca9" (UID: "0480133a-7588-4a60-921e-e53de8998ca9"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.417662 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5eca408c-c86b-4dd6-a1f9-81ef16e12da4-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "5eca408c-c86b-4dd6-a1f9-81ef16e12da4" (UID: "5eca408c-c86b-4dd6-a1f9-81ef16e12da4"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.417737 4716 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0480133a-7588-4a60-921e-e53de8998ca9-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.417755 4716 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5eca408c-c86b-4dd6-a1f9-81ef16e12da4-logs\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.417764 4716 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0480133a-7588-4a60-921e-e53de8998ca9-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.426800 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5488d7fd99-pwmf5" event={"ID":"826c0644-3046-453f-a139-5bd9c95216d0","Type":"ContainerStarted","Data":"f55c738d362ac9d660299e998c40e190d8ec9c41771e651722625184ecb57d08"} Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.427308 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "5eca408c-c86b-4dd6-a1f9-81ef16e12da4" (UID: "5eca408c-c86b-4dd6-a1f9-81ef16e12da4"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.428385 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0480133a-7588-4a60-921e-e53de8998ca9-scripts" (OuterVolumeSpecName: "scripts") pod "0480133a-7588-4a60-921e-e53de8998ca9" (UID: "0480133a-7588-4a60-921e-e53de8998ca9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.428501 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-5488d7fd99-pwmf5" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.428525 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-5488d7fd99-pwmf5" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.430712 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0480133a-7588-4a60-921e-e53de8998ca9-kube-api-access-ltfdw" (OuterVolumeSpecName: "kube-api-access-ltfdw") pod "0480133a-7588-4a60-921e-e53de8998ca9" (UID: "0480133a-7588-4a60-921e-e53de8998ca9"). InnerVolumeSpecName "kube-api-access-ltfdw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.432153 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"6a6ecf77-dd60-497a-8670-919d23db95dd","Type":"ContainerStarted","Data":"84a98a040c47fc7dc5bbf205ee6ec33d0c52409999d151cce43b2132bbd05a37"} Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.436645 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0480133a-7588-4a60-921e-e53de8998ca9","Type":"ContainerDied","Data":"e042f3a14b4b726e2937b7c1edb5cf38b368164ff2e4f333aa761c8edd18ee9b"} Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.436725 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.442055 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5eca408c-c86b-4dd6-a1f9-81ef16e12da4-kube-api-access-hnll7" (OuterVolumeSpecName: "kube-api-access-hnll7") pod "5eca408c-c86b-4dd6-a1f9-81ef16e12da4" (UID: "5eca408c-c86b-4dd6-a1f9-81ef16e12da4"). InnerVolumeSpecName "kube-api-access-hnll7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.442339 4716 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-5488d7fd99-pwmf5" podUID="826c0644-3046-453f-a139-5bd9c95216d0" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.453775 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5eca408c-c86b-4dd6-a1f9-81ef16e12da4-scripts" (OuterVolumeSpecName: "scripts") pod "5eca408c-c86b-4dd6-a1f9-81ef16e12da4" (UID: "5eca408c-c86b-4dd6-a1f9-81ef16e12da4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.466456 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-5488d7fd99-pwmf5" podStartSLOduration=7.463327444 podStartE2EDuration="7.463327444s" podCreationTimestamp="2025-12-07 16:20:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:20:28.450876586 +0000 UTC m=+1091.141161498" watchObservedRunningTime="2025-12-07 16:20:28.463327444 +0000 UTC m=+1091.153612356" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.477889 4716 scope.go:117] "RemoveContainer" containerID="887f5cff319603ea4d4794148d1a5340881877c3bc68e5d25f280bb442fe7459" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.484115 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5eca408c-c86b-4dd6-a1f9-81ef16e12da4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5eca408c-c86b-4dd6-a1f9-81ef16e12da4" (UID: "5eca408c-c86b-4dd6-a1f9-81ef16e12da4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.493538 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0480133a-7588-4a60-921e-e53de8998ca9-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "0480133a-7588-4a60-921e-e53de8998ca9" (UID: "0480133a-7588-4a60-921e-e53de8998ca9"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.495234 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.3325732869999998 podStartE2EDuration="12.495213507s" podCreationTimestamp="2025-12-07 16:20:16 +0000 UTC" firstStartedPulling="2025-12-07 16:20:17.837243188 +0000 UTC m=+1080.527528090" lastFinishedPulling="2025-12-07 16:20:27.999883408 +0000 UTC m=+1090.690168310" observedRunningTime="2025-12-07 16:20:28.488268249 +0000 UTC m=+1091.178553181" watchObservedRunningTime="2025-12-07 16:20:28.495213507 +0000 UTC m=+1091.185498419" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.507710 4716 scope.go:117] "RemoveContainer" containerID="bbbcc700531f528352244efa205cfeeb3355cef62d7bf14bab368ee8dbe7b12a" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.523811 4716 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0480133a-7588-4a60-921e-e53de8998ca9-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.523846 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ltfdw\" (UniqueName: \"kubernetes.io/projected/0480133a-7588-4a60-921e-e53de8998ca9-kube-api-access-ltfdw\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.523861 4716 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5eca408c-c86b-4dd6-a1f9-81ef16e12da4-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.523887 4716 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.523899 4716 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0480133a-7588-4a60-921e-e53de8998ca9-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.523914 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hnll7\" (UniqueName: \"kubernetes.io/projected/5eca408c-c86b-4dd6-a1f9-81ef16e12da4-kube-api-access-hnll7\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.523925 4716 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5eca408c-c86b-4dd6-a1f9-81ef16e12da4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.523935 4716 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5eca408c-c86b-4dd6-a1f9-81ef16e12da4-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:28 crc kubenswrapper[4716]: E1207 16:20:28.526226 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5eca408c-c86b-4dd6-a1f9-81ef16e12da4-internal-tls-certs podName:5eca408c-c86b-4dd6-a1f9-81ef16e12da4 nodeName:}" failed. No retries permitted until 2025-12-07 16:20:29.025948119 +0000 UTC m=+1091.716233031 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "internal-tls-certs" (UniqueName: "kubernetes.io/secret/5eca408c-c86b-4dd6-a1f9-81ef16e12da4-internal-tls-certs") pod "5eca408c-c86b-4dd6-a1f9-81ef16e12da4" (UID: "5eca408c-c86b-4dd6-a1f9-81ef16e12da4") : error deleting /var/lib/kubelet/pods/5eca408c-c86b-4dd6-a1f9-81ef16e12da4/volume-subpaths: remove /var/lib/kubelet/pods/5eca408c-c86b-4dd6-a1f9-81ef16e12da4/volume-subpaths: no such file or directory Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.528886 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5eca408c-c86b-4dd6-a1f9-81ef16e12da4-config-data" (OuterVolumeSpecName: "config-data") pod "5eca408c-c86b-4dd6-a1f9-81ef16e12da4" (UID: "5eca408c-c86b-4dd6-a1f9-81ef16e12da4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.530232 4716 scope.go:117] "RemoveContainer" containerID="6ccfca76fd46f01f21a671fb4b52ac11b7e1cfd9deceaa24b6f22fff4d8d1881" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.549634 4716 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.563668 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0480133a-7588-4a60-921e-e53de8998ca9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0480133a-7588-4a60-921e-e53de8998ca9" (UID: "0480133a-7588-4a60-921e-e53de8998ca9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.564772 4716 scope.go:117] "RemoveContainer" containerID="9307786cb3b64f126e8b2107ac2b2c179b596785613c0be95dfc95f8b2b5f691" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.577931 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0480133a-7588-4a60-921e-e53de8998ca9-config-data" (OuterVolumeSpecName: "config-data") pod "0480133a-7588-4a60-921e-e53de8998ca9" (UID: "0480133a-7588-4a60-921e-e53de8998ca9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.593663 4716 scope.go:117] "RemoveContainer" containerID="610237798bb1fa0726f011172548d74003aeae66c492ec9ddc28726b612574eb" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.625953 4716 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5eca408c-c86b-4dd6-a1f9-81ef16e12da4-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.625988 4716 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0480133a-7588-4a60-921e-e53de8998ca9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.626197 4716 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.626212 4716 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0480133a-7588-4a60-921e-e53de8998ca9-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.669842 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.732490 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gcgn6\" (UniqueName: \"kubernetes.io/projected/afba3ceb-86bf-4aaf-8ac3-b58ab411850c-kube-api-access-gcgn6\") pod \"afba3ceb-86bf-4aaf-8ac3-b58ab411850c\" (UID: \"afba3ceb-86bf-4aaf-8ac3-b58ab411850c\") " Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.732552 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/afba3ceb-86bf-4aaf-8ac3-b58ab411850c-public-tls-certs\") pod \"afba3ceb-86bf-4aaf-8ac3-b58ab411850c\" (UID: \"afba3ceb-86bf-4aaf-8ac3-b58ab411850c\") " Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.732621 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/afba3ceb-86bf-4aaf-8ac3-b58ab411850c-config-data\") pod \"afba3ceb-86bf-4aaf-8ac3-b58ab411850c\" (UID: \"afba3ceb-86bf-4aaf-8ac3-b58ab411850c\") " Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.732727 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"afba3ceb-86bf-4aaf-8ac3-b58ab411850c\" (UID: \"afba3ceb-86bf-4aaf-8ac3-b58ab411850c\") " Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.732767 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/afba3ceb-86bf-4aaf-8ac3-b58ab411850c-scripts\") pod \"afba3ceb-86bf-4aaf-8ac3-b58ab411850c\" (UID: \"afba3ceb-86bf-4aaf-8ac3-b58ab411850c\") " Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.732884 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afba3ceb-86bf-4aaf-8ac3-b58ab411850c-combined-ca-bundle\") pod \"afba3ceb-86bf-4aaf-8ac3-b58ab411850c\" (UID: \"afba3ceb-86bf-4aaf-8ac3-b58ab411850c\") " Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.732912 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/afba3ceb-86bf-4aaf-8ac3-b58ab411850c-logs\") pod \"afba3ceb-86bf-4aaf-8ac3-b58ab411850c\" (UID: \"afba3ceb-86bf-4aaf-8ac3-b58ab411850c\") " Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.736451 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/afba3ceb-86bf-4aaf-8ac3-b58ab411850c-logs" (OuterVolumeSpecName: "logs") pod "afba3ceb-86bf-4aaf-8ac3-b58ab411850c" (UID: "afba3ceb-86bf-4aaf-8ac3-b58ab411850c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.742596 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/afba3ceb-86bf-4aaf-8ac3-b58ab411850c-scripts" (OuterVolumeSpecName: "scripts") pod "afba3ceb-86bf-4aaf-8ac3-b58ab411850c" (UID: "afba3ceb-86bf-4aaf-8ac3-b58ab411850c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.746562 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/afba3ceb-86bf-4aaf-8ac3-b58ab411850c-kube-api-access-gcgn6" (OuterVolumeSpecName: "kube-api-access-gcgn6") pod "afba3ceb-86bf-4aaf-8ac3-b58ab411850c" (UID: "afba3ceb-86bf-4aaf-8ac3-b58ab411850c"). InnerVolumeSpecName "kube-api-access-gcgn6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.776697 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance") pod "afba3ceb-86bf-4aaf-8ac3-b58ab411850c" (UID: "afba3ceb-86bf-4aaf-8ac3-b58ab411850c"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.793866 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/afba3ceb-86bf-4aaf-8ac3-b58ab411850c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "afba3ceb-86bf-4aaf-8ac3-b58ab411850c" (UID: "afba3ceb-86bf-4aaf-8ac3-b58ab411850c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.835719 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.838613 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/afba3ceb-86bf-4aaf-8ac3-b58ab411850c-httpd-run\") pod \"afba3ceb-86bf-4aaf-8ac3-b58ab411850c\" (UID: \"afba3ceb-86bf-4aaf-8ac3-b58ab411850c\") " Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.845961 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/afba3ceb-86bf-4aaf-8ac3-b58ab411850c-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "afba3ceb-86bf-4aaf-8ac3-b58ab411850c" (UID: "afba3ceb-86bf-4aaf-8ac3-b58ab411850c"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.851194 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.857415 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gcgn6\" (UniqueName: \"kubernetes.io/projected/afba3ceb-86bf-4aaf-8ac3-b58ab411850c-kube-api-access-gcgn6\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.857499 4716 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.857510 4716 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/afba3ceb-86bf-4aaf-8ac3-b58ab411850c-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.857521 4716 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afba3ceb-86bf-4aaf-8ac3-b58ab411850c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.857531 4716 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/afba3ceb-86bf-4aaf-8ac3-b58ab411850c-logs\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.857539 4716 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/afba3ceb-86bf-4aaf-8ac3-b58ab411850c-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.889271 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 07 16:20:28 crc kubenswrapper[4716]: E1207 16:20:28.890264 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5eca408c-c86b-4dd6-a1f9-81ef16e12da4" containerName="glance-httpd" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.890286 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="5eca408c-c86b-4dd6-a1f9-81ef16e12da4" containerName="glance-httpd" Dec 07 16:20:28 crc kubenswrapper[4716]: E1207 16:20:28.890304 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0480133a-7588-4a60-921e-e53de8998ca9" containerName="ceilometer-notification-agent" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.890310 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="0480133a-7588-4a60-921e-e53de8998ca9" containerName="ceilometer-notification-agent" Dec 07 16:20:28 crc kubenswrapper[4716]: E1207 16:20:28.890322 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5eca408c-c86b-4dd6-a1f9-81ef16e12da4" containerName="glance-log" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.890329 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="5eca408c-c86b-4dd6-a1f9-81ef16e12da4" containerName="glance-log" Dec 07 16:20:28 crc kubenswrapper[4716]: E1207 16:20:28.890340 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0480133a-7588-4a60-921e-e53de8998ca9" containerName="ceilometer-central-agent" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.890347 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="0480133a-7588-4a60-921e-e53de8998ca9" containerName="ceilometer-central-agent" Dec 07 16:20:28 crc kubenswrapper[4716]: E1207 16:20:28.890360 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afba3ceb-86bf-4aaf-8ac3-b58ab411850c" containerName="glance-log" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.890366 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="afba3ceb-86bf-4aaf-8ac3-b58ab411850c" containerName="glance-log" Dec 07 16:20:28 crc kubenswrapper[4716]: E1207 16:20:28.890379 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0480133a-7588-4a60-921e-e53de8998ca9" containerName="proxy-httpd" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.890385 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="0480133a-7588-4a60-921e-e53de8998ca9" containerName="proxy-httpd" Dec 07 16:20:28 crc kubenswrapper[4716]: E1207 16:20:28.890405 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0480133a-7588-4a60-921e-e53de8998ca9" containerName="sg-core" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.890410 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="0480133a-7588-4a60-921e-e53de8998ca9" containerName="sg-core" Dec 07 16:20:28 crc kubenswrapper[4716]: E1207 16:20:28.890420 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afba3ceb-86bf-4aaf-8ac3-b58ab411850c" containerName="glance-httpd" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.890426 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="afba3ceb-86bf-4aaf-8ac3-b58ab411850c" containerName="glance-httpd" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.890590 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="0480133a-7588-4a60-921e-e53de8998ca9" containerName="proxy-httpd" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.890606 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="afba3ceb-86bf-4aaf-8ac3-b58ab411850c" containerName="glance-httpd" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.890623 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="5eca408c-c86b-4dd6-a1f9-81ef16e12da4" containerName="glance-log" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.890637 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="afba3ceb-86bf-4aaf-8ac3-b58ab411850c" containerName="glance-log" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.890648 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="0480133a-7588-4a60-921e-e53de8998ca9" containerName="ceilometer-central-agent" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.890662 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="5eca408c-c86b-4dd6-a1f9-81ef16e12da4" containerName="glance-httpd" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.890675 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="0480133a-7588-4a60-921e-e53de8998ca9" containerName="ceilometer-notification-agent" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.890685 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="0480133a-7588-4a60-921e-e53de8998ca9" containerName="sg-core" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.894005 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.903870 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.904172 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.904266 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.914111 4716 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.926143 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-pz462"] Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.927456 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-pz462" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.950764 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-pz462"] Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.958775 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6873cce2-e68e-47ca-8ae8-7b612815a223-run-httpd\") pod \"ceilometer-0\" (UID: \"6873cce2-e68e-47ca-8ae8-7b612815a223\") " pod="openstack/ceilometer-0" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.958843 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6873cce2-e68e-47ca-8ae8-7b612815a223-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6873cce2-e68e-47ca-8ae8-7b612815a223\") " pod="openstack/ceilometer-0" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.958863 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6873cce2-e68e-47ca-8ae8-7b612815a223-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6873cce2-e68e-47ca-8ae8-7b612815a223\") " pod="openstack/ceilometer-0" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.958878 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vlj99\" (UniqueName: \"kubernetes.io/projected/6873cce2-e68e-47ca-8ae8-7b612815a223-kube-api-access-vlj99\") pod \"ceilometer-0\" (UID: \"6873cce2-e68e-47ca-8ae8-7b612815a223\") " pod="openstack/ceilometer-0" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.958898 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6873cce2-e68e-47ca-8ae8-7b612815a223-config-data\") pod \"ceilometer-0\" (UID: \"6873cce2-e68e-47ca-8ae8-7b612815a223\") " pod="openstack/ceilometer-0" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.958933 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6873cce2-e68e-47ca-8ae8-7b612815a223-log-httpd\") pod \"ceilometer-0\" (UID: \"6873cce2-e68e-47ca-8ae8-7b612815a223\") " pod="openstack/ceilometer-0" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.958953 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6873cce2-e68e-47ca-8ae8-7b612815a223-scripts\") pod \"ceilometer-0\" (UID: \"6873cce2-e68e-47ca-8ae8-7b612815a223\") " pod="openstack/ceilometer-0" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.958971 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bwbgs\" (UniqueName: \"kubernetes.io/projected/a14d4b2b-4e79-4258-9d9f-46198eca8a41-kube-api-access-bwbgs\") pod \"nova-api-db-create-pz462\" (UID: \"a14d4b2b-4e79-4258-9d9f-46198eca8a41\") " pod="openstack/nova-api-db-create-pz462" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.959020 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a14d4b2b-4e79-4258-9d9f-46198eca8a41-operator-scripts\") pod \"nova-api-db-create-pz462\" (UID: \"a14d4b2b-4e79-4258-9d9f-46198eca8a41\") " pod="openstack/nova-api-db-create-pz462" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.959094 4716 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.980166 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/afba3ceb-86bf-4aaf-8ac3-b58ab411850c-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "afba3ceb-86bf-4aaf-8ac3-b58ab411850c" (UID: "afba3ceb-86bf-4aaf-8ac3-b58ab411850c"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.983163 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/afba3ceb-86bf-4aaf-8ac3-b58ab411850c-config-data" (OuterVolumeSpecName: "config-data") pod "afba3ceb-86bf-4aaf-8ac3-b58ab411850c" (UID: "afba3ceb-86bf-4aaf-8ac3-b58ab411850c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.985163 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-xbk7g"] Dec 07 16:20:28 crc kubenswrapper[4716]: I1207 16:20:28.986270 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-xbk7g" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.003373 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-c495-account-create-update-cpqvg"] Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.004536 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-c495-account-create-update-cpqvg" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.006412 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.012854 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-xbk7g"] Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.025913 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-c495-account-create-update-cpqvg"] Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.061037 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5eca408c-c86b-4dd6-a1f9-81ef16e12da4-internal-tls-certs\") pod \"5eca408c-c86b-4dd6-a1f9-81ef16e12da4\" (UID: \"5eca408c-c86b-4dd6-a1f9-81ef16e12da4\") " Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.061345 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a14d4b2b-4e79-4258-9d9f-46198eca8a41-operator-scripts\") pod \"nova-api-db-create-pz462\" (UID: \"a14d4b2b-4e79-4258-9d9f-46198eca8a41\") " pod="openstack/nova-api-db-create-pz462" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.061384 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ead59b4-dcff-45b8-8b1b-d5e449278acb-operator-scripts\") pod \"nova-cell0-db-create-xbk7g\" (UID: \"7ead59b4-dcff-45b8-8b1b-d5e449278acb\") " pod="openstack/nova-cell0-db-create-xbk7g" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.061465 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6873cce2-e68e-47ca-8ae8-7b612815a223-run-httpd\") pod \"ceilometer-0\" (UID: \"6873cce2-e68e-47ca-8ae8-7b612815a223\") " pod="openstack/ceilometer-0" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.061485 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/73795e56-2afe-427c-a122-7d604c6ed134-operator-scripts\") pod \"nova-api-c495-account-create-update-cpqvg\" (UID: \"73795e56-2afe-427c-a122-7d604c6ed134\") " pod="openstack/nova-api-c495-account-create-update-cpqvg" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.061521 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjw5h\" (UniqueName: \"kubernetes.io/projected/73795e56-2afe-427c-a122-7d604c6ed134-kube-api-access-vjw5h\") pod \"nova-api-c495-account-create-update-cpqvg\" (UID: \"73795e56-2afe-427c-a122-7d604c6ed134\") " pod="openstack/nova-api-c495-account-create-update-cpqvg" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.061544 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nxjkl\" (UniqueName: \"kubernetes.io/projected/7ead59b4-dcff-45b8-8b1b-d5e449278acb-kube-api-access-nxjkl\") pod \"nova-cell0-db-create-xbk7g\" (UID: \"7ead59b4-dcff-45b8-8b1b-d5e449278acb\") " pod="openstack/nova-cell0-db-create-xbk7g" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.061561 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6873cce2-e68e-47ca-8ae8-7b612815a223-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6873cce2-e68e-47ca-8ae8-7b612815a223\") " pod="openstack/ceilometer-0" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.061578 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6873cce2-e68e-47ca-8ae8-7b612815a223-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6873cce2-e68e-47ca-8ae8-7b612815a223\") " pod="openstack/ceilometer-0" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.061595 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vlj99\" (UniqueName: \"kubernetes.io/projected/6873cce2-e68e-47ca-8ae8-7b612815a223-kube-api-access-vlj99\") pod \"ceilometer-0\" (UID: \"6873cce2-e68e-47ca-8ae8-7b612815a223\") " pod="openstack/ceilometer-0" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.061616 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6873cce2-e68e-47ca-8ae8-7b612815a223-config-data\") pod \"ceilometer-0\" (UID: \"6873cce2-e68e-47ca-8ae8-7b612815a223\") " pod="openstack/ceilometer-0" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.061753 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6873cce2-e68e-47ca-8ae8-7b612815a223-log-httpd\") pod \"ceilometer-0\" (UID: \"6873cce2-e68e-47ca-8ae8-7b612815a223\") " pod="openstack/ceilometer-0" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.061813 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6873cce2-e68e-47ca-8ae8-7b612815a223-scripts\") pod \"ceilometer-0\" (UID: \"6873cce2-e68e-47ca-8ae8-7b612815a223\") " pod="openstack/ceilometer-0" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.061847 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bwbgs\" (UniqueName: \"kubernetes.io/projected/a14d4b2b-4e79-4258-9d9f-46198eca8a41-kube-api-access-bwbgs\") pod \"nova-api-db-create-pz462\" (UID: \"a14d4b2b-4e79-4258-9d9f-46198eca8a41\") " pod="openstack/nova-api-db-create-pz462" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.061932 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6873cce2-e68e-47ca-8ae8-7b612815a223-run-httpd\") pod \"ceilometer-0\" (UID: \"6873cce2-e68e-47ca-8ae8-7b612815a223\") " pod="openstack/ceilometer-0" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.062102 4716 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/afba3ceb-86bf-4aaf-8ac3-b58ab411850c-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.062464 4716 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/afba3ceb-86bf-4aaf-8ac3-b58ab411850c-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.063355 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6873cce2-e68e-47ca-8ae8-7b612815a223-log-httpd\") pod \"ceilometer-0\" (UID: \"6873cce2-e68e-47ca-8ae8-7b612815a223\") " pod="openstack/ceilometer-0" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.064112 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a14d4b2b-4e79-4258-9d9f-46198eca8a41-operator-scripts\") pod \"nova-api-db-create-pz462\" (UID: \"a14d4b2b-4e79-4258-9d9f-46198eca8a41\") " pod="openstack/nova-api-db-create-pz462" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.066230 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5eca408c-c86b-4dd6-a1f9-81ef16e12da4-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "5eca408c-c86b-4dd6-a1f9-81ef16e12da4" (UID: "5eca408c-c86b-4dd6-a1f9-81ef16e12da4"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.067617 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6873cce2-e68e-47ca-8ae8-7b612815a223-scripts\") pod \"ceilometer-0\" (UID: \"6873cce2-e68e-47ca-8ae8-7b612815a223\") " pod="openstack/ceilometer-0" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.074340 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6873cce2-e68e-47ca-8ae8-7b612815a223-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6873cce2-e68e-47ca-8ae8-7b612815a223\") " pod="openstack/ceilometer-0" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.076968 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6873cce2-e68e-47ca-8ae8-7b612815a223-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6873cce2-e68e-47ca-8ae8-7b612815a223\") " pod="openstack/ceilometer-0" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.078502 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6873cce2-e68e-47ca-8ae8-7b612815a223-config-data\") pod \"ceilometer-0\" (UID: \"6873cce2-e68e-47ca-8ae8-7b612815a223\") " pod="openstack/ceilometer-0" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.079383 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vlj99\" (UniqueName: \"kubernetes.io/projected/6873cce2-e68e-47ca-8ae8-7b612815a223-kube-api-access-vlj99\") pod \"ceilometer-0\" (UID: \"6873cce2-e68e-47ca-8ae8-7b612815a223\") " pod="openstack/ceilometer-0" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.079629 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bwbgs\" (UniqueName: \"kubernetes.io/projected/a14d4b2b-4e79-4258-9d9f-46198eca8a41-kube-api-access-bwbgs\") pod \"nova-api-db-create-pz462\" (UID: \"a14d4b2b-4e79-4258-9d9f-46198eca8a41\") " pod="openstack/nova-api-db-create-pz462" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.156829 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-d9hjx"] Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.158129 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-d9hjx" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.164905 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ead59b4-dcff-45b8-8b1b-d5e449278acb-operator-scripts\") pod \"nova-cell0-db-create-xbk7g\" (UID: \"7ead59b4-dcff-45b8-8b1b-d5e449278acb\") " pod="openstack/nova-cell0-db-create-xbk7g" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.172935 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ead59b4-dcff-45b8-8b1b-d5e449278acb-operator-scripts\") pod \"nova-cell0-db-create-xbk7g\" (UID: \"7ead59b4-dcff-45b8-8b1b-d5e449278acb\") " pod="openstack/nova-cell0-db-create-xbk7g" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.173054 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/73795e56-2afe-427c-a122-7d604c6ed134-operator-scripts\") pod \"nova-api-c495-account-create-update-cpqvg\" (UID: \"73795e56-2afe-427c-a122-7d604c6ed134\") " pod="openstack/nova-api-c495-account-create-update-cpqvg" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.173139 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjw5h\" (UniqueName: \"kubernetes.io/projected/73795e56-2afe-427c-a122-7d604c6ed134-kube-api-access-vjw5h\") pod \"nova-api-c495-account-create-update-cpqvg\" (UID: \"73795e56-2afe-427c-a122-7d604c6ed134\") " pod="openstack/nova-api-c495-account-create-update-cpqvg" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.173179 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nxjkl\" (UniqueName: \"kubernetes.io/projected/7ead59b4-dcff-45b8-8b1b-d5e449278acb-kube-api-access-nxjkl\") pod \"nova-cell0-db-create-xbk7g\" (UID: \"7ead59b4-dcff-45b8-8b1b-d5e449278acb\") " pod="openstack/nova-cell0-db-create-xbk7g" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.173356 4716 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5eca408c-c86b-4dd6-a1f9-81ef16e12da4-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.174109 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/73795e56-2afe-427c-a122-7d604c6ed134-operator-scripts\") pod \"nova-api-c495-account-create-update-cpqvg\" (UID: \"73795e56-2afe-427c-a122-7d604c6ed134\") " pod="openstack/nova-api-c495-account-create-update-cpqvg" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.186145 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-d9hjx"] Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.199125 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-5b89-account-create-update-w7tgb"] Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.200394 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-5b89-account-create-update-w7tgb" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.203387 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.207338 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjw5h\" (UniqueName: \"kubernetes.io/projected/73795e56-2afe-427c-a122-7d604c6ed134-kube-api-access-vjw5h\") pod \"nova-api-c495-account-create-update-cpqvg\" (UID: \"73795e56-2afe-427c-a122-7d604c6ed134\") " pod="openstack/nova-api-c495-account-create-update-cpqvg" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.216840 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-5b89-account-create-update-w7tgb"] Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.220316 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nxjkl\" (UniqueName: \"kubernetes.io/projected/7ead59b4-dcff-45b8-8b1b-d5e449278acb-kube-api-access-nxjkl\") pod \"nova-cell0-db-create-xbk7g\" (UID: \"7ead59b4-dcff-45b8-8b1b-d5e449278acb\") " pod="openstack/nova-cell0-db-create-xbk7g" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.275303 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3052dae6-ee76-44ac-bc5c-6dcb6d9995cf-operator-scripts\") pod \"nova-cell0-5b89-account-create-update-w7tgb\" (UID: \"3052dae6-ee76-44ac-bc5c-6dcb6d9995cf\") " pod="openstack/nova-cell0-5b89-account-create-update-w7tgb" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.275357 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wh9tb\" (UniqueName: \"kubernetes.io/projected/a8a5773f-5b23-400b-8911-1d2d40b9bfd0-kube-api-access-wh9tb\") pod \"nova-cell1-db-create-d9hjx\" (UID: \"a8a5773f-5b23-400b-8911-1d2d40b9bfd0\") " pod="openstack/nova-cell1-db-create-d9hjx" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.275465 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a8a5773f-5b23-400b-8911-1d2d40b9bfd0-operator-scripts\") pod \"nova-cell1-db-create-d9hjx\" (UID: \"a8a5773f-5b23-400b-8911-1d2d40b9bfd0\") " pod="openstack/nova-cell1-db-create-d9hjx" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.275631 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djrjq\" (UniqueName: \"kubernetes.io/projected/3052dae6-ee76-44ac-bc5c-6dcb6d9995cf-kube-api-access-djrjq\") pod \"nova-cell0-5b89-account-create-update-w7tgb\" (UID: \"3052dae6-ee76-44ac-bc5c-6dcb6d9995cf\") " pod="openstack/nova-cell0-5b89-account-create-update-w7tgb" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.344891 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.354191 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-pz462" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.359983 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.364982 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-xbk7g" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.376833 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-33d4-account-create-update-2tg2q"] Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.378388 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-33d4-account-create-update-2tg2q" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.379253 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wh9tb\" (UniqueName: \"kubernetes.io/projected/a8a5773f-5b23-400b-8911-1d2d40b9bfd0-kube-api-access-wh9tb\") pod \"nova-cell1-db-create-d9hjx\" (UID: \"a8a5773f-5b23-400b-8911-1d2d40b9bfd0\") " pod="openstack/nova-cell1-db-create-d9hjx" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.381690 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a8a5773f-5b23-400b-8911-1d2d40b9bfd0-operator-scripts\") pod \"nova-cell1-db-create-d9hjx\" (UID: \"a8a5773f-5b23-400b-8911-1d2d40b9bfd0\") " pod="openstack/nova-cell1-db-create-d9hjx" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.382006 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-djrjq\" (UniqueName: \"kubernetes.io/projected/3052dae6-ee76-44ac-bc5c-6dcb6d9995cf-kube-api-access-djrjq\") pod \"nova-cell0-5b89-account-create-update-w7tgb\" (UID: \"3052dae6-ee76-44ac-bc5c-6dcb6d9995cf\") " pod="openstack/nova-cell0-5b89-account-create-update-w7tgb" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.382290 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3052dae6-ee76-44ac-bc5c-6dcb6d9995cf-operator-scripts\") pod \"nova-cell0-5b89-account-create-update-w7tgb\" (UID: \"3052dae6-ee76-44ac-bc5c-6dcb6d9995cf\") " pod="openstack/nova-cell0-5b89-account-create-update-w7tgb" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.382924 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a8a5773f-5b23-400b-8911-1d2d40b9bfd0-operator-scripts\") pod \"nova-cell1-db-create-d9hjx\" (UID: \"a8a5773f-5b23-400b-8911-1d2d40b9bfd0\") " pod="openstack/nova-cell1-db-create-d9hjx" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.383115 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3052dae6-ee76-44ac-bc5c-6dcb6d9995cf-operator-scripts\") pod \"nova-cell0-5b89-account-create-update-w7tgb\" (UID: \"3052dae6-ee76-44ac-bc5c-6dcb6d9995cf\") " pod="openstack/nova-cell0-5b89-account-create-update-w7tgb" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.387179 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.404712 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-djrjq\" (UniqueName: \"kubernetes.io/projected/3052dae6-ee76-44ac-bc5c-6dcb6d9995cf-kube-api-access-djrjq\") pod \"nova-cell0-5b89-account-create-update-w7tgb\" (UID: \"3052dae6-ee76-44ac-bc5c-6dcb6d9995cf\") " pod="openstack/nova-cell0-5b89-account-create-update-w7tgb" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.404799 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.423656 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-33d4-account-create-update-2tg2q"] Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.426810 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wh9tb\" (UniqueName: \"kubernetes.io/projected/a8a5773f-5b23-400b-8911-1d2d40b9bfd0-kube-api-access-wh9tb\") pod \"nova-cell1-db-create-d9hjx\" (UID: \"a8a5773f-5b23-400b-8911-1d2d40b9bfd0\") " pod="openstack/nova-cell1-db-create-d9hjx" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.450479 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.451727 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-c495-account-create-update-cpqvg" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.452007 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.456027 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.456505 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.483612 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0f287d7e-4b08-4c16-8a23-cf9e69225d3c-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"0f287d7e-4b08-4c16-8a23-cf9e69225d3c\") " pod="openstack/glance-default-internal-api-0" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.483764 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f287d7e-4b08-4c16-8a23-cf9e69225d3c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"0f287d7e-4b08-4c16-8a23-cf9e69225d3c\") " pod="openstack/glance-default-internal-api-0" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.483860 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"0f287d7e-4b08-4c16-8a23-cf9e69225d3c\") " pod="openstack/glance-default-internal-api-0" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.483939 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f287d7e-4b08-4c16-8a23-cf9e69225d3c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"0f287d7e-4b08-4c16-8a23-cf9e69225d3c\") " pod="openstack/glance-default-internal-api-0" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.484032 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vh98c\" (UniqueName: \"kubernetes.io/projected/4647e69c-31d0-495e-a294-f9970aae2c32-kube-api-access-vh98c\") pod \"nova-cell1-33d4-account-create-update-2tg2q\" (UID: \"4647e69c-31d0-495e-a294-f9970aae2c32\") " pod="openstack/nova-cell1-33d4-account-create-update-2tg2q" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.484147 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0f287d7e-4b08-4c16-8a23-cf9e69225d3c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"0f287d7e-4b08-4c16-8a23-cf9e69225d3c\") " pod="openstack/glance-default-internal-api-0" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.484260 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f287d7e-4b08-4c16-8a23-cf9e69225d3c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"0f287d7e-4b08-4c16-8a23-cf9e69225d3c\") " pod="openstack/glance-default-internal-api-0" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.484355 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zpqjg\" (UniqueName: \"kubernetes.io/projected/0f287d7e-4b08-4c16-8a23-cf9e69225d3c-kube-api-access-zpqjg\") pod \"glance-default-internal-api-0\" (UID: \"0f287d7e-4b08-4c16-8a23-cf9e69225d3c\") " pod="openstack/glance-default-internal-api-0" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.484431 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0f287d7e-4b08-4c16-8a23-cf9e69225d3c-logs\") pod \"glance-default-internal-api-0\" (UID: \"0f287d7e-4b08-4c16-8a23-cf9e69225d3c\") " pod="openstack/glance-default-internal-api-0" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.484499 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4647e69c-31d0-495e-a294-f9970aae2c32-operator-scripts\") pod \"nova-cell1-33d4-account-create-update-2tg2q\" (UID: \"4647e69c-31d0-495e-a294-f9970aae2c32\") " pod="openstack/nova-cell1-33d4-account-create-update-2tg2q" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.504228 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.529332 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-d9hjx" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.542551 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-5b89-account-create-update-w7tgb" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.585774 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"0f287d7e-4b08-4c16-8a23-cf9e69225d3c\") " pod="openstack/glance-default-internal-api-0" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.585816 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f287d7e-4b08-4c16-8a23-cf9e69225d3c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"0f287d7e-4b08-4c16-8a23-cf9e69225d3c\") " pod="openstack/glance-default-internal-api-0" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.585845 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vh98c\" (UniqueName: \"kubernetes.io/projected/4647e69c-31d0-495e-a294-f9970aae2c32-kube-api-access-vh98c\") pod \"nova-cell1-33d4-account-create-update-2tg2q\" (UID: \"4647e69c-31d0-495e-a294-f9970aae2c32\") " pod="openstack/nova-cell1-33d4-account-create-update-2tg2q" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.585877 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0f287d7e-4b08-4c16-8a23-cf9e69225d3c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"0f287d7e-4b08-4c16-8a23-cf9e69225d3c\") " pod="openstack/glance-default-internal-api-0" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.585927 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f287d7e-4b08-4c16-8a23-cf9e69225d3c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"0f287d7e-4b08-4c16-8a23-cf9e69225d3c\") " pod="openstack/glance-default-internal-api-0" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.585957 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zpqjg\" (UniqueName: \"kubernetes.io/projected/0f287d7e-4b08-4c16-8a23-cf9e69225d3c-kube-api-access-zpqjg\") pod \"glance-default-internal-api-0\" (UID: \"0f287d7e-4b08-4c16-8a23-cf9e69225d3c\") " pod="openstack/glance-default-internal-api-0" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.585979 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0f287d7e-4b08-4c16-8a23-cf9e69225d3c-logs\") pod \"glance-default-internal-api-0\" (UID: \"0f287d7e-4b08-4c16-8a23-cf9e69225d3c\") " pod="openstack/glance-default-internal-api-0" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.585998 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4647e69c-31d0-495e-a294-f9970aae2c32-operator-scripts\") pod \"nova-cell1-33d4-account-create-update-2tg2q\" (UID: \"4647e69c-31d0-495e-a294-f9970aae2c32\") " pod="openstack/nova-cell1-33d4-account-create-update-2tg2q" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.586045 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0f287d7e-4b08-4c16-8a23-cf9e69225d3c-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"0f287d7e-4b08-4c16-8a23-cf9e69225d3c\") " pod="openstack/glance-default-internal-api-0" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.586071 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f287d7e-4b08-4c16-8a23-cf9e69225d3c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"0f287d7e-4b08-4c16-8a23-cf9e69225d3c\") " pod="openstack/glance-default-internal-api-0" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.587059 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0f287d7e-4b08-4c16-8a23-cf9e69225d3c-logs\") pod \"glance-default-internal-api-0\" (UID: \"0f287d7e-4b08-4c16-8a23-cf9e69225d3c\") " pod="openstack/glance-default-internal-api-0" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.587705 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4647e69c-31d0-495e-a294-f9970aae2c32-operator-scripts\") pod \"nova-cell1-33d4-account-create-update-2tg2q\" (UID: \"4647e69c-31d0-495e-a294-f9970aae2c32\") " pod="openstack/nova-cell1-33d4-account-create-update-2tg2q" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.588017 4716 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"0f287d7e-4b08-4c16-8a23-cf9e69225d3c\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-internal-api-0" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.588404 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0f287d7e-4b08-4c16-8a23-cf9e69225d3c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"0f287d7e-4b08-4c16-8a23-cf9e69225d3c\") " pod="openstack/glance-default-internal-api-0" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.592238 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f287d7e-4b08-4c16-8a23-cf9e69225d3c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"0f287d7e-4b08-4c16-8a23-cf9e69225d3c\") " pod="openstack/glance-default-internal-api-0" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.604638 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0f287d7e-4b08-4c16-8a23-cf9e69225d3c-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"0f287d7e-4b08-4c16-8a23-cf9e69225d3c\") " pod="openstack/glance-default-internal-api-0" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.606711 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f287d7e-4b08-4c16-8a23-cf9e69225d3c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"0f287d7e-4b08-4c16-8a23-cf9e69225d3c\") " pod="openstack/glance-default-internal-api-0" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.624842 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.627496 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"afba3ceb-86bf-4aaf-8ac3-b58ab411850c","Type":"ContainerDied","Data":"c28bdf8eceedbc243af87786b8abfd72c50c2a43117daebec6acb9d7c89977a0"} Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.627554 4716 scope.go:117] "RemoveContainer" containerID="9d342a91e7dc00c4eac8190839f9ea8a4d0e4c05fb04b03355f7cb519bf04840" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.630979 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vh98c\" (UniqueName: \"kubernetes.io/projected/4647e69c-31d0-495e-a294-f9970aae2c32-kube-api-access-vh98c\") pod \"nova-cell1-33d4-account-create-update-2tg2q\" (UID: \"4647e69c-31d0-495e-a294-f9970aae2c32\") " pod="openstack/nova-cell1-33d4-account-create-update-2tg2q" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.668868 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zpqjg\" (UniqueName: \"kubernetes.io/projected/0f287d7e-4b08-4c16-8a23-cf9e69225d3c-kube-api-access-zpqjg\") pod \"glance-default-internal-api-0\" (UID: \"0f287d7e-4b08-4c16-8a23-cf9e69225d3c\") " pod="openstack/glance-default-internal-api-0" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.690321 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f287d7e-4b08-4c16-8a23-cf9e69225d3c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"0f287d7e-4b08-4c16-8a23-cf9e69225d3c\") " pod="openstack/glance-default-internal-api-0" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.716905 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-33d4-account-create-update-2tg2q" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.743608 4716 scope.go:117] "RemoveContainer" containerID="7e62702f6546007c2a5d5db6138318f092dfcf6f6759db6ed04b36ce56729518" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.811726 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"0f287d7e-4b08-4c16-8a23-cf9e69225d3c\") " pod="openstack/glance-default-internal-api-0" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.816271 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0480133a-7588-4a60-921e-e53de8998ca9" path="/var/lib/kubelet/pods/0480133a-7588-4a60-921e-e53de8998ca9/volumes" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.817036 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5eca408c-c86b-4dd6-a1f9-81ef16e12da4" path="/var/lib/kubelet/pods/5eca408c-c86b-4dd6-a1f9-81ef16e12da4/volumes" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.817749 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.825621 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.830801 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.835997 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.838594 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.840901 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.841680 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.847148 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.901547 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-5488d7fd99-pwmf5" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.915919 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f4f8c7b-38d4-4054-bb66-c51f32448567-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"9f4f8c7b-38d4-4054-bb66-c51f32448567\") " pod="openstack/glance-default-external-api-0" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.916626 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f4f8c7b-38d4-4054-bb66-c51f32448567-logs\") pod \"glance-default-external-api-0\" (UID: \"9f4f8c7b-38d4-4054-bb66-c51f32448567\") " pod="openstack/glance-default-external-api-0" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.916835 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f4f8c7b-38d4-4054-bb66-c51f32448567-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"9f4f8c7b-38d4-4054-bb66-c51f32448567\") " pod="openstack/glance-default-external-api-0" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.917067 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9f4f8c7b-38d4-4054-bb66-c51f32448567-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"9f4f8c7b-38d4-4054-bb66-c51f32448567\") " pod="openstack/glance-default-external-api-0" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.917200 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f4f8c7b-38d4-4054-bb66-c51f32448567-scripts\") pod \"glance-default-external-api-0\" (UID: \"9f4f8c7b-38d4-4054-bb66-c51f32448567\") " pod="openstack/glance-default-external-api-0" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.917287 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f4f8c7b-38d4-4054-bb66-c51f32448567-config-data\") pod \"glance-default-external-api-0\" (UID: \"9f4f8c7b-38d4-4054-bb66-c51f32448567\") " pod="openstack/glance-default-external-api-0" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.917332 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"9f4f8c7b-38d4-4054-bb66-c51f32448567\") " pod="openstack/glance-default-external-api-0" Dec 07 16:20:29 crc kubenswrapper[4716]: I1207 16:20:29.917505 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9n9fk\" (UniqueName: \"kubernetes.io/projected/9f4f8c7b-38d4-4054-bb66-c51f32448567-kube-api-access-9n9fk\") pod \"glance-default-external-api-0\" (UID: \"9f4f8c7b-38d4-4054-bb66-c51f32448567\") " pod="openstack/glance-default-external-api-0" Dec 07 16:20:30 crc kubenswrapper[4716]: I1207 16:20:30.019209 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f4f8c7b-38d4-4054-bb66-c51f32448567-logs\") pod \"glance-default-external-api-0\" (UID: \"9f4f8c7b-38d4-4054-bb66-c51f32448567\") " pod="openstack/glance-default-external-api-0" Dec 07 16:20:30 crc kubenswrapper[4716]: I1207 16:20:30.019269 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f4f8c7b-38d4-4054-bb66-c51f32448567-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"9f4f8c7b-38d4-4054-bb66-c51f32448567\") " pod="openstack/glance-default-external-api-0" Dec 07 16:20:30 crc kubenswrapper[4716]: I1207 16:20:30.019315 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9f4f8c7b-38d4-4054-bb66-c51f32448567-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"9f4f8c7b-38d4-4054-bb66-c51f32448567\") " pod="openstack/glance-default-external-api-0" Dec 07 16:20:30 crc kubenswrapper[4716]: I1207 16:20:30.019347 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f4f8c7b-38d4-4054-bb66-c51f32448567-scripts\") pod \"glance-default-external-api-0\" (UID: \"9f4f8c7b-38d4-4054-bb66-c51f32448567\") " pod="openstack/glance-default-external-api-0" Dec 07 16:20:30 crc kubenswrapper[4716]: I1207 16:20:30.019377 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f4f8c7b-38d4-4054-bb66-c51f32448567-config-data\") pod \"glance-default-external-api-0\" (UID: \"9f4f8c7b-38d4-4054-bb66-c51f32448567\") " pod="openstack/glance-default-external-api-0" Dec 07 16:20:30 crc kubenswrapper[4716]: I1207 16:20:30.019395 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"9f4f8c7b-38d4-4054-bb66-c51f32448567\") " pod="openstack/glance-default-external-api-0" Dec 07 16:20:30 crc kubenswrapper[4716]: I1207 16:20:30.019436 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9n9fk\" (UniqueName: \"kubernetes.io/projected/9f4f8c7b-38d4-4054-bb66-c51f32448567-kube-api-access-9n9fk\") pod \"glance-default-external-api-0\" (UID: \"9f4f8c7b-38d4-4054-bb66-c51f32448567\") " pod="openstack/glance-default-external-api-0" Dec 07 16:20:30 crc kubenswrapper[4716]: I1207 16:20:30.019471 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f4f8c7b-38d4-4054-bb66-c51f32448567-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"9f4f8c7b-38d4-4054-bb66-c51f32448567\") " pod="openstack/glance-default-external-api-0" Dec 07 16:20:30 crc kubenswrapper[4716]: I1207 16:20:30.022525 4716 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"9f4f8c7b-38d4-4054-bb66-c51f32448567\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-external-api-0" Dec 07 16:20:30 crc kubenswrapper[4716]: I1207 16:20:30.022990 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f4f8c7b-38d4-4054-bb66-c51f32448567-logs\") pod \"glance-default-external-api-0\" (UID: \"9f4f8c7b-38d4-4054-bb66-c51f32448567\") " pod="openstack/glance-default-external-api-0" Dec 07 16:20:30 crc kubenswrapper[4716]: I1207 16:20:30.023570 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9f4f8c7b-38d4-4054-bb66-c51f32448567-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"9f4f8c7b-38d4-4054-bb66-c51f32448567\") " pod="openstack/glance-default-external-api-0" Dec 07 16:20:30 crc kubenswrapper[4716]: I1207 16:20:30.024938 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f4f8c7b-38d4-4054-bb66-c51f32448567-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"9f4f8c7b-38d4-4054-bb66-c51f32448567\") " pod="openstack/glance-default-external-api-0" Dec 07 16:20:30 crc kubenswrapper[4716]: I1207 16:20:30.030610 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f4f8c7b-38d4-4054-bb66-c51f32448567-scripts\") pod \"glance-default-external-api-0\" (UID: \"9f4f8c7b-38d4-4054-bb66-c51f32448567\") " pod="openstack/glance-default-external-api-0" Dec 07 16:20:30 crc kubenswrapper[4716]: I1207 16:20:30.038419 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f4f8c7b-38d4-4054-bb66-c51f32448567-config-data\") pod \"glance-default-external-api-0\" (UID: \"9f4f8c7b-38d4-4054-bb66-c51f32448567\") " pod="openstack/glance-default-external-api-0" Dec 07 16:20:30 crc kubenswrapper[4716]: I1207 16:20:30.039093 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f4f8c7b-38d4-4054-bb66-c51f32448567-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"9f4f8c7b-38d4-4054-bb66-c51f32448567\") " pod="openstack/glance-default-external-api-0" Dec 07 16:20:30 crc kubenswrapper[4716]: I1207 16:20:30.055716 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9n9fk\" (UniqueName: \"kubernetes.io/projected/9f4f8c7b-38d4-4054-bb66-c51f32448567-kube-api-access-9n9fk\") pod \"glance-default-external-api-0\" (UID: \"9f4f8c7b-38d4-4054-bb66-c51f32448567\") " pod="openstack/glance-default-external-api-0" Dec 07 16:20:30 crc kubenswrapper[4716]: I1207 16:20:30.070369 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"9f4f8c7b-38d4-4054-bb66-c51f32448567\") " pod="openstack/glance-default-external-api-0" Dec 07 16:20:30 crc kubenswrapper[4716]: I1207 16:20:30.070396 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-xbk7g"] Dec 07 16:20:30 crc kubenswrapper[4716]: I1207 16:20:30.125271 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 07 16:20:30 crc kubenswrapper[4716]: I1207 16:20:30.155821 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-c495-account-create-update-cpqvg"] Dec 07 16:20:30 crc kubenswrapper[4716]: I1207 16:20:30.192090 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 07 16:20:30 crc kubenswrapper[4716]: I1207 16:20:30.208066 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 07 16:20:30 crc kubenswrapper[4716]: I1207 16:20:30.208395 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="c088cd1b-83a8-4f05-b7db-11c8ffb59fab" containerName="kube-state-metrics" containerID="cri-o://5695d3e0ddfd3193c3f76ef6bfbe23181c8fd9bc6f004904a15d8d152feadcd5" gracePeriod=30 Dec 07 16:20:30 crc kubenswrapper[4716]: I1207 16:20:30.372953 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-33d4-account-create-update-2tg2q"] Dec 07 16:20:30 crc kubenswrapper[4716]: I1207 16:20:30.520232 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-d9hjx"] Dec 07 16:20:30 crc kubenswrapper[4716]: I1207 16:20:30.538474 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-5b89-account-create-update-w7tgb"] Dec 07 16:20:30 crc kubenswrapper[4716]: I1207 16:20:30.553773 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-pz462"] Dec 07 16:20:30 crc kubenswrapper[4716]: I1207 16:20:30.666586 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-33d4-account-create-update-2tg2q" event={"ID":"4647e69c-31d0-495e-a294-f9970aae2c32","Type":"ContainerStarted","Data":"7ad63c67cc7344b576e5c104851f19bd0e23971ef38e833eba4570bbe4d63af5"} Dec 07 16:20:30 crc kubenswrapper[4716]: I1207 16:20:30.666660 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-33d4-account-create-update-2tg2q" event={"ID":"4647e69c-31d0-495e-a294-f9970aae2c32","Type":"ContainerStarted","Data":"03e061a83e97940fa65de350a932fb9e219e83dd182f5a483272e6d896518a13"} Dec 07 16:20:30 crc kubenswrapper[4716]: I1207 16:20:30.668417 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-d9hjx" event={"ID":"a8a5773f-5b23-400b-8911-1d2d40b9bfd0","Type":"ContainerStarted","Data":"4c16c36568b5d98c100b2910a47f5ed901cb5394863acbfec17644d823ba62eb"} Dec 07 16:20:30 crc kubenswrapper[4716]: I1207 16:20:30.676716 4716 generic.go:334] "Generic (PLEG): container finished" podID="7ead59b4-dcff-45b8-8b1b-d5e449278acb" containerID="a93e9710e35e18568b8b3e95c22a35043085b7c042fb81b1aa8b550b36ab9a3f" exitCode=0 Dec 07 16:20:30 crc kubenswrapper[4716]: I1207 16:20:30.676878 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-xbk7g" event={"ID":"7ead59b4-dcff-45b8-8b1b-d5e449278acb","Type":"ContainerDied","Data":"a93e9710e35e18568b8b3e95c22a35043085b7c042fb81b1aa8b550b36ab9a3f"} Dec 07 16:20:30 crc kubenswrapper[4716]: I1207 16:20:30.676911 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-xbk7g" event={"ID":"7ead59b4-dcff-45b8-8b1b-d5e449278acb","Type":"ContainerStarted","Data":"eaeb45e5bf8e6a4649a47c98490e81e63cf929deb78275a0ea041aa2d4f2bb2b"} Dec 07 16:20:30 crc kubenswrapper[4716]: I1207 16:20:30.678905 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-c495-account-create-update-cpqvg" event={"ID":"73795e56-2afe-427c-a122-7d604c6ed134","Type":"ContainerStarted","Data":"290aaea26096362bace3d7ab97853c9057fe0afaa3cf08806cd429ea3ae13f7a"} Dec 07 16:20:30 crc kubenswrapper[4716]: I1207 16:20:30.678939 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-c495-account-create-update-cpqvg" event={"ID":"73795e56-2afe-427c-a122-7d604c6ed134","Type":"ContainerStarted","Data":"e238b8abff70cca389d44c3f2f52dfe683051c07f6d961bcf2d8c5fc84af2f82"} Dec 07 16:20:30 crc kubenswrapper[4716]: I1207 16:20:30.686597 4716 generic.go:334] "Generic (PLEG): container finished" podID="c088cd1b-83a8-4f05-b7db-11c8ffb59fab" containerID="5695d3e0ddfd3193c3f76ef6bfbe23181c8fd9bc6f004904a15d8d152feadcd5" exitCode=2 Dec 07 16:20:30 crc kubenswrapper[4716]: I1207 16:20:30.686704 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"c088cd1b-83a8-4f05-b7db-11c8ffb59fab","Type":"ContainerDied","Data":"5695d3e0ddfd3193c3f76ef6bfbe23181c8fd9bc6f004904a15d8d152feadcd5"} Dec 07 16:20:30 crc kubenswrapper[4716]: I1207 16:20:30.692657 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-33d4-account-create-update-2tg2q" podStartSLOduration=1.692633083 podStartE2EDuration="1.692633083s" podCreationTimestamp="2025-12-07 16:20:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:20:30.685570872 +0000 UTC m=+1093.375855784" watchObservedRunningTime="2025-12-07 16:20:30.692633083 +0000 UTC m=+1093.382917985" Dec 07 16:20:30 crc kubenswrapper[4716]: I1207 16:20:30.696722 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6873cce2-e68e-47ca-8ae8-7b612815a223","Type":"ContainerStarted","Data":"8d03bbfd8f4f7e0d28b7cba869d384348e81879a56ac18e7ca03abde08694745"} Dec 07 16:20:30 crc kubenswrapper[4716]: I1207 16:20:30.719435 4716 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-86c7567d4-99rx9" podUID="c291e114-7940-46bc-91d3-a8fa256549ff" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Dec 07 16:20:30 crc kubenswrapper[4716]: I1207 16:20:30.719557 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-86c7567d4-99rx9" Dec 07 16:20:30 crc kubenswrapper[4716]: I1207 16:20:30.734723 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-5b89-account-create-update-w7tgb" event={"ID":"3052dae6-ee76-44ac-bc5c-6dcb6d9995cf","Type":"ContainerStarted","Data":"743f1108a3f30863c154c0f4a97117edde65287b6ee02ad37d6dcfc8bf7296e4"} Dec 07 16:20:30 crc kubenswrapper[4716]: I1207 16:20:30.737804 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-pz462" event={"ID":"a14d4b2b-4e79-4258-9d9f-46198eca8a41","Type":"ContainerStarted","Data":"1b5af05ebc40d55cb25fde5aa45b802ff769ba95dd1ebd29f7afee1bdf69adb4"} Dec 07 16:20:30 crc kubenswrapper[4716]: I1207 16:20:30.741056 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-c495-account-create-update-cpqvg" podStartSLOduration=2.741032703 podStartE2EDuration="2.741032703s" podCreationTimestamp="2025-12-07 16:20:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:20:30.724872305 +0000 UTC m=+1093.415157217" watchObservedRunningTime="2025-12-07 16:20:30.741032703 +0000 UTC m=+1093.431317615" Dec 07 16:20:30 crc kubenswrapper[4716]: I1207 16:20:30.780844 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 07 16:20:30 crc kubenswrapper[4716]: I1207 16:20:30.783431 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 07 16:20:30 crc kubenswrapper[4716]: I1207 16:20:30.848547 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-glmlq\" (UniqueName: \"kubernetes.io/projected/c088cd1b-83a8-4f05-b7db-11c8ffb59fab-kube-api-access-glmlq\") pod \"c088cd1b-83a8-4f05-b7db-11c8ffb59fab\" (UID: \"c088cd1b-83a8-4f05-b7db-11c8ffb59fab\") " Dec 07 16:20:30 crc kubenswrapper[4716]: I1207 16:20:30.859319 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 07 16:20:30 crc kubenswrapper[4716]: I1207 16:20:30.882726 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c088cd1b-83a8-4f05-b7db-11c8ffb59fab-kube-api-access-glmlq" (OuterVolumeSpecName: "kube-api-access-glmlq") pod "c088cd1b-83a8-4f05-b7db-11c8ffb59fab" (UID: "c088cd1b-83a8-4f05-b7db-11c8ffb59fab"). InnerVolumeSpecName "kube-api-access-glmlq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:20:30 crc kubenswrapper[4716]: I1207 16:20:30.951062 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-glmlq\" (UniqueName: \"kubernetes.io/projected/c088cd1b-83a8-4f05-b7db-11c8ffb59fab-kube-api-access-glmlq\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:31 crc kubenswrapper[4716]: I1207 16:20:31.693135 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="afba3ceb-86bf-4aaf-8ac3-b58ab411850c" path="/var/lib/kubelet/pods/afba3ceb-86bf-4aaf-8ac3-b58ab411850c/volumes" Dec 07 16:20:31 crc kubenswrapper[4716]: I1207 16:20:31.747499 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0f287d7e-4b08-4c16-8a23-cf9e69225d3c","Type":"ContainerStarted","Data":"4c4d5b7abc7f17e8d7d39c7f506f9675d0631728b9edc52f0cf95533b20ce4d7"} Dec 07 16:20:31 crc kubenswrapper[4716]: I1207 16:20:31.747549 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0f287d7e-4b08-4c16-8a23-cf9e69225d3c","Type":"ContainerStarted","Data":"912e3aab38540c5542bc19e257d3dd827a56b746d153caf4284ce65b37b7127f"} Dec 07 16:20:31 crc kubenswrapper[4716]: I1207 16:20:31.751181 4716 generic.go:334] "Generic (PLEG): container finished" podID="3052dae6-ee76-44ac-bc5c-6dcb6d9995cf" containerID="0e6d7d18623b527169bf2eaaab3b349ad13833aa0638e2e22b0ab5fcd0b0f715" exitCode=0 Dec 07 16:20:31 crc kubenswrapper[4716]: I1207 16:20:31.751229 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-5b89-account-create-update-w7tgb" event={"ID":"3052dae6-ee76-44ac-bc5c-6dcb6d9995cf","Type":"ContainerDied","Data":"0e6d7d18623b527169bf2eaaab3b349ad13833aa0638e2e22b0ab5fcd0b0f715"} Dec 07 16:20:31 crc kubenswrapper[4716]: I1207 16:20:31.768756 4716 generic.go:334] "Generic (PLEG): container finished" podID="73795e56-2afe-427c-a122-7d604c6ed134" containerID="290aaea26096362bace3d7ab97853c9057fe0afaa3cf08806cd429ea3ae13f7a" exitCode=0 Dec 07 16:20:31 crc kubenswrapper[4716]: I1207 16:20:31.768832 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-c495-account-create-update-cpqvg" event={"ID":"73795e56-2afe-427c-a122-7d604c6ed134","Type":"ContainerDied","Data":"290aaea26096362bace3d7ab97853c9057fe0afaa3cf08806cd429ea3ae13f7a"} Dec 07 16:20:31 crc kubenswrapper[4716]: I1207 16:20:31.771419 4716 generic.go:334] "Generic (PLEG): container finished" podID="4647e69c-31d0-495e-a294-f9970aae2c32" containerID="7ad63c67cc7344b576e5c104851f19bd0e23971ef38e833eba4570bbe4d63af5" exitCode=0 Dec 07 16:20:31 crc kubenswrapper[4716]: I1207 16:20:31.771478 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-33d4-account-create-update-2tg2q" event={"ID":"4647e69c-31d0-495e-a294-f9970aae2c32","Type":"ContainerDied","Data":"7ad63c67cc7344b576e5c104851f19bd0e23971ef38e833eba4570bbe4d63af5"} Dec 07 16:20:31 crc kubenswrapper[4716]: I1207 16:20:31.777728 4716 generic.go:334] "Generic (PLEG): container finished" podID="a8a5773f-5b23-400b-8911-1d2d40b9bfd0" containerID="25a2fe086275ad4a71a835f4d1d54853927f11009001e16c4bde04a32db4975d" exitCode=0 Dec 07 16:20:31 crc kubenswrapper[4716]: I1207 16:20:31.777781 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-d9hjx" event={"ID":"a8a5773f-5b23-400b-8911-1d2d40b9bfd0","Type":"ContainerDied","Data":"25a2fe086275ad4a71a835f4d1d54853927f11009001e16c4bde04a32db4975d"} Dec 07 16:20:31 crc kubenswrapper[4716]: I1207 16:20:31.782920 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6873cce2-e68e-47ca-8ae8-7b612815a223","Type":"ContainerStarted","Data":"ef45036db19a07ce203b9018d7f545d5aaaeb31f8ee89c8dd4a93b441821fa1a"} Dec 07 16:20:31 crc kubenswrapper[4716]: I1207 16:20:31.787763 4716 generic.go:334] "Generic (PLEG): container finished" podID="a14d4b2b-4e79-4258-9d9f-46198eca8a41" containerID="2df3471b336e2c8ff319c52b304c9fbe472180f8eb3d143bcee10f51a7b68902" exitCode=0 Dec 07 16:20:31 crc kubenswrapper[4716]: I1207 16:20:31.787871 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-pz462" event={"ID":"a14d4b2b-4e79-4258-9d9f-46198eca8a41","Type":"ContainerDied","Data":"2df3471b336e2c8ff319c52b304c9fbe472180f8eb3d143bcee10f51a7b68902"} Dec 07 16:20:31 crc kubenswrapper[4716]: I1207 16:20:31.802412 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"c088cd1b-83a8-4f05-b7db-11c8ffb59fab","Type":"ContainerDied","Data":"bb8e5f62eece4be6882fee91108453b363425e0c0ff925e89657243b1889ca2a"} Dec 07 16:20:31 crc kubenswrapper[4716]: I1207 16:20:31.802481 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 07 16:20:31 crc kubenswrapper[4716]: I1207 16:20:31.802653 4716 scope.go:117] "RemoveContainer" containerID="5695d3e0ddfd3193c3f76ef6bfbe23181c8fd9bc6f004904a15d8d152feadcd5" Dec 07 16:20:31 crc kubenswrapper[4716]: I1207 16:20:31.805697 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9f4f8c7b-38d4-4054-bb66-c51f32448567","Type":"ContainerStarted","Data":"4fc53af8eda9ef77cb6d25514ed9d34134b3a2bc2293a357d7f1a17c557f8b99"} Dec 07 16:20:31 crc kubenswrapper[4716]: I1207 16:20:31.805761 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9f4f8c7b-38d4-4054-bb66-c51f32448567","Type":"ContainerStarted","Data":"ebb081251d30ea6a76ba1d9c66eeaedf019c6366da5335fd66f272fbce852804"} Dec 07 16:20:31 crc kubenswrapper[4716]: I1207 16:20:31.861211 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 07 16:20:31 crc kubenswrapper[4716]: I1207 16:20:31.875152 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 07 16:20:31 crc kubenswrapper[4716]: I1207 16:20:31.883350 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 07 16:20:31 crc kubenswrapper[4716]: E1207 16:20:31.883913 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c088cd1b-83a8-4f05-b7db-11c8ffb59fab" containerName="kube-state-metrics" Dec 07 16:20:31 crc kubenswrapper[4716]: I1207 16:20:31.883941 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="c088cd1b-83a8-4f05-b7db-11c8ffb59fab" containerName="kube-state-metrics" Dec 07 16:20:31 crc kubenswrapper[4716]: I1207 16:20:31.884356 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="c088cd1b-83a8-4f05-b7db-11c8ffb59fab" containerName="kube-state-metrics" Dec 07 16:20:31 crc kubenswrapper[4716]: I1207 16:20:31.887011 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 07 16:20:31 crc kubenswrapper[4716]: I1207 16:20:31.889710 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Dec 07 16:20:31 crc kubenswrapper[4716]: I1207 16:20:31.891007 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Dec 07 16:20:31 crc kubenswrapper[4716]: I1207 16:20:31.900613 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 07 16:20:32 crc kubenswrapper[4716]: I1207 16:20:32.086324 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/7953b7b7-a0c6-452c-868a-dcc4a1bda1fa-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"7953b7b7-a0c6-452c-868a-dcc4a1bda1fa\") " pod="openstack/kube-state-metrics-0" Dec 07 16:20:32 crc kubenswrapper[4716]: I1207 16:20:32.086693 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7953b7b7-a0c6-452c-868a-dcc4a1bda1fa-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"7953b7b7-a0c6-452c-868a-dcc4a1bda1fa\") " pod="openstack/kube-state-metrics-0" Dec 07 16:20:32 crc kubenswrapper[4716]: I1207 16:20:32.086844 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6zvm5\" (UniqueName: \"kubernetes.io/projected/7953b7b7-a0c6-452c-868a-dcc4a1bda1fa-kube-api-access-6zvm5\") pod \"kube-state-metrics-0\" (UID: \"7953b7b7-a0c6-452c-868a-dcc4a1bda1fa\") " pod="openstack/kube-state-metrics-0" Dec 07 16:20:32 crc kubenswrapper[4716]: I1207 16:20:32.086969 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/7953b7b7-a0c6-452c-868a-dcc4a1bda1fa-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"7953b7b7-a0c6-452c-868a-dcc4a1bda1fa\") " pod="openstack/kube-state-metrics-0" Dec 07 16:20:32 crc kubenswrapper[4716]: I1207 16:20:32.189724 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7953b7b7-a0c6-452c-868a-dcc4a1bda1fa-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"7953b7b7-a0c6-452c-868a-dcc4a1bda1fa\") " pod="openstack/kube-state-metrics-0" Dec 07 16:20:32 crc kubenswrapper[4716]: I1207 16:20:32.190013 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6zvm5\" (UniqueName: \"kubernetes.io/projected/7953b7b7-a0c6-452c-868a-dcc4a1bda1fa-kube-api-access-6zvm5\") pod \"kube-state-metrics-0\" (UID: \"7953b7b7-a0c6-452c-868a-dcc4a1bda1fa\") " pod="openstack/kube-state-metrics-0" Dec 07 16:20:32 crc kubenswrapper[4716]: I1207 16:20:32.190163 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/7953b7b7-a0c6-452c-868a-dcc4a1bda1fa-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"7953b7b7-a0c6-452c-868a-dcc4a1bda1fa\") " pod="openstack/kube-state-metrics-0" Dec 07 16:20:32 crc kubenswrapper[4716]: I1207 16:20:32.190598 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/7953b7b7-a0c6-452c-868a-dcc4a1bda1fa-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"7953b7b7-a0c6-452c-868a-dcc4a1bda1fa\") " pod="openstack/kube-state-metrics-0" Dec 07 16:20:32 crc kubenswrapper[4716]: I1207 16:20:32.194928 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7953b7b7-a0c6-452c-868a-dcc4a1bda1fa-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"7953b7b7-a0c6-452c-868a-dcc4a1bda1fa\") " pod="openstack/kube-state-metrics-0" Dec 07 16:20:32 crc kubenswrapper[4716]: I1207 16:20:32.202031 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/7953b7b7-a0c6-452c-868a-dcc4a1bda1fa-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"7953b7b7-a0c6-452c-868a-dcc4a1bda1fa\") " pod="openstack/kube-state-metrics-0" Dec 07 16:20:32 crc kubenswrapper[4716]: I1207 16:20:32.202545 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/7953b7b7-a0c6-452c-868a-dcc4a1bda1fa-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"7953b7b7-a0c6-452c-868a-dcc4a1bda1fa\") " pod="openstack/kube-state-metrics-0" Dec 07 16:20:32 crc kubenswrapper[4716]: I1207 16:20:32.209286 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6zvm5\" (UniqueName: \"kubernetes.io/projected/7953b7b7-a0c6-452c-868a-dcc4a1bda1fa-kube-api-access-6zvm5\") pod \"kube-state-metrics-0\" (UID: \"7953b7b7-a0c6-452c-868a-dcc4a1bda1fa\") " pod="openstack/kube-state-metrics-0" Dec 07 16:20:32 crc kubenswrapper[4716]: I1207 16:20:32.235125 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 07 16:20:32 crc kubenswrapper[4716]: I1207 16:20:32.265288 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-xbk7g" Dec 07 16:20:32 crc kubenswrapper[4716]: E1207 16:20:32.290302 4716 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podafba3ceb_86bf_4aaf_8ac3_b58ab411850c.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podafba3ceb_86bf_4aaf_8ac3_b58ab411850c.slice/crio-c28bdf8eceedbc243af87786b8abfd72c50c2a43117daebec6acb9d7c89977a0\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5eca408c_c86b_4dd6_a1f9_81ef16e12da4.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5eca408c_c86b_4dd6_a1f9_81ef16e12da4.slice/crio-74f894d324ce5d8ac07b7d951146a82fe1f7fc35ef5ec12b146942ecc20cd47a\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode1dd089c_abb1_4317_aa2f_d301b7ae730d.slice/crio-53eb298d61c184858cf42d228a84ba335e5e977399e58be6ed004d8ed973b4ba.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc088cd1b_83a8_4f05_b7db_11c8ffb59fab.slice/crio-5695d3e0ddfd3193c3f76ef6bfbe23181c8fd9bc6f004904a15d8d152feadcd5.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc088cd1b_83a8_4f05_b7db_11c8ffb59fab.slice/crio-conmon-5695d3e0ddfd3193c3f76ef6bfbe23181c8fd9bc6f004904a15d8d152feadcd5.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode1dd089c_abb1_4317_aa2f_d301b7ae730d.slice/crio-conmon-53eb298d61c184858cf42d228a84ba335e5e977399e58be6ed004d8ed973b4ba.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc088cd1b_83a8_4f05_b7db_11c8ffb59fab.slice/crio-bb8e5f62eece4be6882fee91108453b363425e0c0ff925e89657243b1889ca2a\": RecentStats: unable to find data in memory cache]" Dec 07 16:20:32 crc kubenswrapper[4716]: I1207 16:20:32.393022 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nxjkl\" (UniqueName: \"kubernetes.io/projected/7ead59b4-dcff-45b8-8b1b-d5e449278acb-kube-api-access-nxjkl\") pod \"7ead59b4-dcff-45b8-8b1b-d5e449278acb\" (UID: \"7ead59b4-dcff-45b8-8b1b-d5e449278acb\") " Dec 07 16:20:32 crc kubenswrapper[4716]: I1207 16:20:32.393365 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ead59b4-dcff-45b8-8b1b-d5e449278acb-operator-scripts\") pod \"7ead59b4-dcff-45b8-8b1b-d5e449278acb\" (UID: \"7ead59b4-dcff-45b8-8b1b-d5e449278acb\") " Dec 07 16:20:32 crc kubenswrapper[4716]: I1207 16:20:32.395395 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ead59b4-dcff-45b8-8b1b-d5e449278acb-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7ead59b4-dcff-45b8-8b1b-d5e449278acb" (UID: "7ead59b4-dcff-45b8-8b1b-d5e449278acb"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:20:32 crc kubenswrapper[4716]: I1207 16:20:32.401630 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ead59b4-dcff-45b8-8b1b-d5e449278acb-kube-api-access-nxjkl" (OuterVolumeSpecName: "kube-api-access-nxjkl") pod "7ead59b4-dcff-45b8-8b1b-d5e449278acb" (UID: "7ead59b4-dcff-45b8-8b1b-d5e449278acb"). InnerVolumeSpecName "kube-api-access-nxjkl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:20:32 crc kubenswrapper[4716]: I1207 16:20:32.496342 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nxjkl\" (UniqueName: \"kubernetes.io/projected/7ead59b4-dcff-45b8-8b1b-d5e449278acb-kube-api-access-nxjkl\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:32 crc kubenswrapper[4716]: I1207 16:20:32.496610 4716 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ead59b4-dcff-45b8-8b1b-d5e449278acb-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:32 crc kubenswrapper[4716]: I1207 16:20:32.500977 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 07 16:20:32 crc kubenswrapper[4716]: I1207 16:20:32.597346 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1dd089c-abb1-4317-aa2f-d301b7ae730d-combined-ca-bundle\") pod \"e1dd089c-abb1-4317-aa2f-d301b7ae730d\" (UID: \"e1dd089c-abb1-4317-aa2f-d301b7ae730d\") " Dec 07 16:20:32 crc kubenswrapper[4716]: I1207 16:20:32.597456 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e1dd089c-abb1-4317-aa2f-d301b7ae730d-logs\") pod \"e1dd089c-abb1-4317-aa2f-d301b7ae730d\" (UID: \"e1dd089c-abb1-4317-aa2f-d301b7ae730d\") " Dec 07 16:20:32 crc kubenswrapper[4716]: I1207 16:20:32.597502 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nvvr4\" (UniqueName: \"kubernetes.io/projected/e1dd089c-abb1-4317-aa2f-d301b7ae730d-kube-api-access-nvvr4\") pod \"e1dd089c-abb1-4317-aa2f-d301b7ae730d\" (UID: \"e1dd089c-abb1-4317-aa2f-d301b7ae730d\") " Dec 07 16:20:32 crc kubenswrapper[4716]: I1207 16:20:32.597574 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e1dd089c-abb1-4317-aa2f-d301b7ae730d-config-data-custom\") pod \"e1dd089c-abb1-4317-aa2f-d301b7ae730d\" (UID: \"e1dd089c-abb1-4317-aa2f-d301b7ae730d\") " Dec 07 16:20:32 crc kubenswrapper[4716]: I1207 16:20:32.597708 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1dd089c-abb1-4317-aa2f-d301b7ae730d-config-data\") pod \"e1dd089c-abb1-4317-aa2f-d301b7ae730d\" (UID: \"e1dd089c-abb1-4317-aa2f-d301b7ae730d\") " Dec 07 16:20:32 crc kubenswrapper[4716]: I1207 16:20:32.597757 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e1dd089c-abb1-4317-aa2f-d301b7ae730d-scripts\") pod \"e1dd089c-abb1-4317-aa2f-d301b7ae730d\" (UID: \"e1dd089c-abb1-4317-aa2f-d301b7ae730d\") " Dec 07 16:20:32 crc kubenswrapper[4716]: I1207 16:20:32.597802 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e1dd089c-abb1-4317-aa2f-d301b7ae730d-etc-machine-id\") pod \"e1dd089c-abb1-4317-aa2f-d301b7ae730d\" (UID: \"e1dd089c-abb1-4317-aa2f-d301b7ae730d\") " Dec 07 16:20:32 crc kubenswrapper[4716]: I1207 16:20:32.598267 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e1dd089c-abb1-4317-aa2f-d301b7ae730d-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "e1dd089c-abb1-4317-aa2f-d301b7ae730d" (UID: "e1dd089c-abb1-4317-aa2f-d301b7ae730d"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 16:20:32 crc kubenswrapper[4716]: I1207 16:20:32.599975 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e1dd089c-abb1-4317-aa2f-d301b7ae730d-logs" (OuterVolumeSpecName: "logs") pod "e1dd089c-abb1-4317-aa2f-d301b7ae730d" (UID: "e1dd089c-abb1-4317-aa2f-d301b7ae730d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:20:32 crc kubenswrapper[4716]: I1207 16:20:32.605360 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1dd089c-abb1-4317-aa2f-d301b7ae730d-kube-api-access-nvvr4" (OuterVolumeSpecName: "kube-api-access-nvvr4") pod "e1dd089c-abb1-4317-aa2f-d301b7ae730d" (UID: "e1dd089c-abb1-4317-aa2f-d301b7ae730d"). InnerVolumeSpecName "kube-api-access-nvvr4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:20:32 crc kubenswrapper[4716]: I1207 16:20:32.607274 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1dd089c-abb1-4317-aa2f-d301b7ae730d-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "e1dd089c-abb1-4317-aa2f-d301b7ae730d" (UID: "e1dd089c-abb1-4317-aa2f-d301b7ae730d"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:20:32 crc kubenswrapper[4716]: I1207 16:20:32.611843 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1dd089c-abb1-4317-aa2f-d301b7ae730d-scripts" (OuterVolumeSpecName: "scripts") pod "e1dd089c-abb1-4317-aa2f-d301b7ae730d" (UID: "e1dd089c-abb1-4317-aa2f-d301b7ae730d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:20:32 crc kubenswrapper[4716]: I1207 16:20:32.635208 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1dd089c-abb1-4317-aa2f-d301b7ae730d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e1dd089c-abb1-4317-aa2f-d301b7ae730d" (UID: "e1dd089c-abb1-4317-aa2f-d301b7ae730d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:20:32 crc kubenswrapper[4716]: I1207 16:20:32.663032 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1dd089c-abb1-4317-aa2f-d301b7ae730d-config-data" (OuterVolumeSpecName: "config-data") pod "e1dd089c-abb1-4317-aa2f-d301b7ae730d" (UID: "e1dd089c-abb1-4317-aa2f-d301b7ae730d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:20:32 crc kubenswrapper[4716]: I1207 16:20:32.701741 4716 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e1dd089c-abb1-4317-aa2f-d301b7ae730d-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:32 crc kubenswrapper[4716]: I1207 16:20:32.702744 4716 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1dd089c-abb1-4317-aa2f-d301b7ae730d-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:32 crc kubenswrapper[4716]: I1207 16:20:32.702806 4716 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e1dd089c-abb1-4317-aa2f-d301b7ae730d-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:32 crc kubenswrapper[4716]: I1207 16:20:32.702860 4716 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e1dd089c-abb1-4317-aa2f-d301b7ae730d-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:32 crc kubenswrapper[4716]: I1207 16:20:32.702911 4716 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1dd089c-abb1-4317-aa2f-d301b7ae730d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:32 crc kubenswrapper[4716]: I1207 16:20:32.702987 4716 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e1dd089c-abb1-4317-aa2f-d301b7ae730d-logs\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:32 crc kubenswrapper[4716]: I1207 16:20:32.703040 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nvvr4\" (UniqueName: \"kubernetes.io/projected/e1dd089c-abb1-4317-aa2f-d301b7ae730d-kube-api-access-nvvr4\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:32 crc kubenswrapper[4716]: I1207 16:20:32.821767 4716 generic.go:334] "Generic (PLEG): container finished" podID="e1dd089c-abb1-4317-aa2f-d301b7ae730d" containerID="53eb298d61c184858cf42d228a84ba335e5e977399e58be6ed004d8ed973b4ba" exitCode=137 Dec 07 16:20:32 crc kubenswrapper[4716]: I1207 16:20:32.821901 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 07 16:20:32 crc kubenswrapper[4716]: I1207 16:20:32.830398 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 07 16:20:32 crc kubenswrapper[4716]: I1207 16:20:32.832953 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e1dd089c-abb1-4317-aa2f-d301b7ae730d","Type":"ContainerDied","Data":"53eb298d61c184858cf42d228a84ba335e5e977399e58be6ed004d8ed973b4ba"} Dec 07 16:20:32 crc kubenswrapper[4716]: I1207 16:20:32.833004 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e1dd089c-abb1-4317-aa2f-d301b7ae730d","Type":"ContainerDied","Data":"2021e1edd583374810e404464702da8eb2de94144082e419cd74ae66b0abbc18"} Dec 07 16:20:32 crc kubenswrapper[4716]: I1207 16:20:32.833023 4716 scope.go:117] "RemoveContainer" containerID="53eb298d61c184858cf42d228a84ba335e5e977399e58be6ed004d8ed973b4ba" Dec 07 16:20:32 crc kubenswrapper[4716]: I1207 16:20:32.841142 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 07 16:20:32 crc kubenswrapper[4716]: I1207 16:20:32.841408 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6873cce2-e68e-47ca-8ae8-7b612815a223","Type":"ContainerStarted","Data":"086e0a12638bc85b5222676a602a725a665991b0f030d5aaa151a6e50aff57eb"} Dec 07 16:20:32 crc kubenswrapper[4716]: I1207 16:20:32.851388 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0f287d7e-4b08-4c16-8a23-cf9e69225d3c","Type":"ContainerStarted","Data":"517f3baaccdc8d1550a27d23895cb8ea5d32f401cf0c0d611a87b98f81290277"} Dec 07 16:20:32 crc kubenswrapper[4716]: I1207 16:20:32.854047 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-xbk7g" event={"ID":"7ead59b4-dcff-45b8-8b1b-d5e449278acb","Type":"ContainerDied","Data":"eaeb45e5bf8e6a4649a47c98490e81e63cf929deb78275a0ea041aa2d4f2bb2b"} Dec 07 16:20:32 crc kubenswrapper[4716]: I1207 16:20:32.854154 4716 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eaeb45e5bf8e6a4649a47c98490e81e63cf929deb78275a0ea041aa2d4f2bb2b" Dec 07 16:20:32 crc kubenswrapper[4716]: I1207 16:20:32.854260 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-xbk7g" Dec 07 16:20:32 crc kubenswrapper[4716]: I1207 16:20:32.867579 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9f4f8c7b-38d4-4054-bb66-c51f32448567","Type":"ContainerStarted","Data":"3eede10a9208a4ad0d4ebfad62a311034601df16478fb1b5399e7040f12d65ad"} Dec 07 16:20:32 crc kubenswrapper[4716]: I1207 16:20:32.873261 4716 scope.go:117] "RemoveContainer" containerID="2d303dd7a108d3836287cc7f009cc60f84e17a4753a55a815b95ca9a76d71969" Dec 07 16:20:32 crc kubenswrapper[4716]: I1207 16:20:32.878042 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.878030653 podStartE2EDuration="3.878030653s" podCreationTimestamp="2025-12-07 16:20:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:20:32.8738773 +0000 UTC m=+1095.564162212" watchObservedRunningTime="2025-12-07 16:20:32.878030653 +0000 UTC m=+1095.568315565" Dec 07 16:20:32 crc kubenswrapper[4716]: I1207 16:20:32.898037 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.8980149539999998 podStartE2EDuration="3.898014954s" podCreationTimestamp="2025-12-07 16:20:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:20:32.893499012 +0000 UTC m=+1095.583783914" watchObservedRunningTime="2025-12-07 16:20:32.898014954 +0000 UTC m=+1095.588299866" Dec 07 16:20:32 crc kubenswrapper[4716]: I1207 16:20:32.964442 4716 scope.go:117] "RemoveContainer" containerID="53eb298d61c184858cf42d228a84ba335e5e977399e58be6ed004d8ed973b4ba" Dec 07 16:20:32 crc kubenswrapper[4716]: E1207 16:20:32.967898 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53eb298d61c184858cf42d228a84ba335e5e977399e58be6ed004d8ed973b4ba\": container with ID starting with 53eb298d61c184858cf42d228a84ba335e5e977399e58be6ed004d8ed973b4ba not found: ID does not exist" containerID="53eb298d61c184858cf42d228a84ba335e5e977399e58be6ed004d8ed973b4ba" Dec 07 16:20:32 crc kubenswrapper[4716]: I1207 16:20:32.967935 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53eb298d61c184858cf42d228a84ba335e5e977399e58be6ed004d8ed973b4ba"} err="failed to get container status \"53eb298d61c184858cf42d228a84ba335e5e977399e58be6ed004d8ed973b4ba\": rpc error: code = NotFound desc = could not find container \"53eb298d61c184858cf42d228a84ba335e5e977399e58be6ed004d8ed973b4ba\": container with ID starting with 53eb298d61c184858cf42d228a84ba335e5e977399e58be6ed004d8ed973b4ba not found: ID does not exist" Dec 07 16:20:32 crc kubenswrapper[4716]: I1207 16:20:32.967961 4716 scope.go:117] "RemoveContainer" containerID="2d303dd7a108d3836287cc7f009cc60f84e17a4753a55a815b95ca9a76d71969" Dec 07 16:20:32 crc kubenswrapper[4716]: E1207 16:20:32.970239 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d303dd7a108d3836287cc7f009cc60f84e17a4753a55a815b95ca9a76d71969\": container with ID starting with 2d303dd7a108d3836287cc7f009cc60f84e17a4753a55a815b95ca9a76d71969 not found: ID does not exist" containerID="2d303dd7a108d3836287cc7f009cc60f84e17a4753a55a815b95ca9a76d71969" Dec 07 16:20:32 crc kubenswrapper[4716]: I1207 16:20:32.970269 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d303dd7a108d3836287cc7f009cc60f84e17a4753a55a815b95ca9a76d71969"} err="failed to get container status \"2d303dd7a108d3836287cc7f009cc60f84e17a4753a55a815b95ca9a76d71969\": rpc error: code = NotFound desc = could not find container \"2d303dd7a108d3836287cc7f009cc60f84e17a4753a55a815b95ca9a76d71969\": container with ID starting with 2d303dd7a108d3836287cc7f009cc60f84e17a4753a55a815b95ca9a76d71969 not found: ID does not exist" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.006550 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.018021 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.024559 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 07 16:20:33 crc kubenswrapper[4716]: E1207 16:20:33.025160 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1dd089c-abb1-4317-aa2f-d301b7ae730d" containerName="cinder-api" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.025208 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1dd089c-abb1-4317-aa2f-d301b7ae730d" containerName="cinder-api" Dec 07 16:20:33 crc kubenswrapper[4716]: E1207 16:20:33.025237 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ead59b4-dcff-45b8-8b1b-d5e449278acb" containerName="mariadb-database-create" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.025245 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ead59b4-dcff-45b8-8b1b-d5e449278acb" containerName="mariadb-database-create" Dec 07 16:20:33 crc kubenswrapper[4716]: E1207 16:20:33.025262 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1dd089c-abb1-4317-aa2f-d301b7ae730d" containerName="cinder-api-log" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.025271 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1dd089c-abb1-4317-aa2f-d301b7ae730d" containerName="cinder-api-log" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.025499 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1dd089c-abb1-4317-aa2f-d301b7ae730d" containerName="cinder-api" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.025532 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ead59b4-dcff-45b8-8b1b-d5e449278acb" containerName="mariadb-database-create" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.025562 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1dd089c-abb1-4317-aa2f-d301b7ae730d" containerName="cinder-api-log" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.035232 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.035355 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.037384 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.037750 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.037828 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.214238 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7htn\" (UniqueName: \"kubernetes.io/projected/9237e175-c046-4f00-9535-474448e79076-kube-api-access-b7htn\") pod \"cinder-api-0\" (UID: \"9237e175-c046-4f00-9535-474448e79076\") " pod="openstack/cinder-api-0" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.214286 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9237e175-c046-4f00-9535-474448e79076-public-tls-certs\") pod \"cinder-api-0\" (UID: \"9237e175-c046-4f00-9535-474448e79076\") " pod="openstack/cinder-api-0" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.214306 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9237e175-c046-4f00-9535-474448e79076-config-data-custom\") pod \"cinder-api-0\" (UID: \"9237e175-c046-4f00-9535-474448e79076\") " pod="openstack/cinder-api-0" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.214325 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9237e175-c046-4f00-9535-474448e79076-logs\") pod \"cinder-api-0\" (UID: \"9237e175-c046-4f00-9535-474448e79076\") " pod="openstack/cinder-api-0" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.214374 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9237e175-c046-4f00-9535-474448e79076-etc-machine-id\") pod \"cinder-api-0\" (UID: \"9237e175-c046-4f00-9535-474448e79076\") " pod="openstack/cinder-api-0" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.214419 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9237e175-c046-4f00-9535-474448e79076-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"9237e175-c046-4f00-9535-474448e79076\") " pod="openstack/cinder-api-0" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.214456 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9237e175-c046-4f00-9535-474448e79076-scripts\") pod \"cinder-api-0\" (UID: \"9237e175-c046-4f00-9535-474448e79076\") " pod="openstack/cinder-api-0" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.214480 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9237e175-c046-4f00-9535-474448e79076-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"9237e175-c046-4f00-9535-474448e79076\") " pod="openstack/cinder-api-0" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.214495 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9237e175-c046-4f00-9535-474448e79076-config-data\") pod \"cinder-api-0\" (UID: \"9237e175-c046-4f00-9535-474448e79076\") " pod="openstack/cinder-api-0" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.317200 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9237e175-c046-4f00-9535-474448e79076-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"9237e175-c046-4f00-9535-474448e79076\") " pod="openstack/cinder-api-0" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.317485 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9237e175-c046-4f00-9535-474448e79076-scripts\") pod \"cinder-api-0\" (UID: \"9237e175-c046-4f00-9535-474448e79076\") " pod="openstack/cinder-api-0" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.317514 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9237e175-c046-4f00-9535-474448e79076-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"9237e175-c046-4f00-9535-474448e79076\") " pod="openstack/cinder-api-0" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.317529 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9237e175-c046-4f00-9535-474448e79076-config-data\") pod \"cinder-api-0\" (UID: \"9237e175-c046-4f00-9535-474448e79076\") " pod="openstack/cinder-api-0" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.317580 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7htn\" (UniqueName: \"kubernetes.io/projected/9237e175-c046-4f00-9535-474448e79076-kube-api-access-b7htn\") pod \"cinder-api-0\" (UID: \"9237e175-c046-4f00-9535-474448e79076\") " pod="openstack/cinder-api-0" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.317603 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9237e175-c046-4f00-9535-474448e79076-public-tls-certs\") pod \"cinder-api-0\" (UID: \"9237e175-c046-4f00-9535-474448e79076\") " pod="openstack/cinder-api-0" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.317621 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9237e175-c046-4f00-9535-474448e79076-config-data-custom\") pod \"cinder-api-0\" (UID: \"9237e175-c046-4f00-9535-474448e79076\") " pod="openstack/cinder-api-0" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.317639 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9237e175-c046-4f00-9535-474448e79076-logs\") pod \"cinder-api-0\" (UID: \"9237e175-c046-4f00-9535-474448e79076\") " pod="openstack/cinder-api-0" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.317682 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9237e175-c046-4f00-9535-474448e79076-etc-machine-id\") pod \"cinder-api-0\" (UID: \"9237e175-c046-4f00-9535-474448e79076\") " pod="openstack/cinder-api-0" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.317767 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9237e175-c046-4f00-9535-474448e79076-etc-machine-id\") pod \"cinder-api-0\" (UID: \"9237e175-c046-4f00-9535-474448e79076\") " pod="openstack/cinder-api-0" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.326101 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9237e175-c046-4f00-9535-474448e79076-config-data-custom\") pod \"cinder-api-0\" (UID: \"9237e175-c046-4f00-9535-474448e79076\") " pod="openstack/cinder-api-0" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.328359 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9237e175-c046-4f00-9535-474448e79076-logs\") pod \"cinder-api-0\" (UID: \"9237e175-c046-4f00-9535-474448e79076\") " pod="openstack/cinder-api-0" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.328831 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9237e175-c046-4f00-9535-474448e79076-scripts\") pod \"cinder-api-0\" (UID: \"9237e175-c046-4f00-9535-474448e79076\") " pod="openstack/cinder-api-0" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.345728 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9237e175-c046-4f00-9535-474448e79076-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"9237e175-c046-4f00-9535-474448e79076\") " pod="openstack/cinder-api-0" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.352976 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9237e175-c046-4f00-9535-474448e79076-config-data\") pod \"cinder-api-0\" (UID: \"9237e175-c046-4f00-9535-474448e79076\") " pod="openstack/cinder-api-0" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.359930 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7htn\" (UniqueName: \"kubernetes.io/projected/9237e175-c046-4f00-9535-474448e79076-kube-api-access-b7htn\") pod \"cinder-api-0\" (UID: \"9237e175-c046-4f00-9535-474448e79076\") " pod="openstack/cinder-api-0" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.365621 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9237e175-c046-4f00-9535-474448e79076-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"9237e175-c046-4f00-9535-474448e79076\") " pod="openstack/cinder-api-0" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.366726 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9237e175-c046-4f00-9535-474448e79076-public-tls-certs\") pod \"cinder-api-0\" (UID: \"9237e175-c046-4f00-9535-474448e79076\") " pod="openstack/cinder-api-0" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.452739 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-c495-account-create-update-cpqvg" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.540977 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-pz462" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.622910 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-5b89-account-create-update-w7tgb" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.623901 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a14d4b2b-4e79-4258-9d9f-46198eca8a41-operator-scripts\") pod \"a14d4b2b-4e79-4258-9d9f-46198eca8a41\" (UID: \"a14d4b2b-4e79-4258-9d9f-46198eca8a41\") " Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.623929 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bwbgs\" (UniqueName: \"kubernetes.io/projected/a14d4b2b-4e79-4258-9d9f-46198eca8a41-kube-api-access-bwbgs\") pod \"a14d4b2b-4e79-4258-9d9f-46198eca8a41\" (UID: \"a14d4b2b-4e79-4258-9d9f-46198eca8a41\") " Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.624174 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/73795e56-2afe-427c-a122-7d604c6ed134-operator-scripts\") pod \"73795e56-2afe-427c-a122-7d604c6ed134\" (UID: \"73795e56-2afe-427c-a122-7d604c6ed134\") " Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.624421 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vjw5h\" (UniqueName: \"kubernetes.io/projected/73795e56-2afe-427c-a122-7d604c6ed134-kube-api-access-vjw5h\") pod \"73795e56-2afe-427c-a122-7d604c6ed134\" (UID: \"73795e56-2afe-427c-a122-7d604c6ed134\") " Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.625789 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a14d4b2b-4e79-4258-9d9f-46198eca8a41-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a14d4b2b-4e79-4258-9d9f-46198eca8a41" (UID: "a14d4b2b-4e79-4258-9d9f-46198eca8a41"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.626553 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73795e56-2afe-427c-a122-7d604c6ed134-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "73795e56-2afe-427c-a122-7d604c6ed134" (UID: "73795e56-2afe-427c-a122-7d604c6ed134"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.628512 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-d9hjx" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.629920 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73795e56-2afe-427c-a122-7d604c6ed134-kube-api-access-vjw5h" (OuterVolumeSpecName: "kube-api-access-vjw5h") pod "73795e56-2afe-427c-a122-7d604c6ed134" (UID: "73795e56-2afe-427c-a122-7d604c6ed134"). InnerVolumeSpecName "kube-api-access-vjw5h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.630560 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a14d4b2b-4e79-4258-9d9f-46198eca8a41-kube-api-access-bwbgs" (OuterVolumeSpecName: "kube-api-access-bwbgs") pod "a14d4b2b-4e79-4258-9d9f-46198eca8a41" (UID: "a14d4b2b-4e79-4258-9d9f-46198eca8a41"). InnerVolumeSpecName "kube-api-access-bwbgs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.654610 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-33d4-account-create-update-2tg2q" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.666925 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.673666 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c088cd1b-83a8-4f05-b7db-11c8ffb59fab" path="/var/lib/kubelet/pods/c088cd1b-83a8-4f05-b7db-11c8ffb59fab/volumes" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.674746 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1dd089c-abb1-4317-aa2f-d301b7ae730d" path="/var/lib/kubelet/pods/e1dd089c-abb1-4317-aa2f-d301b7ae730d/volumes" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.728349 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wh9tb\" (UniqueName: \"kubernetes.io/projected/a8a5773f-5b23-400b-8911-1d2d40b9bfd0-kube-api-access-wh9tb\") pod \"a8a5773f-5b23-400b-8911-1d2d40b9bfd0\" (UID: \"a8a5773f-5b23-400b-8911-1d2d40b9bfd0\") " Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.728541 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3052dae6-ee76-44ac-bc5c-6dcb6d9995cf-operator-scripts\") pod \"3052dae6-ee76-44ac-bc5c-6dcb6d9995cf\" (UID: \"3052dae6-ee76-44ac-bc5c-6dcb6d9995cf\") " Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.728691 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a8a5773f-5b23-400b-8911-1d2d40b9bfd0-operator-scripts\") pod \"a8a5773f-5b23-400b-8911-1d2d40b9bfd0\" (UID: \"a8a5773f-5b23-400b-8911-1d2d40b9bfd0\") " Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.728747 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-djrjq\" (UniqueName: \"kubernetes.io/projected/3052dae6-ee76-44ac-bc5c-6dcb6d9995cf-kube-api-access-djrjq\") pod \"3052dae6-ee76-44ac-bc5c-6dcb6d9995cf\" (UID: \"3052dae6-ee76-44ac-bc5c-6dcb6d9995cf\") " Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.728920 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3052dae6-ee76-44ac-bc5c-6dcb6d9995cf-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3052dae6-ee76-44ac-bc5c-6dcb6d9995cf" (UID: "3052dae6-ee76-44ac-bc5c-6dcb6d9995cf"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.729191 4716 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a14d4b2b-4e79-4258-9d9f-46198eca8a41-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.729208 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bwbgs\" (UniqueName: \"kubernetes.io/projected/a14d4b2b-4e79-4258-9d9f-46198eca8a41-kube-api-access-bwbgs\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.729221 4716 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3052dae6-ee76-44ac-bc5c-6dcb6d9995cf-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.729232 4716 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/73795e56-2afe-427c-a122-7d604c6ed134-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.729244 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vjw5h\" (UniqueName: \"kubernetes.io/projected/73795e56-2afe-427c-a122-7d604c6ed134-kube-api-access-vjw5h\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.729401 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a8a5773f-5b23-400b-8911-1d2d40b9bfd0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a8a5773f-5b23-400b-8911-1d2d40b9bfd0" (UID: "a8a5773f-5b23-400b-8911-1d2d40b9bfd0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.734367 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3052dae6-ee76-44ac-bc5c-6dcb6d9995cf-kube-api-access-djrjq" (OuterVolumeSpecName: "kube-api-access-djrjq") pod "3052dae6-ee76-44ac-bc5c-6dcb6d9995cf" (UID: "3052dae6-ee76-44ac-bc5c-6dcb6d9995cf"). InnerVolumeSpecName "kube-api-access-djrjq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.734401 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8a5773f-5b23-400b-8911-1d2d40b9bfd0-kube-api-access-wh9tb" (OuterVolumeSpecName: "kube-api-access-wh9tb") pod "a8a5773f-5b23-400b-8911-1d2d40b9bfd0" (UID: "a8a5773f-5b23-400b-8911-1d2d40b9bfd0"). InnerVolumeSpecName "kube-api-access-wh9tb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.830584 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4647e69c-31d0-495e-a294-f9970aae2c32-operator-scripts\") pod \"4647e69c-31d0-495e-a294-f9970aae2c32\" (UID: \"4647e69c-31d0-495e-a294-f9970aae2c32\") " Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.830764 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vh98c\" (UniqueName: \"kubernetes.io/projected/4647e69c-31d0-495e-a294-f9970aae2c32-kube-api-access-vh98c\") pod \"4647e69c-31d0-495e-a294-f9970aae2c32\" (UID: \"4647e69c-31d0-495e-a294-f9970aae2c32\") " Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.831948 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4647e69c-31d0-495e-a294-f9970aae2c32-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4647e69c-31d0-495e-a294-f9970aae2c32" (UID: "4647e69c-31d0-495e-a294-f9970aae2c32"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.832508 4716 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4647e69c-31d0-495e-a294-f9970aae2c32-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.832550 4716 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a8a5773f-5b23-400b-8911-1d2d40b9bfd0-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.832563 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-djrjq\" (UniqueName: \"kubernetes.io/projected/3052dae6-ee76-44ac-bc5c-6dcb6d9995cf-kube-api-access-djrjq\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.832591 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wh9tb\" (UniqueName: \"kubernetes.io/projected/a8a5773f-5b23-400b-8911-1d2d40b9bfd0-kube-api-access-wh9tb\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.838797 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4647e69c-31d0-495e-a294-f9970aae2c32-kube-api-access-vh98c" (OuterVolumeSpecName: "kube-api-access-vh98c") pod "4647e69c-31d0-495e-a294-f9970aae2c32" (UID: "4647e69c-31d0-495e-a294-f9970aae2c32"). InnerVolumeSpecName "kube-api-access-vh98c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.893372 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"7953b7b7-a0c6-452c-868a-dcc4a1bda1fa","Type":"ContainerStarted","Data":"96b05370e41dfaca784602b891ceab12ea87e8d750d8936b5715764e2be53bac"} Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.893424 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"7953b7b7-a0c6-452c-868a-dcc4a1bda1fa","Type":"ContainerStarted","Data":"563c57073d2dfa31a4134eac1dbcae94ccaf460763c0fa35f43345bd9dfb052f"} Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.894833 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.908636 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-d9hjx" event={"ID":"a8a5773f-5b23-400b-8911-1d2d40b9bfd0","Type":"ContainerDied","Data":"4c16c36568b5d98c100b2910a47f5ed901cb5394863acbfec17644d823ba62eb"} Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.908683 4716 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4c16c36568b5d98c100b2910a47f5ed901cb5394863acbfec17644d823ba62eb" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.908747 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-d9hjx" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.917892 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6873cce2-e68e-47ca-8ae8-7b612815a223","Type":"ContainerStarted","Data":"81d54b420e9a0506bc7d4efba865f53f5c07ac678546167d73a039558c995086"} Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.921639 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.445240648 podStartE2EDuration="2.921620594s" podCreationTimestamp="2025-12-07 16:20:31 +0000 UTC" firstStartedPulling="2025-12-07 16:20:32.817965457 +0000 UTC m=+1095.508250369" lastFinishedPulling="2025-12-07 16:20:33.294345403 +0000 UTC m=+1095.984630315" observedRunningTime="2025-12-07 16:20:33.909502696 +0000 UTC m=+1096.599787608" watchObservedRunningTime="2025-12-07 16:20:33.921620594 +0000 UTC m=+1096.611905506" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.921767 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-5b89-account-create-update-w7tgb" event={"ID":"3052dae6-ee76-44ac-bc5c-6dcb6d9995cf","Type":"ContainerDied","Data":"743f1108a3f30863c154c0f4a97117edde65287b6ee02ad37d6dcfc8bf7296e4"} Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.922199 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-5b89-account-create-update-w7tgb" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.922533 4716 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="743f1108a3f30863c154c0f4a97117edde65287b6ee02ad37d6dcfc8bf7296e4" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.931349 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-pz462" event={"ID":"a14d4b2b-4e79-4258-9d9f-46198eca8a41","Type":"ContainerDied","Data":"1b5af05ebc40d55cb25fde5aa45b802ff769ba95dd1ebd29f7afee1bdf69adb4"} Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.931382 4716 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1b5af05ebc40d55cb25fde5aa45b802ff769ba95dd1ebd29f7afee1bdf69adb4" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.931447 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-pz462" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.933677 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vh98c\" (UniqueName: \"kubernetes.io/projected/4647e69c-31d0-495e-a294-f9970aae2c32-kube-api-access-vh98c\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.938569 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-c495-account-create-update-cpqvg" event={"ID":"73795e56-2afe-427c-a122-7d604c6ed134","Type":"ContainerDied","Data":"e238b8abff70cca389d44c3f2f52dfe683051c07f6d961bcf2d8c5fc84af2f82"} Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.938608 4716 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e238b8abff70cca389d44c3f2f52dfe683051c07f6d961bcf2d8c5fc84af2f82" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.938693 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-c495-account-create-update-cpqvg" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.943595 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-33d4-account-create-update-2tg2q" event={"ID":"4647e69c-31d0-495e-a294-f9970aae2c32","Type":"ContainerDied","Data":"03e061a83e97940fa65de350a932fb9e219e83dd182f5a483272e6d896518a13"} Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.943637 4716 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="03e061a83e97940fa65de350a932fb9e219e83dd182f5a483272e6d896518a13" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.943700 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-33d4-account-create-update-2tg2q" Dec 07 16:20:33 crc kubenswrapper[4716]: I1207 16:20:33.967306 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 07 16:20:34 crc kubenswrapper[4716]: I1207 16:20:34.974139 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6873cce2-e68e-47ca-8ae8-7b612815a223","Type":"ContainerStarted","Data":"cb8b887082e5b10ffc83cc548108c87e4c5071bbffd61194184f3b3a53de524d"} Dec 07 16:20:34 crc kubenswrapper[4716]: I1207 16:20:34.975604 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 07 16:20:34 crc kubenswrapper[4716]: I1207 16:20:34.975319 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6873cce2-e68e-47ca-8ae8-7b612815a223" containerName="sg-core" containerID="cri-o://81d54b420e9a0506bc7d4efba865f53f5c07ac678546167d73a039558c995086" gracePeriod=30 Dec 07 16:20:34 crc kubenswrapper[4716]: I1207 16:20:34.975380 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6873cce2-e68e-47ca-8ae8-7b612815a223" containerName="ceilometer-notification-agent" containerID="cri-o://086e0a12638bc85b5222676a602a725a665991b0f030d5aaa151a6e50aff57eb" gracePeriod=30 Dec 07 16:20:34 crc kubenswrapper[4716]: I1207 16:20:34.975265 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6873cce2-e68e-47ca-8ae8-7b612815a223" containerName="ceilometer-central-agent" containerID="cri-o://ef45036db19a07ce203b9018d7f545d5aaaeb31f8ee89c8dd4a93b441821fa1a" gracePeriod=30 Dec 07 16:20:34 crc kubenswrapper[4716]: I1207 16:20:34.975338 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6873cce2-e68e-47ca-8ae8-7b612815a223" containerName="proxy-httpd" containerID="cri-o://cb8b887082e5b10ffc83cc548108c87e4c5071bbffd61194184f3b3a53de524d" gracePeriod=30 Dec 07 16:20:34 crc kubenswrapper[4716]: I1207 16:20:34.985876 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"9237e175-c046-4f00-9535-474448e79076","Type":"ContainerStarted","Data":"035d85b994e52079bf1e3bf033bf6eede7507853464a261e9b72cf4664299f1b"} Dec 07 16:20:34 crc kubenswrapper[4716]: I1207 16:20:34.986069 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"9237e175-c046-4f00-9535-474448e79076","Type":"ContainerStarted","Data":"926776112dff026d1d8b15e92b3cac3d140dfd042744c6edff409d5e427e682a"} Dec 07 16:20:35 crc kubenswrapper[4716]: I1207 16:20:34.998005 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.226711451 podStartE2EDuration="6.997953011s" podCreationTimestamp="2025-12-07 16:20:28 +0000 UTC" firstStartedPulling="2025-12-07 16:20:30.223117453 +0000 UTC m=+1092.913402365" lastFinishedPulling="2025-12-07 16:20:33.994359013 +0000 UTC m=+1096.684643925" observedRunningTime="2025-12-07 16:20:34.996479251 +0000 UTC m=+1097.686764183" watchObservedRunningTime="2025-12-07 16:20:34.997953011 +0000 UTC m=+1097.688237923" Dec 07 16:20:35 crc kubenswrapper[4716]: I1207 16:20:35.995059 4716 generic.go:334] "Generic (PLEG): container finished" podID="6873cce2-e68e-47ca-8ae8-7b612815a223" containerID="cb8b887082e5b10ffc83cc548108c87e4c5071bbffd61194184f3b3a53de524d" exitCode=0 Dec 07 16:20:35 crc kubenswrapper[4716]: I1207 16:20:35.995394 4716 generic.go:334] "Generic (PLEG): container finished" podID="6873cce2-e68e-47ca-8ae8-7b612815a223" containerID="81d54b420e9a0506bc7d4efba865f53f5c07ac678546167d73a039558c995086" exitCode=2 Dec 07 16:20:35 crc kubenswrapper[4716]: I1207 16:20:35.995403 4716 generic.go:334] "Generic (PLEG): container finished" podID="6873cce2-e68e-47ca-8ae8-7b612815a223" containerID="086e0a12638bc85b5222676a602a725a665991b0f030d5aaa151a6e50aff57eb" exitCode=0 Dec 07 16:20:35 crc kubenswrapper[4716]: I1207 16:20:35.995440 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6873cce2-e68e-47ca-8ae8-7b612815a223","Type":"ContainerDied","Data":"cb8b887082e5b10ffc83cc548108c87e4c5071bbffd61194184f3b3a53de524d"} Dec 07 16:20:35 crc kubenswrapper[4716]: I1207 16:20:35.995470 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6873cce2-e68e-47ca-8ae8-7b612815a223","Type":"ContainerDied","Data":"81d54b420e9a0506bc7d4efba865f53f5c07ac678546167d73a039558c995086"} Dec 07 16:20:35 crc kubenswrapper[4716]: I1207 16:20:35.995481 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6873cce2-e68e-47ca-8ae8-7b612815a223","Type":"ContainerDied","Data":"086e0a12638bc85b5222676a602a725a665991b0f030d5aaa151a6e50aff57eb"} Dec 07 16:20:36 crc kubenswrapper[4716]: I1207 16:20:36.000962 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"9237e175-c046-4f00-9535-474448e79076","Type":"ContainerStarted","Data":"fef27f9b78869d70e20fecc88992d591a0951f2b3b38193d0750e6f69a6adffb"} Dec 07 16:20:36 crc kubenswrapper[4716]: I1207 16:20:36.001050 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 07 16:20:36 crc kubenswrapper[4716]: I1207 16:20:36.031275 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.031248773 podStartE2EDuration="4.031248773s" podCreationTimestamp="2025-12-07 16:20:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:20:36.023976596 +0000 UTC m=+1098.714261528" watchObservedRunningTime="2025-12-07 16:20:36.031248773 +0000 UTC m=+1098.721533705" Dec 07 16:20:36 crc kubenswrapper[4716]: I1207 16:20:36.445797 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-86c7567d4-99rx9" Dec 07 16:20:36 crc kubenswrapper[4716]: I1207 16:20:36.533419 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c291e114-7940-46bc-91d3-a8fa256549ff-horizon-secret-key\") pod \"c291e114-7940-46bc-91d3-a8fa256549ff\" (UID: \"c291e114-7940-46bc-91d3-a8fa256549ff\") " Dec 07 16:20:36 crc kubenswrapper[4716]: I1207 16:20:36.533788 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c291e114-7940-46bc-91d3-a8fa256549ff-combined-ca-bundle\") pod \"c291e114-7940-46bc-91d3-a8fa256549ff\" (UID: \"c291e114-7940-46bc-91d3-a8fa256549ff\") " Dec 07 16:20:36 crc kubenswrapper[4716]: I1207 16:20:36.533832 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/c291e114-7940-46bc-91d3-a8fa256549ff-horizon-tls-certs\") pod \"c291e114-7940-46bc-91d3-a8fa256549ff\" (UID: \"c291e114-7940-46bc-91d3-a8fa256549ff\") " Dec 07 16:20:36 crc kubenswrapper[4716]: I1207 16:20:36.533852 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c291e114-7940-46bc-91d3-a8fa256549ff-logs\") pod \"c291e114-7940-46bc-91d3-a8fa256549ff\" (UID: \"c291e114-7940-46bc-91d3-a8fa256549ff\") " Dec 07 16:20:36 crc kubenswrapper[4716]: I1207 16:20:36.533907 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c291e114-7940-46bc-91d3-a8fa256549ff-config-data\") pod \"c291e114-7940-46bc-91d3-a8fa256549ff\" (UID: \"c291e114-7940-46bc-91d3-a8fa256549ff\") " Dec 07 16:20:36 crc kubenswrapper[4716]: I1207 16:20:36.533998 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c291e114-7940-46bc-91d3-a8fa256549ff-scripts\") pod \"c291e114-7940-46bc-91d3-a8fa256549ff\" (UID: \"c291e114-7940-46bc-91d3-a8fa256549ff\") " Dec 07 16:20:36 crc kubenswrapper[4716]: I1207 16:20:36.534058 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8xcrh\" (UniqueName: \"kubernetes.io/projected/c291e114-7940-46bc-91d3-a8fa256549ff-kube-api-access-8xcrh\") pod \"c291e114-7940-46bc-91d3-a8fa256549ff\" (UID: \"c291e114-7940-46bc-91d3-a8fa256549ff\") " Dec 07 16:20:36 crc kubenswrapper[4716]: I1207 16:20:36.536069 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c291e114-7940-46bc-91d3-a8fa256549ff-logs" (OuterVolumeSpecName: "logs") pod "c291e114-7940-46bc-91d3-a8fa256549ff" (UID: "c291e114-7940-46bc-91d3-a8fa256549ff"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:20:36 crc kubenswrapper[4716]: I1207 16:20:36.552760 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c291e114-7940-46bc-91d3-a8fa256549ff-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "c291e114-7940-46bc-91d3-a8fa256549ff" (UID: "c291e114-7940-46bc-91d3-a8fa256549ff"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:20:36 crc kubenswrapper[4716]: I1207 16:20:36.554433 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c291e114-7940-46bc-91d3-a8fa256549ff-kube-api-access-8xcrh" (OuterVolumeSpecName: "kube-api-access-8xcrh") pod "c291e114-7940-46bc-91d3-a8fa256549ff" (UID: "c291e114-7940-46bc-91d3-a8fa256549ff"). InnerVolumeSpecName "kube-api-access-8xcrh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:20:36 crc kubenswrapper[4716]: I1207 16:20:36.583412 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c291e114-7940-46bc-91d3-a8fa256549ff-scripts" (OuterVolumeSpecName: "scripts") pod "c291e114-7940-46bc-91d3-a8fa256549ff" (UID: "c291e114-7940-46bc-91d3-a8fa256549ff"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:20:36 crc kubenswrapper[4716]: I1207 16:20:36.593927 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c291e114-7940-46bc-91d3-a8fa256549ff-config-data" (OuterVolumeSpecName: "config-data") pod "c291e114-7940-46bc-91d3-a8fa256549ff" (UID: "c291e114-7940-46bc-91d3-a8fa256549ff"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:20:36 crc kubenswrapper[4716]: I1207 16:20:36.601490 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c291e114-7940-46bc-91d3-a8fa256549ff-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c291e114-7940-46bc-91d3-a8fa256549ff" (UID: "c291e114-7940-46bc-91d3-a8fa256549ff"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:20:36 crc kubenswrapper[4716]: I1207 16:20:36.611842 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-5488d7fd99-pwmf5" Dec 07 16:20:36 crc kubenswrapper[4716]: I1207 16:20:36.627026 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c291e114-7940-46bc-91d3-a8fa256549ff-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "c291e114-7940-46bc-91d3-a8fa256549ff" (UID: "c291e114-7940-46bc-91d3-a8fa256549ff"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:20:36 crc kubenswrapper[4716]: I1207 16:20:36.641775 4716 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c291e114-7940-46bc-91d3-a8fa256549ff-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:36 crc kubenswrapper[4716]: I1207 16:20:36.641817 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8xcrh\" (UniqueName: \"kubernetes.io/projected/c291e114-7940-46bc-91d3-a8fa256549ff-kube-api-access-8xcrh\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:36 crc kubenswrapper[4716]: I1207 16:20:36.641830 4716 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c291e114-7940-46bc-91d3-a8fa256549ff-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:36 crc kubenswrapper[4716]: I1207 16:20:36.641842 4716 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c291e114-7940-46bc-91d3-a8fa256549ff-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:36 crc kubenswrapper[4716]: I1207 16:20:36.641854 4716 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/c291e114-7940-46bc-91d3-a8fa256549ff-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:36 crc kubenswrapper[4716]: I1207 16:20:36.641865 4716 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c291e114-7940-46bc-91d3-a8fa256549ff-logs\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:36 crc kubenswrapper[4716]: I1207 16:20:36.641876 4716 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c291e114-7940-46bc-91d3-a8fa256549ff-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:36 crc kubenswrapper[4716]: I1207 16:20:36.665571 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 07 16:20:36 crc kubenswrapper[4716]: I1207 16:20:36.742684 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6873cce2-e68e-47ca-8ae8-7b612815a223-scripts\") pod \"6873cce2-e68e-47ca-8ae8-7b612815a223\" (UID: \"6873cce2-e68e-47ca-8ae8-7b612815a223\") " Dec 07 16:20:36 crc kubenswrapper[4716]: I1207 16:20:36.742751 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6873cce2-e68e-47ca-8ae8-7b612815a223-run-httpd\") pod \"6873cce2-e68e-47ca-8ae8-7b612815a223\" (UID: \"6873cce2-e68e-47ca-8ae8-7b612815a223\") " Dec 07 16:20:36 crc kubenswrapper[4716]: I1207 16:20:36.742791 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6873cce2-e68e-47ca-8ae8-7b612815a223-combined-ca-bundle\") pod \"6873cce2-e68e-47ca-8ae8-7b612815a223\" (UID: \"6873cce2-e68e-47ca-8ae8-7b612815a223\") " Dec 07 16:20:36 crc kubenswrapper[4716]: I1207 16:20:36.742815 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6873cce2-e68e-47ca-8ae8-7b612815a223-config-data\") pod \"6873cce2-e68e-47ca-8ae8-7b612815a223\" (UID: \"6873cce2-e68e-47ca-8ae8-7b612815a223\") " Dec 07 16:20:36 crc kubenswrapper[4716]: I1207 16:20:36.742854 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6873cce2-e68e-47ca-8ae8-7b612815a223-sg-core-conf-yaml\") pod \"6873cce2-e68e-47ca-8ae8-7b612815a223\" (UID: \"6873cce2-e68e-47ca-8ae8-7b612815a223\") " Dec 07 16:20:36 crc kubenswrapper[4716]: I1207 16:20:36.742892 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6873cce2-e68e-47ca-8ae8-7b612815a223-log-httpd\") pod \"6873cce2-e68e-47ca-8ae8-7b612815a223\" (UID: \"6873cce2-e68e-47ca-8ae8-7b612815a223\") " Dec 07 16:20:36 crc kubenswrapper[4716]: I1207 16:20:36.743191 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vlj99\" (UniqueName: \"kubernetes.io/projected/6873cce2-e68e-47ca-8ae8-7b612815a223-kube-api-access-vlj99\") pod \"6873cce2-e68e-47ca-8ae8-7b612815a223\" (UID: \"6873cce2-e68e-47ca-8ae8-7b612815a223\") " Dec 07 16:20:36 crc kubenswrapper[4716]: I1207 16:20:36.744442 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6873cce2-e68e-47ca-8ae8-7b612815a223-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "6873cce2-e68e-47ca-8ae8-7b612815a223" (UID: "6873cce2-e68e-47ca-8ae8-7b612815a223"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:20:36 crc kubenswrapper[4716]: I1207 16:20:36.744721 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6873cce2-e68e-47ca-8ae8-7b612815a223-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "6873cce2-e68e-47ca-8ae8-7b612815a223" (UID: "6873cce2-e68e-47ca-8ae8-7b612815a223"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:20:36 crc kubenswrapper[4716]: I1207 16:20:36.747455 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6873cce2-e68e-47ca-8ae8-7b612815a223-scripts" (OuterVolumeSpecName: "scripts") pod "6873cce2-e68e-47ca-8ae8-7b612815a223" (UID: "6873cce2-e68e-47ca-8ae8-7b612815a223"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:20:36 crc kubenswrapper[4716]: I1207 16:20:36.748287 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6873cce2-e68e-47ca-8ae8-7b612815a223-kube-api-access-vlj99" (OuterVolumeSpecName: "kube-api-access-vlj99") pod "6873cce2-e68e-47ca-8ae8-7b612815a223" (UID: "6873cce2-e68e-47ca-8ae8-7b612815a223"). InnerVolumeSpecName "kube-api-access-vlj99". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:20:36 crc kubenswrapper[4716]: I1207 16:20:36.782602 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6873cce2-e68e-47ca-8ae8-7b612815a223-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "6873cce2-e68e-47ca-8ae8-7b612815a223" (UID: "6873cce2-e68e-47ca-8ae8-7b612815a223"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:20:36 crc kubenswrapper[4716]: I1207 16:20:36.812174 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6873cce2-e68e-47ca-8ae8-7b612815a223-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6873cce2-e68e-47ca-8ae8-7b612815a223" (UID: "6873cce2-e68e-47ca-8ae8-7b612815a223"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:20:36 crc kubenswrapper[4716]: I1207 16:20:36.844886 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vlj99\" (UniqueName: \"kubernetes.io/projected/6873cce2-e68e-47ca-8ae8-7b612815a223-kube-api-access-vlj99\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:36 crc kubenswrapper[4716]: I1207 16:20:36.844912 4716 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6873cce2-e68e-47ca-8ae8-7b612815a223-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:36 crc kubenswrapper[4716]: I1207 16:20:36.844921 4716 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6873cce2-e68e-47ca-8ae8-7b612815a223-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:36 crc kubenswrapper[4716]: I1207 16:20:36.844930 4716 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6873cce2-e68e-47ca-8ae8-7b612815a223-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:36 crc kubenswrapper[4716]: I1207 16:20:36.844940 4716 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6873cce2-e68e-47ca-8ae8-7b612815a223-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:36 crc kubenswrapper[4716]: I1207 16:20:36.844963 4716 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6873cce2-e68e-47ca-8ae8-7b612815a223-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:36 crc kubenswrapper[4716]: I1207 16:20:36.846766 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6873cce2-e68e-47ca-8ae8-7b612815a223-config-data" (OuterVolumeSpecName: "config-data") pod "6873cce2-e68e-47ca-8ae8-7b612815a223" (UID: "6873cce2-e68e-47ca-8ae8-7b612815a223"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:20:36 crc kubenswrapper[4716]: I1207 16:20:36.946522 4716 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6873cce2-e68e-47ca-8ae8-7b612815a223-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.009298 4716 generic.go:334] "Generic (PLEG): container finished" podID="c291e114-7940-46bc-91d3-a8fa256549ff" containerID="00893cd59537178a447a7d7ca931b0f9922afd0ddd98cb384ea1435f3b09a2f9" exitCode=137 Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.009353 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-86c7567d4-99rx9" event={"ID":"c291e114-7940-46bc-91d3-a8fa256549ff","Type":"ContainerDied","Data":"00893cd59537178a447a7d7ca931b0f9922afd0ddd98cb384ea1435f3b09a2f9"} Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.009384 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-86c7567d4-99rx9" event={"ID":"c291e114-7940-46bc-91d3-a8fa256549ff","Type":"ContainerDied","Data":"3738b4121edde6afb7201317a053ff7656912527a99eb08cd861bd6904dd5dc4"} Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.009401 4716 scope.go:117] "RemoveContainer" containerID="bc89298b4c5f29666ca1176ee3a6613d8d2303c5623d2bec2695ed37ae7b6619" Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.009511 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-86c7567d4-99rx9" Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.018386 4716 generic.go:334] "Generic (PLEG): container finished" podID="6873cce2-e68e-47ca-8ae8-7b612815a223" containerID="ef45036db19a07ce203b9018d7f545d5aaaeb31f8ee89c8dd4a93b441821fa1a" exitCode=0 Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.018478 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6873cce2-e68e-47ca-8ae8-7b612815a223","Type":"ContainerDied","Data":"ef45036db19a07ce203b9018d7f545d5aaaeb31f8ee89c8dd4a93b441821fa1a"} Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.018528 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6873cce2-e68e-47ca-8ae8-7b612815a223","Type":"ContainerDied","Data":"8d03bbfd8f4f7e0d28b7cba869d384348e81879a56ac18e7ca03abde08694745"} Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.018492 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.045896 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-86c7567d4-99rx9"] Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.053667 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-86c7567d4-99rx9"] Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.065843 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.070999 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.096033 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 07 16:20:37 crc kubenswrapper[4716]: E1207 16:20:37.096613 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c291e114-7940-46bc-91d3-a8fa256549ff" containerName="horizon-log" Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.096678 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="c291e114-7940-46bc-91d3-a8fa256549ff" containerName="horizon-log" Dec 07 16:20:37 crc kubenswrapper[4716]: E1207 16:20:37.096771 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8a5773f-5b23-400b-8911-1d2d40b9bfd0" containerName="mariadb-database-create" Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.096823 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8a5773f-5b23-400b-8911-1d2d40b9bfd0" containerName="mariadb-database-create" Dec 07 16:20:37 crc kubenswrapper[4716]: E1207 16:20:37.096887 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c291e114-7940-46bc-91d3-a8fa256549ff" containerName="horizon" Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.096974 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="c291e114-7940-46bc-91d3-a8fa256549ff" containerName="horizon" Dec 07 16:20:37 crc kubenswrapper[4716]: E1207 16:20:37.097035 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a14d4b2b-4e79-4258-9d9f-46198eca8a41" containerName="mariadb-database-create" Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.098829 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="a14d4b2b-4e79-4258-9d9f-46198eca8a41" containerName="mariadb-database-create" Dec 07 16:20:37 crc kubenswrapper[4716]: E1207 16:20:37.098917 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73795e56-2afe-427c-a122-7d604c6ed134" containerName="mariadb-account-create-update" Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.098991 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="73795e56-2afe-427c-a122-7d604c6ed134" containerName="mariadb-account-create-update" Dec 07 16:20:37 crc kubenswrapper[4716]: E1207 16:20:37.099047 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3052dae6-ee76-44ac-bc5c-6dcb6d9995cf" containerName="mariadb-account-create-update" Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.099110 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="3052dae6-ee76-44ac-bc5c-6dcb6d9995cf" containerName="mariadb-account-create-update" Dec 07 16:20:37 crc kubenswrapper[4716]: E1207 16:20:37.099185 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6873cce2-e68e-47ca-8ae8-7b612815a223" containerName="proxy-httpd" Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.099238 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="6873cce2-e68e-47ca-8ae8-7b612815a223" containerName="proxy-httpd" Dec 07 16:20:37 crc kubenswrapper[4716]: E1207 16:20:37.099295 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6873cce2-e68e-47ca-8ae8-7b612815a223" containerName="ceilometer-notification-agent" Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.099348 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="6873cce2-e68e-47ca-8ae8-7b612815a223" containerName="ceilometer-notification-agent" Dec 07 16:20:37 crc kubenswrapper[4716]: E1207 16:20:37.099417 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6873cce2-e68e-47ca-8ae8-7b612815a223" containerName="sg-core" Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.099478 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="6873cce2-e68e-47ca-8ae8-7b612815a223" containerName="sg-core" Dec 07 16:20:37 crc kubenswrapper[4716]: E1207 16:20:37.099544 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6873cce2-e68e-47ca-8ae8-7b612815a223" containerName="ceilometer-central-agent" Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.099592 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="6873cce2-e68e-47ca-8ae8-7b612815a223" containerName="ceilometer-central-agent" Dec 07 16:20:37 crc kubenswrapper[4716]: E1207 16:20:37.099643 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4647e69c-31d0-495e-a294-f9970aae2c32" containerName="mariadb-account-create-update" Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.099708 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="4647e69c-31d0-495e-a294-f9970aae2c32" containerName="mariadb-account-create-update" Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.101054 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="3052dae6-ee76-44ac-bc5c-6dcb6d9995cf" containerName="mariadb-account-create-update" Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.104823 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8a5773f-5b23-400b-8911-1d2d40b9bfd0" containerName="mariadb-database-create" Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.104907 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="6873cce2-e68e-47ca-8ae8-7b612815a223" containerName="ceilometer-central-agent" Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.104989 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="6873cce2-e68e-47ca-8ae8-7b612815a223" containerName="proxy-httpd" Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.105045 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="c291e114-7940-46bc-91d3-a8fa256549ff" containerName="horizon" Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.105122 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="6873cce2-e68e-47ca-8ae8-7b612815a223" containerName="ceilometer-notification-agent" Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.105192 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="4647e69c-31d0-495e-a294-f9970aae2c32" containerName="mariadb-account-create-update" Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.105248 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="c291e114-7940-46bc-91d3-a8fa256549ff" containerName="horizon-log" Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.105305 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="6873cce2-e68e-47ca-8ae8-7b612815a223" containerName="sg-core" Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.105368 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="a14d4b2b-4e79-4258-9d9f-46198eca8a41" containerName="mariadb-database-create" Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.105422 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="73795e56-2afe-427c-a122-7d604c6ed134" containerName="mariadb-account-create-update" Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.107164 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.110277 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.110485 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.110713 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.113451 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.160151 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0f39334f-2b6b-4362-9f8b-378c5859c786-log-httpd\") pod \"ceilometer-0\" (UID: \"0f39334f-2b6b-4362-9f8b-378c5859c786\") " pod="openstack/ceilometer-0" Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.160244 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0f39334f-2b6b-4362-9f8b-378c5859c786-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0f39334f-2b6b-4362-9f8b-378c5859c786\") " pod="openstack/ceilometer-0" Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.160271 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f39334f-2b6b-4362-9f8b-378c5859c786-scripts\") pod \"ceilometer-0\" (UID: \"0f39334f-2b6b-4362-9f8b-378c5859c786\") " pod="openstack/ceilometer-0" Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.160465 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0f39334f-2b6b-4362-9f8b-378c5859c786-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0f39334f-2b6b-4362-9f8b-378c5859c786\") " pod="openstack/ceilometer-0" Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.160498 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f39334f-2b6b-4362-9f8b-378c5859c786-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0f39334f-2b6b-4362-9f8b-378c5859c786\") " pod="openstack/ceilometer-0" Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.160550 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0f39334f-2b6b-4362-9f8b-378c5859c786-run-httpd\") pod \"ceilometer-0\" (UID: \"0f39334f-2b6b-4362-9f8b-378c5859c786\") " pod="openstack/ceilometer-0" Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.160571 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hsxz9\" (UniqueName: \"kubernetes.io/projected/0f39334f-2b6b-4362-9f8b-378c5859c786-kube-api-access-hsxz9\") pod \"ceilometer-0\" (UID: \"0f39334f-2b6b-4362-9f8b-378c5859c786\") " pod="openstack/ceilometer-0" Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.160589 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f39334f-2b6b-4362-9f8b-378c5859c786-config-data\") pod \"ceilometer-0\" (UID: \"0f39334f-2b6b-4362-9f8b-378c5859c786\") " pod="openstack/ceilometer-0" Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.224326 4716 scope.go:117] "RemoveContainer" containerID="00893cd59537178a447a7d7ca931b0f9922afd0ddd98cb384ea1435f3b09a2f9" Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.244028 4716 scope.go:117] "RemoveContainer" containerID="bc89298b4c5f29666ca1176ee3a6613d8d2303c5623d2bec2695ed37ae7b6619" Dec 07 16:20:37 crc kubenswrapper[4716]: E1207 16:20:37.244573 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc89298b4c5f29666ca1176ee3a6613d8d2303c5623d2bec2695ed37ae7b6619\": container with ID starting with bc89298b4c5f29666ca1176ee3a6613d8d2303c5623d2bec2695ed37ae7b6619 not found: ID does not exist" containerID="bc89298b4c5f29666ca1176ee3a6613d8d2303c5623d2bec2695ed37ae7b6619" Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.244667 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc89298b4c5f29666ca1176ee3a6613d8d2303c5623d2bec2695ed37ae7b6619"} err="failed to get container status \"bc89298b4c5f29666ca1176ee3a6613d8d2303c5623d2bec2695ed37ae7b6619\": rpc error: code = NotFound desc = could not find container \"bc89298b4c5f29666ca1176ee3a6613d8d2303c5623d2bec2695ed37ae7b6619\": container with ID starting with bc89298b4c5f29666ca1176ee3a6613d8d2303c5623d2bec2695ed37ae7b6619 not found: ID does not exist" Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.244746 4716 scope.go:117] "RemoveContainer" containerID="00893cd59537178a447a7d7ca931b0f9922afd0ddd98cb384ea1435f3b09a2f9" Dec 07 16:20:37 crc kubenswrapper[4716]: E1207 16:20:37.245207 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00893cd59537178a447a7d7ca931b0f9922afd0ddd98cb384ea1435f3b09a2f9\": container with ID starting with 00893cd59537178a447a7d7ca931b0f9922afd0ddd98cb384ea1435f3b09a2f9 not found: ID does not exist" containerID="00893cd59537178a447a7d7ca931b0f9922afd0ddd98cb384ea1435f3b09a2f9" Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.245321 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00893cd59537178a447a7d7ca931b0f9922afd0ddd98cb384ea1435f3b09a2f9"} err="failed to get container status \"00893cd59537178a447a7d7ca931b0f9922afd0ddd98cb384ea1435f3b09a2f9\": rpc error: code = NotFound desc = could not find container \"00893cd59537178a447a7d7ca931b0f9922afd0ddd98cb384ea1435f3b09a2f9\": container with ID starting with 00893cd59537178a447a7d7ca931b0f9922afd0ddd98cb384ea1435f3b09a2f9 not found: ID does not exist" Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.245353 4716 scope.go:117] "RemoveContainer" containerID="cb8b887082e5b10ffc83cc548108c87e4c5071bbffd61194184f3b3a53de524d" Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.261791 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0f39334f-2b6b-4362-9f8b-378c5859c786-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0f39334f-2b6b-4362-9f8b-378c5859c786\") " pod="openstack/ceilometer-0" Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.261945 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f39334f-2b6b-4362-9f8b-378c5859c786-scripts\") pod \"ceilometer-0\" (UID: \"0f39334f-2b6b-4362-9f8b-378c5859c786\") " pod="openstack/ceilometer-0" Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.262051 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0f39334f-2b6b-4362-9f8b-378c5859c786-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0f39334f-2b6b-4362-9f8b-378c5859c786\") " pod="openstack/ceilometer-0" Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.262389 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f39334f-2b6b-4362-9f8b-378c5859c786-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0f39334f-2b6b-4362-9f8b-378c5859c786\") " pod="openstack/ceilometer-0" Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.262497 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0f39334f-2b6b-4362-9f8b-378c5859c786-run-httpd\") pod \"ceilometer-0\" (UID: \"0f39334f-2b6b-4362-9f8b-378c5859c786\") " pod="openstack/ceilometer-0" Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.262713 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hsxz9\" (UniqueName: \"kubernetes.io/projected/0f39334f-2b6b-4362-9f8b-378c5859c786-kube-api-access-hsxz9\") pod \"ceilometer-0\" (UID: \"0f39334f-2b6b-4362-9f8b-378c5859c786\") " pod="openstack/ceilometer-0" Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.262770 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f39334f-2b6b-4362-9f8b-378c5859c786-config-data\") pod \"ceilometer-0\" (UID: \"0f39334f-2b6b-4362-9f8b-378c5859c786\") " pod="openstack/ceilometer-0" Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.262928 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0f39334f-2b6b-4362-9f8b-378c5859c786-log-httpd\") pod \"ceilometer-0\" (UID: \"0f39334f-2b6b-4362-9f8b-378c5859c786\") " pod="openstack/ceilometer-0" Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.263154 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0f39334f-2b6b-4362-9f8b-378c5859c786-run-httpd\") pod \"ceilometer-0\" (UID: \"0f39334f-2b6b-4362-9f8b-378c5859c786\") " pod="openstack/ceilometer-0" Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.263454 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0f39334f-2b6b-4362-9f8b-378c5859c786-log-httpd\") pod \"ceilometer-0\" (UID: \"0f39334f-2b6b-4362-9f8b-378c5859c786\") " pod="openstack/ceilometer-0" Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.268003 4716 scope.go:117] "RemoveContainer" containerID="81d54b420e9a0506bc7d4efba865f53f5c07ac678546167d73a039558c995086" Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.268587 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0f39334f-2b6b-4362-9f8b-378c5859c786-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0f39334f-2b6b-4362-9f8b-378c5859c786\") " pod="openstack/ceilometer-0" Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.268786 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f39334f-2b6b-4362-9f8b-378c5859c786-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0f39334f-2b6b-4362-9f8b-378c5859c786\") " pod="openstack/ceilometer-0" Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.268958 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f39334f-2b6b-4362-9f8b-378c5859c786-config-data\") pod \"ceilometer-0\" (UID: \"0f39334f-2b6b-4362-9f8b-378c5859c786\") " pod="openstack/ceilometer-0" Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.273053 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f39334f-2b6b-4362-9f8b-378c5859c786-scripts\") pod \"ceilometer-0\" (UID: \"0f39334f-2b6b-4362-9f8b-378c5859c786\") " pod="openstack/ceilometer-0" Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.277668 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0f39334f-2b6b-4362-9f8b-378c5859c786-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0f39334f-2b6b-4362-9f8b-378c5859c786\") " pod="openstack/ceilometer-0" Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.286375 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hsxz9\" (UniqueName: \"kubernetes.io/projected/0f39334f-2b6b-4362-9f8b-378c5859c786-kube-api-access-hsxz9\") pod \"ceilometer-0\" (UID: \"0f39334f-2b6b-4362-9f8b-378c5859c786\") " pod="openstack/ceilometer-0" Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.368898 4716 scope.go:117] "RemoveContainer" containerID="086e0a12638bc85b5222676a602a725a665991b0f030d5aaa151a6e50aff57eb" Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.398257 4716 scope.go:117] "RemoveContainer" containerID="ef45036db19a07ce203b9018d7f545d5aaaeb31f8ee89c8dd4a93b441821fa1a" Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.425917 4716 scope.go:117] "RemoveContainer" containerID="cb8b887082e5b10ffc83cc548108c87e4c5071bbffd61194184f3b3a53de524d" Dec 07 16:20:37 crc kubenswrapper[4716]: E1207 16:20:37.426377 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb8b887082e5b10ffc83cc548108c87e4c5071bbffd61194184f3b3a53de524d\": container with ID starting with cb8b887082e5b10ffc83cc548108c87e4c5071bbffd61194184f3b3a53de524d not found: ID does not exist" containerID="cb8b887082e5b10ffc83cc548108c87e4c5071bbffd61194184f3b3a53de524d" Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.426462 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb8b887082e5b10ffc83cc548108c87e4c5071bbffd61194184f3b3a53de524d"} err="failed to get container status \"cb8b887082e5b10ffc83cc548108c87e4c5071bbffd61194184f3b3a53de524d\": rpc error: code = NotFound desc = could not find container \"cb8b887082e5b10ffc83cc548108c87e4c5071bbffd61194184f3b3a53de524d\": container with ID starting with cb8b887082e5b10ffc83cc548108c87e4c5071bbffd61194184f3b3a53de524d not found: ID does not exist" Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.426545 4716 scope.go:117] "RemoveContainer" containerID="81d54b420e9a0506bc7d4efba865f53f5c07ac678546167d73a039558c995086" Dec 07 16:20:37 crc kubenswrapper[4716]: E1207 16:20:37.426882 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"81d54b420e9a0506bc7d4efba865f53f5c07ac678546167d73a039558c995086\": container with ID starting with 81d54b420e9a0506bc7d4efba865f53f5c07ac678546167d73a039558c995086 not found: ID does not exist" containerID="81d54b420e9a0506bc7d4efba865f53f5c07ac678546167d73a039558c995086" Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.426970 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"81d54b420e9a0506bc7d4efba865f53f5c07ac678546167d73a039558c995086"} err="failed to get container status \"81d54b420e9a0506bc7d4efba865f53f5c07ac678546167d73a039558c995086\": rpc error: code = NotFound desc = could not find container \"81d54b420e9a0506bc7d4efba865f53f5c07ac678546167d73a039558c995086\": container with ID starting with 81d54b420e9a0506bc7d4efba865f53f5c07ac678546167d73a039558c995086 not found: ID does not exist" Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.427032 4716 scope.go:117] "RemoveContainer" containerID="086e0a12638bc85b5222676a602a725a665991b0f030d5aaa151a6e50aff57eb" Dec 07 16:20:37 crc kubenswrapper[4716]: E1207 16:20:37.427315 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"086e0a12638bc85b5222676a602a725a665991b0f030d5aaa151a6e50aff57eb\": container with ID starting with 086e0a12638bc85b5222676a602a725a665991b0f030d5aaa151a6e50aff57eb not found: ID does not exist" containerID="086e0a12638bc85b5222676a602a725a665991b0f030d5aaa151a6e50aff57eb" Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.427388 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"086e0a12638bc85b5222676a602a725a665991b0f030d5aaa151a6e50aff57eb"} err="failed to get container status \"086e0a12638bc85b5222676a602a725a665991b0f030d5aaa151a6e50aff57eb\": rpc error: code = NotFound desc = could not find container \"086e0a12638bc85b5222676a602a725a665991b0f030d5aaa151a6e50aff57eb\": container with ID starting with 086e0a12638bc85b5222676a602a725a665991b0f030d5aaa151a6e50aff57eb not found: ID does not exist" Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.427462 4716 scope.go:117] "RemoveContainer" containerID="ef45036db19a07ce203b9018d7f545d5aaaeb31f8ee89c8dd4a93b441821fa1a" Dec 07 16:20:37 crc kubenswrapper[4716]: E1207 16:20:37.429781 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef45036db19a07ce203b9018d7f545d5aaaeb31f8ee89c8dd4a93b441821fa1a\": container with ID starting with ef45036db19a07ce203b9018d7f545d5aaaeb31f8ee89c8dd4a93b441821fa1a not found: ID does not exist" containerID="ef45036db19a07ce203b9018d7f545d5aaaeb31f8ee89c8dd4a93b441821fa1a" Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.429831 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef45036db19a07ce203b9018d7f545d5aaaeb31f8ee89c8dd4a93b441821fa1a"} err="failed to get container status \"ef45036db19a07ce203b9018d7f545d5aaaeb31f8ee89c8dd4a93b441821fa1a\": rpc error: code = NotFound desc = could not find container \"ef45036db19a07ce203b9018d7f545d5aaaeb31f8ee89c8dd4a93b441821fa1a\": container with ID starting with ef45036db19a07ce203b9018d7f545d5aaaeb31f8ee89c8dd4a93b441821fa1a not found: ID does not exist" Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.482191 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.674518 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6873cce2-e68e-47ca-8ae8-7b612815a223" path="/var/lib/kubelet/pods/6873cce2-e68e-47ca-8ae8-7b612815a223/volumes" Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.675606 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c291e114-7940-46bc-91d3-a8fa256549ff" path="/var/lib/kubelet/pods/c291e114-7940-46bc-91d3-a8fa256549ff/volumes" Dec 07 16:20:37 crc kubenswrapper[4716]: I1207 16:20:37.956297 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 07 16:20:37 crc kubenswrapper[4716]: W1207 16:20:37.958719 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0f39334f_2b6b_4362_9f8b_378c5859c786.slice/crio-6cba6b0d57392ac128d3a6fe3a9b4b2b256bb54c13e7aa566673a277900df1e4 WatchSource:0}: Error finding container 6cba6b0d57392ac128d3a6fe3a9b4b2b256bb54c13e7aa566673a277900df1e4: Status 404 returned error can't find the container with id 6cba6b0d57392ac128d3a6fe3a9b4b2b256bb54c13e7aa566673a277900df1e4 Dec 07 16:20:38 crc kubenswrapper[4716]: I1207 16:20:38.040570 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0f39334f-2b6b-4362-9f8b-378c5859c786","Type":"ContainerStarted","Data":"6cba6b0d57392ac128d3a6fe3a9b4b2b256bb54c13e7aa566673a277900df1e4"} Dec 07 16:20:39 crc kubenswrapper[4716]: I1207 16:20:39.052891 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0f39334f-2b6b-4362-9f8b-378c5859c786","Type":"ContainerStarted","Data":"97a4ff00b29a9e6830f5eccd1b76474a7cdb96bbd591c32dff188cd18ec3a223"} Dec 07 16:20:39 crc kubenswrapper[4716]: I1207 16:20:39.557058 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-6kcdj"] Dec 07 16:20:39 crc kubenswrapper[4716]: I1207 16:20:39.558242 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-6kcdj" Dec 07 16:20:39 crc kubenswrapper[4716]: I1207 16:20:39.567178 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Dec 07 16:20:39 crc kubenswrapper[4716]: I1207 16:20:39.567348 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-j5z2b" Dec 07 16:20:39 crc kubenswrapper[4716]: I1207 16:20:39.569326 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 07 16:20:39 crc kubenswrapper[4716]: I1207 16:20:39.587835 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-6kcdj"] Dec 07 16:20:39 crc kubenswrapper[4716]: I1207 16:20:39.626422 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a587b1d-a86e-4ee2-907e-6b014c61b09b-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-6kcdj\" (UID: \"7a587b1d-a86e-4ee2-907e-6b014c61b09b\") " pod="openstack/nova-cell0-conductor-db-sync-6kcdj" Dec 07 16:20:39 crc kubenswrapper[4716]: I1207 16:20:39.626458 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a587b1d-a86e-4ee2-907e-6b014c61b09b-scripts\") pod \"nova-cell0-conductor-db-sync-6kcdj\" (UID: \"7a587b1d-a86e-4ee2-907e-6b014c61b09b\") " pod="openstack/nova-cell0-conductor-db-sync-6kcdj" Dec 07 16:20:39 crc kubenswrapper[4716]: I1207 16:20:39.626491 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d6j4h\" (UniqueName: \"kubernetes.io/projected/7a587b1d-a86e-4ee2-907e-6b014c61b09b-kube-api-access-d6j4h\") pod \"nova-cell0-conductor-db-sync-6kcdj\" (UID: \"7a587b1d-a86e-4ee2-907e-6b014c61b09b\") " pod="openstack/nova-cell0-conductor-db-sync-6kcdj" Dec 07 16:20:39 crc kubenswrapper[4716]: I1207 16:20:39.626510 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a587b1d-a86e-4ee2-907e-6b014c61b09b-config-data\") pod \"nova-cell0-conductor-db-sync-6kcdj\" (UID: \"7a587b1d-a86e-4ee2-907e-6b014c61b09b\") " pod="openstack/nova-cell0-conductor-db-sync-6kcdj" Dec 07 16:20:39 crc kubenswrapper[4716]: I1207 16:20:39.730131 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a587b1d-a86e-4ee2-907e-6b014c61b09b-scripts\") pod \"nova-cell0-conductor-db-sync-6kcdj\" (UID: \"7a587b1d-a86e-4ee2-907e-6b014c61b09b\") " pod="openstack/nova-cell0-conductor-db-sync-6kcdj" Dec 07 16:20:39 crc kubenswrapper[4716]: I1207 16:20:39.730174 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a587b1d-a86e-4ee2-907e-6b014c61b09b-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-6kcdj\" (UID: \"7a587b1d-a86e-4ee2-907e-6b014c61b09b\") " pod="openstack/nova-cell0-conductor-db-sync-6kcdj" Dec 07 16:20:39 crc kubenswrapper[4716]: I1207 16:20:39.730266 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d6j4h\" (UniqueName: \"kubernetes.io/projected/7a587b1d-a86e-4ee2-907e-6b014c61b09b-kube-api-access-d6j4h\") pod \"nova-cell0-conductor-db-sync-6kcdj\" (UID: \"7a587b1d-a86e-4ee2-907e-6b014c61b09b\") " pod="openstack/nova-cell0-conductor-db-sync-6kcdj" Dec 07 16:20:39 crc kubenswrapper[4716]: I1207 16:20:39.730300 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a587b1d-a86e-4ee2-907e-6b014c61b09b-config-data\") pod \"nova-cell0-conductor-db-sync-6kcdj\" (UID: \"7a587b1d-a86e-4ee2-907e-6b014c61b09b\") " pod="openstack/nova-cell0-conductor-db-sync-6kcdj" Dec 07 16:20:39 crc kubenswrapper[4716]: I1207 16:20:39.735608 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a587b1d-a86e-4ee2-907e-6b014c61b09b-scripts\") pod \"nova-cell0-conductor-db-sync-6kcdj\" (UID: \"7a587b1d-a86e-4ee2-907e-6b014c61b09b\") " pod="openstack/nova-cell0-conductor-db-sync-6kcdj" Dec 07 16:20:39 crc kubenswrapper[4716]: I1207 16:20:39.735838 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a587b1d-a86e-4ee2-907e-6b014c61b09b-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-6kcdj\" (UID: \"7a587b1d-a86e-4ee2-907e-6b014c61b09b\") " pod="openstack/nova-cell0-conductor-db-sync-6kcdj" Dec 07 16:20:39 crc kubenswrapper[4716]: I1207 16:20:39.746407 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a587b1d-a86e-4ee2-907e-6b014c61b09b-config-data\") pod \"nova-cell0-conductor-db-sync-6kcdj\" (UID: \"7a587b1d-a86e-4ee2-907e-6b014c61b09b\") " pod="openstack/nova-cell0-conductor-db-sync-6kcdj" Dec 07 16:20:39 crc kubenswrapper[4716]: I1207 16:20:39.746850 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d6j4h\" (UniqueName: \"kubernetes.io/projected/7a587b1d-a86e-4ee2-907e-6b014c61b09b-kube-api-access-d6j4h\") pod \"nova-cell0-conductor-db-sync-6kcdj\" (UID: \"7a587b1d-a86e-4ee2-907e-6b014c61b09b\") " pod="openstack/nova-cell0-conductor-db-sync-6kcdj" Dec 07 16:20:39 crc kubenswrapper[4716]: I1207 16:20:39.832705 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 07 16:20:39 crc kubenswrapper[4716]: I1207 16:20:39.833067 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 07 16:20:39 crc kubenswrapper[4716]: I1207 16:20:39.877912 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-6kcdj" Dec 07 16:20:39 crc kubenswrapper[4716]: I1207 16:20:39.878635 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 07 16:20:39 crc kubenswrapper[4716]: I1207 16:20:39.878719 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 07 16:20:40 crc kubenswrapper[4716]: I1207 16:20:40.070270 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0f39334f-2b6b-4362-9f8b-378c5859c786","Type":"ContainerStarted","Data":"d36af3dba8a21ff6778f2ff44630ca22ed77f777413511d9a2f6bf2cc1cdfa21"} Dec 07 16:20:40 crc kubenswrapper[4716]: I1207 16:20:40.070580 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 07 16:20:40 crc kubenswrapper[4716]: I1207 16:20:40.070612 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 07 16:20:40 crc kubenswrapper[4716]: I1207 16:20:40.128239 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 07 16:20:40 crc kubenswrapper[4716]: I1207 16:20:40.128299 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 07 16:20:40 crc kubenswrapper[4716]: I1207 16:20:40.178134 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 07 16:20:40 crc kubenswrapper[4716]: I1207 16:20:40.185897 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 07 16:20:40 crc kubenswrapper[4716]: I1207 16:20:40.359450 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-6kcdj"] Dec 07 16:20:41 crc kubenswrapper[4716]: I1207 16:20:41.079814 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0f39334f-2b6b-4362-9f8b-378c5859c786","Type":"ContainerStarted","Data":"e9dd18c93c5f88ae45851ac9038d1eb97781daeb7d6d56a286df308aba33f41e"} Dec 07 16:20:41 crc kubenswrapper[4716]: I1207 16:20:41.081457 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-6kcdj" event={"ID":"7a587b1d-a86e-4ee2-907e-6b014c61b09b","Type":"ContainerStarted","Data":"a86371ed238c219743c0204426f459d4c2eaaada8d805204592093b84261ccad"} Dec 07 16:20:41 crc kubenswrapper[4716]: I1207 16:20:41.082280 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 07 16:20:41 crc kubenswrapper[4716]: I1207 16:20:41.082304 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 07 16:20:41 crc kubenswrapper[4716]: I1207 16:20:41.534348 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 07 16:20:41 crc kubenswrapper[4716]: I1207 16:20:41.968832 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 07 16:20:42 crc kubenswrapper[4716]: I1207 16:20:42.094756 4716 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 07 16:20:42 crc kubenswrapper[4716]: I1207 16:20:42.244990 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 07 16:20:42 crc kubenswrapper[4716]: I1207 16:20:42.284232 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 07 16:20:43 crc kubenswrapper[4716]: I1207 16:20:43.103151 4716 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 07 16:20:43 crc kubenswrapper[4716]: I1207 16:20:43.103382 4716 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 07 16:20:43 crc kubenswrapper[4716]: I1207 16:20:43.307473 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 07 16:20:43 crc kubenswrapper[4716]: I1207 16:20:43.312743 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 07 16:20:44 crc kubenswrapper[4716]: I1207 16:20:44.116129 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0f39334f-2b6b-4362-9f8b-378c5859c786","Type":"ContainerStarted","Data":"99ea5996aeac68059135d341c182767af46241b8086eb8def3d008c55f82b8dc"} Dec 07 16:20:44 crc kubenswrapper[4716]: I1207 16:20:44.116465 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0f39334f-2b6b-4362-9f8b-378c5859c786" containerName="ceilometer-central-agent" containerID="cri-o://97a4ff00b29a9e6830f5eccd1b76474a7cdb96bbd591c32dff188cd18ec3a223" gracePeriod=30 Dec 07 16:20:44 crc kubenswrapper[4716]: I1207 16:20:44.116832 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0f39334f-2b6b-4362-9f8b-378c5859c786" containerName="proxy-httpd" containerID="cri-o://99ea5996aeac68059135d341c182767af46241b8086eb8def3d008c55f82b8dc" gracePeriod=30 Dec 07 16:20:44 crc kubenswrapper[4716]: I1207 16:20:44.116940 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0f39334f-2b6b-4362-9f8b-378c5859c786" containerName="ceilometer-notification-agent" containerID="cri-o://d36af3dba8a21ff6778f2ff44630ca22ed77f777413511d9a2f6bf2cc1cdfa21" gracePeriod=30 Dec 07 16:20:44 crc kubenswrapper[4716]: I1207 16:20:44.116992 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0f39334f-2b6b-4362-9f8b-378c5859c786" containerName="sg-core" containerID="cri-o://e9dd18c93c5f88ae45851ac9038d1eb97781daeb7d6d56a286df308aba33f41e" gracePeriod=30 Dec 07 16:20:45 crc kubenswrapper[4716]: I1207 16:20:45.127882 4716 generic.go:334] "Generic (PLEG): container finished" podID="0f39334f-2b6b-4362-9f8b-378c5859c786" containerID="99ea5996aeac68059135d341c182767af46241b8086eb8def3d008c55f82b8dc" exitCode=0 Dec 07 16:20:45 crc kubenswrapper[4716]: I1207 16:20:45.128185 4716 generic.go:334] "Generic (PLEG): container finished" podID="0f39334f-2b6b-4362-9f8b-378c5859c786" containerID="e9dd18c93c5f88ae45851ac9038d1eb97781daeb7d6d56a286df308aba33f41e" exitCode=2 Dec 07 16:20:45 crc kubenswrapper[4716]: I1207 16:20:45.128193 4716 generic.go:334] "Generic (PLEG): container finished" podID="0f39334f-2b6b-4362-9f8b-378c5859c786" containerID="d36af3dba8a21ff6778f2ff44630ca22ed77f777413511d9a2f6bf2cc1cdfa21" exitCode=0 Dec 07 16:20:45 crc kubenswrapper[4716]: I1207 16:20:45.127951 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0f39334f-2b6b-4362-9f8b-378c5859c786","Type":"ContainerDied","Data":"99ea5996aeac68059135d341c182767af46241b8086eb8def3d008c55f82b8dc"} Dec 07 16:20:45 crc kubenswrapper[4716]: I1207 16:20:45.128285 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0f39334f-2b6b-4362-9f8b-378c5859c786","Type":"ContainerDied","Data":"e9dd18c93c5f88ae45851ac9038d1eb97781daeb7d6d56a286df308aba33f41e"} Dec 07 16:20:45 crc kubenswrapper[4716]: I1207 16:20:45.128298 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0f39334f-2b6b-4362-9f8b-378c5859c786","Type":"ContainerDied","Data":"d36af3dba8a21ff6778f2ff44630ca22ed77f777413511d9a2f6bf2cc1cdfa21"} Dec 07 16:20:45 crc kubenswrapper[4716]: I1207 16:20:45.587427 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 07 16:20:45 crc kubenswrapper[4716]: I1207 16:20:45.606639 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.248380835 podStartE2EDuration="8.606611832s" podCreationTimestamp="2025-12-07 16:20:37 +0000 UTC" firstStartedPulling="2025-12-07 16:20:37.96198933 +0000 UTC m=+1100.652274242" lastFinishedPulling="2025-12-07 16:20:43.320220327 +0000 UTC m=+1106.010505239" observedRunningTime="2025-12-07 16:20:44.147774154 +0000 UTC m=+1106.838059076" watchObservedRunningTime="2025-12-07 16:20:45.606611832 +0000 UTC m=+1108.296896744" Dec 07 16:20:51 crc kubenswrapper[4716]: I1207 16:20:51.192582 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-6kcdj" event={"ID":"7a587b1d-a86e-4ee2-907e-6b014c61b09b","Type":"ContainerStarted","Data":"3f6d2d6aa158ff9d5af36e3ebe2ea76619cd872f551bf4eba62dd2343c17d952"} Dec 07 16:20:51 crc kubenswrapper[4716]: I1207 16:20:51.218174 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-6kcdj" podStartSLOduration=2.428500111 podStartE2EDuration="12.218152545s" podCreationTimestamp="2025-12-07 16:20:39 +0000 UTC" firstStartedPulling="2025-12-07 16:20:40.372054512 +0000 UTC m=+1103.062339424" lastFinishedPulling="2025-12-07 16:20:50.161706946 +0000 UTC m=+1112.851991858" observedRunningTime="2025-12-07 16:20:51.212586295 +0000 UTC m=+1113.902871237" watchObservedRunningTime="2025-12-07 16:20:51.218152545 +0000 UTC m=+1113.908437467" Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.141405 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.218455 4716 generic.go:334] "Generic (PLEG): container finished" podID="0f39334f-2b6b-4362-9f8b-378c5859c786" containerID="97a4ff00b29a9e6830f5eccd1b76474a7cdb96bbd591c32dff188cd18ec3a223" exitCode=0 Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.218501 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0f39334f-2b6b-4362-9f8b-378c5859c786","Type":"ContainerDied","Data":"97a4ff00b29a9e6830f5eccd1b76474a7cdb96bbd591c32dff188cd18ec3a223"} Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.218530 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0f39334f-2b6b-4362-9f8b-378c5859c786","Type":"ContainerDied","Data":"6cba6b0d57392ac128d3a6fe3a9b4b2b256bb54c13e7aa566673a277900df1e4"} Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.218548 4716 scope.go:117] "RemoveContainer" containerID="99ea5996aeac68059135d341c182767af46241b8086eb8def3d008c55f82b8dc" Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.218682 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.238057 4716 scope.go:117] "RemoveContainer" containerID="e9dd18c93c5f88ae45851ac9038d1eb97781daeb7d6d56a286df308aba33f41e" Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.255864 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f39334f-2b6b-4362-9f8b-378c5859c786-scripts\") pod \"0f39334f-2b6b-4362-9f8b-378c5859c786\" (UID: \"0f39334f-2b6b-4362-9f8b-378c5859c786\") " Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.256015 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hsxz9\" (UniqueName: \"kubernetes.io/projected/0f39334f-2b6b-4362-9f8b-378c5859c786-kube-api-access-hsxz9\") pod \"0f39334f-2b6b-4362-9f8b-378c5859c786\" (UID: \"0f39334f-2b6b-4362-9f8b-378c5859c786\") " Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.256051 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0f39334f-2b6b-4362-9f8b-378c5859c786-run-httpd\") pod \"0f39334f-2b6b-4362-9f8b-378c5859c786\" (UID: \"0f39334f-2b6b-4362-9f8b-378c5859c786\") " Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.256129 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0f39334f-2b6b-4362-9f8b-378c5859c786-log-httpd\") pod \"0f39334f-2b6b-4362-9f8b-378c5859c786\" (UID: \"0f39334f-2b6b-4362-9f8b-378c5859c786\") " Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.256158 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0f39334f-2b6b-4362-9f8b-378c5859c786-ceilometer-tls-certs\") pod \"0f39334f-2b6b-4362-9f8b-378c5859c786\" (UID: \"0f39334f-2b6b-4362-9f8b-378c5859c786\") " Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.256218 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f39334f-2b6b-4362-9f8b-378c5859c786-combined-ca-bundle\") pod \"0f39334f-2b6b-4362-9f8b-378c5859c786\" (UID: \"0f39334f-2b6b-4362-9f8b-378c5859c786\") " Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.256247 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f39334f-2b6b-4362-9f8b-378c5859c786-config-data\") pod \"0f39334f-2b6b-4362-9f8b-378c5859c786\" (UID: \"0f39334f-2b6b-4362-9f8b-378c5859c786\") " Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.256261 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0f39334f-2b6b-4362-9f8b-378c5859c786-sg-core-conf-yaml\") pod \"0f39334f-2b6b-4362-9f8b-378c5859c786\" (UID: \"0f39334f-2b6b-4362-9f8b-378c5859c786\") " Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.257180 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f39334f-2b6b-4362-9f8b-378c5859c786-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "0f39334f-2b6b-4362-9f8b-378c5859c786" (UID: "0f39334f-2b6b-4362-9f8b-378c5859c786"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.257412 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f39334f-2b6b-4362-9f8b-378c5859c786-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "0f39334f-2b6b-4362-9f8b-378c5859c786" (UID: "0f39334f-2b6b-4362-9f8b-378c5859c786"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.259445 4716 scope.go:117] "RemoveContainer" containerID="d36af3dba8a21ff6778f2ff44630ca22ed77f777413511d9a2f6bf2cc1cdfa21" Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.261927 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f39334f-2b6b-4362-9f8b-378c5859c786-scripts" (OuterVolumeSpecName: "scripts") pod "0f39334f-2b6b-4362-9f8b-378c5859c786" (UID: "0f39334f-2b6b-4362-9f8b-378c5859c786"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.263305 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f39334f-2b6b-4362-9f8b-378c5859c786-kube-api-access-hsxz9" (OuterVolumeSpecName: "kube-api-access-hsxz9") pod "0f39334f-2b6b-4362-9f8b-378c5859c786" (UID: "0f39334f-2b6b-4362-9f8b-378c5859c786"). InnerVolumeSpecName "kube-api-access-hsxz9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.289279 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f39334f-2b6b-4362-9f8b-378c5859c786-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "0f39334f-2b6b-4362-9f8b-378c5859c786" (UID: "0f39334f-2b6b-4362-9f8b-378c5859c786"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.315171 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f39334f-2b6b-4362-9f8b-378c5859c786-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "0f39334f-2b6b-4362-9f8b-378c5859c786" (UID: "0f39334f-2b6b-4362-9f8b-378c5859c786"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.336277 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f39334f-2b6b-4362-9f8b-378c5859c786-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0f39334f-2b6b-4362-9f8b-378c5859c786" (UID: "0f39334f-2b6b-4362-9f8b-378c5859c786"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.357931 4716 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f39334f-2b6b-4362-9f8b-378c5859c786-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.357959 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hsxz9\" (UniqueName: \"kubernetes.io/projected/0f39334f-2b6b-4362-9f8b-378c5859c786-kube-api-access-hsxz9\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.357970 4716 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0f39334f-2b6b-4362-9f8b-378c5859c786-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.357981 4716 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0f39334f-2b6b-4362-9f8b-378c5859c786-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.357990 4716 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0f39334f-2b6b-4362-9f8b-378c5859c786-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.357998 4716 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f39334f-2b6b-4362-9f8b-378c5859c786-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.358007 4716 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0f39334f-2b6b-4362-9f8b-378c5859c786-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.358704 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f39334f-2b6b-4362-9f8b-378c5859c786-config-data" (OuterVolumeSpecName: "config-data") pod "0f39334f-2b6b-4362-9f8b-378c5859c786" (UID: "0f39334f-2b6b-4362-9f8b-378c5859c786"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.360543 4716 scope.go:117] "RemoveContainer" containerID="97a4ff00b29a9e6830f5eccd1b76474a7cdb96bbd591c32dff188cd18ec3a223" Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.376596 4716 scope.go:117] "RemoveContainer" containerID="99ea5996aeac68059135d341c182767af46241b8086eb8def3d008c55f82b8dc" Dec 07 16:20:53 crc kubenswrapper[4716]: E1207 16:20:53.377180 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"99ea5996aeac68059135d341c182767af46241b8086eb8def3d008c55f82b8dc\": container with ID starting with 99ea5996aeac68059135d341c182767af46241b8086eb8def3d008c55f82b8dc not found: ID does not exist" containerID="99ea5996aeac68059135d341c182767af46241b8086eb8def3d008c55f82b8dc" Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.377229 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99ea5996aeac68059135d341c182767af46241b8086eb8def3d008c55f82b8dc"} err="failed to get container status \"99ea5996aeac68059135d341c182767af46241b8086eb8def3d008c55f82b8dc\": rpc error: code = NotFound desc = could not find container \"99ea5996aeac68059135d341c182767af46241b8086eb8def3d008c55f82b8dc\": container with ID starting with 99ea5996aeac68059135d341c182767af46241b8086eb8def3d008c55f82b8dc not found: ID does not exist" Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.377286 4716 scope.go:117] "RemoveContainer" containerID="e9dd18c93c5f88ae45851ac9038d1eb97781daeb7d6d56a286df308aba33f41e" Dec 07 16:20:53 crc kubenswrapper[4716]: E1207 16:20:53.377609 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e9dd18c93c5f88ae45851ac9038d1eb97781daeb7d6d56a286df308aba33f41e\": container with ID starting with e9dd18c93c5f88ae45851ac9038d1eb97781daeb7d6d56a286df308aba33f41e not found: ID does not exist" containerID="e9dd18c93c5f88ae45851ac9038d1eb97781daeb7d6d56a286df308aba33f41e" Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.377647 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9dd18c93c5f88ae45851ac9038d1eb97781daeb7d6d56a286df308aba33f41e"} err="failed to get container status \"e9dd18c93c5f88ae45851ac9038d1eb97781daeb7d6d56a286df308aba33f41e\": rpc error: code = NotFound desc = could not find container \"e9dd18c93c5f88ae45851ac9038d1eb97781daeb7d6d56a286df308aba33f41e\": container with ID starting with e9dd18c93c5f88ae45851ac9038d1eb97781daeb7d6d56a286df308aba33f41e not found: ID does not exist" Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.377675 4716 scope.go:117] "RemoveContainer" containerID="d36af3dba8a21ff6778f2ff44630ca22ed77f777413511d9a2f6bf2cc1cdfa21" Dec 07 16:20:53 crc kubenswrapper[4716]: E1207 16:20:53.377942 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d36af3dba8a21ff6778f2ff44630ca22ed77f777413511d9a2f6bf2cc1cdfa21\": container with ID starting with d36af3dba8a21ff6778f2ff44630ca22ed77f777413511d9a2f6bf2cc1cdfa21 not found: ID does not exist" containerID="d36af3dba8a21ff6778f2ff44630ca22ed77f777413511d9a2f6bf2cc1cdfa21" Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.377970 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d36af3dba8a21ff6778f2ff44630ca22ed77f777413511d9a2f6bf2cc1cdfa21"} err="failed to get container status \"d36af3dba8a21ff6778f2ff44630ca22ed77f777413511d9a2f6bf2cc1cdfa21\": rpc error: code = NotFound desc = could not find container \"d36af3dba8a21ff6778f2ff44630ca22ed77f777413511d9a2f6bf2cc1cdfa21\": container with ID starting with d36af3dba8a21ff6778f2ff44630ca22ed77f777413511d9a2f6bf2cc1cdfa21 not found: ID does not exist" Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.377986 4716 scope.go:117] "RemoveContainer" containerID="97a4ff00b29a9e6830f5eccd1b76474a7cdb96bbd591c32dff188cd18ec3a223" Dec 07 16:20:53 crc kubenswrapper[4716]: E1207 16:20:53.378340 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"97a4ff00b29a9e6830f5eccd1b76474a7cdb96bbd591c32dff188cd18ec3a223\": container with ID starting with 97a4ff00b29a9e6830f5eccd1b76474a7cdb96bbd591c32dff188cd18ec3a223 not found: ID does not exist" containerID="97a4ff00b29a9e6830f5eccd1b76474a7cdb96bbd591c32dff188cd18ec3a223" Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.378373 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"97a4ff00b29a9e6830f5eccd1b76474a7cdb96bbd591c32dff188cd18ec3a223"} err="failed to get container status \"97a4ff00b29a9e6830f5eccd1b76474a7cdb96bbd591c32dff188cd18ec3a223\": rpc error: code = NotFound desc = could not find container \"97a4ff00b29a9e6830f5eccd1b76474a7cdb96bbd591c32dff188cd18ec3a223\": container with ID starting with 97a4ff00b29a9e6830f5eccd1b76474a7cdb96bbd591c32dff188cd18ec3a223 not found: ID does not exist" Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.460119 4716 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f39334f-2b6b-4362-9f8b-378c5859c786-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.552216 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.560548 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.586010 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 07 16:20:53 crc kubenswrapper[4716]: E1207 16:20:53.586406 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f39334f-2b6b-4362-9f8b-378c5859c786" containerName="sg-core" Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.586426 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f39334f-2b6b-4362-9f8b-378c5859c786" containerName="sg-core" Dec 07 16:20:53 crc kubenswrapper[4716]: E1207 16:20:53.586449 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f39334f-2b6b-4362-9f8b-378c5859c786" containerName="ceilometer-notification-agent" Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.586456 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f39334f-2b6b-4362-9f8b-378c5859c786" containerName="ceilometer-notification-agent" Dec 07 16:20:53 crc kubenswrapper[4716]: E1207 16:20:53.586468 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f39334f-2b6b-4362-9f8b-378c5859c786" containerName="ceilometer-central-agent" Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.586474 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f39334f-2b6b-4362-9f8b-378c5859c786" containerName="ceilometer-central-agent" Dec 07 16:20:53 crc kubenswrapper[4716]: E1207 16:20:53.586495 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f39334f-2b6b-4362-9f8b-378c5859c786" containerName="proxy-httpd" Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.586501 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f39334f-2b6b-4362-9f8b-378c5859c786" containerName="proxy-httpd" Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.586655 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f39334f-2b6b-4362-9f8b-378c5859c786" containerName="sg-core" Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.586665 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f39334f-2b6b-4362-9f8b-378c5859c786" containerName="proxy-httpd" Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.586678 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f39334f-2b6b-4362-9f8b-378c5859c786" containerName="ceilometer-notification-agent" Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.586697 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f39334f-2b6b-4362-9f8b-378c5859c786" containerName="ceilometer-central-agent" Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.588588 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.590425 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.590754 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.597066 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.599001 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.667450 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f39334f-2b6b-4362-9f8b-378c5859c786" path="/var/lib/kubelet/pods/0f39334f-2b6b-4362-9f8b-378c5859c786/volumes" Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.765198 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/859fd7aa-6477-4335-a671-a760a1dd7cf8-scripts\") pod \"ceilometer-0\" (UID: \"859fd7aa-6477-4335-a671-a760a1dd7cf8\") " pod="openstack/ceilometer-0" Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.765261 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/859fd7aa-6477-4335-a671-a760a1dd7cf8-config-data\") pod \"ceilometer-0\" (UID: \"859fd7aa-6477-4335-a671-a760a1dd7cf8\") " pod="openstack/ceilometer-0" Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.765328 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xqnkh\" (UniqueName: \"kubernetes.io/projected/859fd7aa-6477-4335-a671-a760a1dd7cf8-kube-api-access-xqnkh\") pod \"ceilometer-0\" (UID: \"859fd7aa-6477-4335-a671-a760a1dd7cf8\") " pod="openstack/ceilometer-0" Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.765378 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/859fd7aa-6477-4335-a671-a760a1dd7cf8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"859fd7aa-6477-4335-a671-a760a1dd7cf8\") " pod="openstack/ceilometer-0" Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.765397 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/859fd7aa-6477-4335-a671-a760a1dd7cf8-log-httpd\") pod \"ceilometer-0\" (UID: \"859fd7aa-6477-4335-a671-a760a1dd7cf8\") " pod="openstack/ceilometer-0" Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.765415 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/859fd7aa-6477-4335-a671-a760a1dd7cf8-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"859fd7aa-6477-4335-a671-a760a1dd7cf8\") " pod="openstack/ceilometer-0" Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.765462 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/859fd7aa-6477-4335-a671-a760a1dd7cf8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"859fd7aa-6477-4335-a671-a760a1dd7cf8\") " pod="openstack/ceilometer-0" Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.765502 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/859fd7aa-6477-4335-a671-a760a1dd7cf8-run-httpd\") pod \"ceilometer-0\" (UID: \"859fd7aa-6477-4335-a671-a760a1dd7cf8\") " pod="openstack/ceilometer-0" Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.866827 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xqnkh\" (UniqueName: \"kubernetes.io/projected/859fd7aa-6477-4335-a671-a760a1dd7cf8-kube-api-access-xqnkh\") pod \"ceilometer-0\" (UID: \"859fd7aa-6477-4335-a671-a760a1dd7cf8\") " pod="openstack/ceilometer-0" Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.866916 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/859fd7aa-6477-4335-a671-a760a1dd7cf8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"859fd7aa-6477-4335-a671-a760a1dd7cf8\") " pod="openstack/ceilometer-0" Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.867137 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/859fd7aa-6477-4335-a671-a760a1dd7cf8-log-httpd\") pod \"ceilometer-0\" (UID: \"859fd7aa-6477-4335-a671-a760a1dd7cf8\") " pod="openstack/ceilometer-0" Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.867168 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/859fd7aa-6477-4335-a671-a760a1dd7cf8-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"859fd7aa-6477-4335-a671-a760a1dd7cf8\") " pod="openstack/ceilometer-0" Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.867213 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/859fd7aa-6477-4335-a671-a760a1dd7cf8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"859fd7aa-6477-4335-a671-a760a1dd7cf8\") " pod="openstack/ceilometer-0" Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.867272 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/859fd7aa-6477-4335-a671-a760a1dd7cf8-run-httpd\") pod \"ceilometer-0\" (UID: \"859fd7aa-6477-4335-a671-a760a1dd7cf8\") " pod="openstack/ceilometer-0" Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.867317 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/859fd7aa-6477-4335-a671-a760a1dd7cf8-scripts\") pod \"ceilometer-0\" (UID: \"859fd7aa-6477-4335-a671-a760a1dd7cf8\") " pod="openstack/ceilometer-0" Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.867348 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/859fd7aa-6477-4335-a671-a760a1dd7cf8-config-data\") pod \"ceilometer-0\" (UID: \"859fd7aa-6477-4335-a671-a760a1dd7cf8\") " pod="openstack/ceilometer-0" Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.870116 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/859fd7aa-6477-4335-a671-a760a1dd7cf8-run-httpd\") pod \"ceilometer-0\" (UID: \"859fd7aa-6477-4335-a671-a760a1dd7cf8\") " pod="openstack/ceilometer-0" Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.870410 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/859fd7aa-6477-4335-a671-a760a1dd7cf8-log-httpd\") pod \"ceilometer-0\" (UID: \"859fd7aa-6477-4335-a671-a760a1dd7cf8\") " pod="openstack/ceilometer-0" Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.872724 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/859fd7aa-6477-4335-a671-a760a1dd7cf8-scripts\") pod \"ceilometer-0\" (UID: \"859fd7aa-6477-4335-a671-a760a1dd7cf8\") " pod="openstack/ceilometer-0" Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.873013 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/859fd7aa-6477-4335-a671-a760a1dd7cf8-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"859fd7aa-6477-4335-a671-a760a1dd7cf8\") " pod="openstack/ceilometer-0" Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.874264 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/859fd7aa-6477-4335-a671-a760a1dd7cf8-config-data\") pod \"ceilometer-0\" (UID: \"859fd7aa-6477-4335-a671-a760a1dd7cf8\") " pod="openstack/ceilometer-0" Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.874338 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/859fd7aa-6477-4335-a671-a760a1dd7cf8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"859fd7aa-6477-4335-a671-a760a1dd7cf8\") " pod="openstack/ceilometer-0" Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.875402 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/859fd7aa-6477-4335-a671-a760a1dd7cf8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"859fd7aa-6477-4335-a671-a760a1dd7cf8\") " pod="openstack/ceilometer-0" Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.884969 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xqnkh\" (UniqueName: \"kubernetes.io/projected/859fd7aa-6477-4335-a671-a760a1dd7cf8-kube-api-access-xqnkh\") pod \"ceilometer-0\" (UID: \"859fd7aa-6477-4335-a671-a760a1dd7cf8\") " pod="openstack/ceilometer-0" Dec 07 16:20:53 crc kubenswrapper[4716]: I1207 16:20:53.914243 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 07 16:20:54 crc kubenswrapper[4716]: I1207 16:20:54.415229 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 07 16:20:54 crc kubenswrapper[4716]: W1207 16:20:54.422555 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod859fd7aa_6477_4335_a671_a760a1dd7cf8.slice/crio-4b8b03b5a4bdc7bc29fc1ed279f7d918e015de09a8da939857e3cf9383be859e WatchSource:0}: Error finding container 4b8b03b5a4bdc7bc29fc1ed279f7d918e015de09a8da939857e3cf9383be859e: Status 404 returned error can't find the container with id 4b8b03b5a4bdc7bc29fc1ed279f7d918e015de09a8da939857e3cf9383be859e Dec 07 16:20:55 crc kubenswrapper[4716]: I1207 16:20:55.257240 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"859fd7aa-6477-4335-a671-a760a1dd7cf8","Type":"ContainerStarted","Data":"11ff631224cb8de030199944568fa2d79ef30c81639510b9f243eec113ca4ca3"} Dec 07 16:20:55 crc kubenswrapper[4716]: I1207 16:20:55.257537 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"859fd7aa-6477-4335-a671-a760a1dd7cf8","Type":"ContainerStarted","Data":"4b8b03b5a4bdc7bc29fc1ed279f7d918e015de09a8da939857e3cf9383be859e"} Dec 07 16:20:56 crc kubenswrapper[4716]: I1207 16:20:56.270327 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"859fd7aa-6477-4335-a671-a760a1dd7cf8","Type":"ContainerStarted","Data":"f7b515d9c5e9dcc7f85ce5e46b76c9af49507ad42f2d3fb0fa16bcf2f017f206"} Dec 07 16:20:57 crc kubenswrapper[4716]: I1207 16:20:57.288037 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"859fd7aa-6477-4335-a671-a760a1dd7cf8","Type":"ContainerStarted","Data":"01ea3429fecbd3394b8b9df1392f6e7969d731311de9f2029960bc25ad215ba0"} Dec 07 16:20:58 crc kubenswrapper[4716]: I1207 16:20:58.297548 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"859fd7aa-6477-4335-a671-a760a1dd7cf8","Type":"ContainerStarted","Data":"f7ee11dae27dd6d0d58a7912b0ae41e6c384cc2424cdf86625543cd75053e7c6"} Dec 07 16:20:58 crc kubenswrapper[4716]: I1207 16:20:58.298153 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 07 16:20:58 crc kubenswrapper[4716]: I1207 16:20:58.317244 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.047363654 podStartE2EDuration="5.317221639s" podCreationTimestamp="2025-12-07 16:20:53 +0000 UTC" firstStartedPulling="2025-12-07 16:20:54.425458054 +0000 UTC m=+1117.115742966" lastFinishedPulling="2025-12-07 16:20:57.695316029 +0000 UTC m=+1120.385600951" observedRunningTime="2025-12-07 16:20:58.314539817 +0000 UTC m=+1121.004824729" watchObservedRunningTime="2025-12-07 16:20:58.317221639 +0000 UTC m=+1121.007506551" Dec 07 16:21:01 crc kubenswrapper[4716]: I1207 16:21:01.339978 4716 generic.go:334] "Generic (PLEG): container finished" podID="7a587b1d-a86e-4ee2-907e-6b014c61b09b" containerID="3f6d2d6aa158ff9d5af36e3ebe2ea76619cd872f551bf4eba62dd2343c17d952" exitCode=0 Dec 07 16:21:01 crc kubenswrapper[4716]: I1207 16:21:01.340308 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-6kcdj" event={"ID":"7a587b1d-a86e-4ee2-907e-6b014c61b09b","Type":"ContainerDied","Data":"3f6d2d6aa158ff9d5af36e3ebe2ea76619cd872f551bf4eba62dd2343c17d952"} Dec 07 16:21:02 crc kubenswrapper[4716]: I1207 16:21:02.728827 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-6kcdj" Dec 07 16:21:02 crc kubenswrapper[4716]: I1207 16:21:02.763278 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6j4h\" (UniqueName: \"kubernetes.io/projected/7a587b1d-a86e-4ee2-907e-6b014c61b09b-kube-api-access-d6j4h\") pod \"7a587b1d-a86e-4ee2-907e-6b014c61b09b\" (UID: \"7a587b1d-a86e-4ee2-907e-6b014c61b09b\") " Dec 07 16:21:02 crc kubenswrapper[4716]: I1207 16:21:02.763515 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a587b1d-a86e-4ee2-907e-6b014c61b09b-config-data\") pod \"7a587b1d-a86e-4ee2-907e-6b014c61b09b\" (UID: \"7a587b1d-a86e-4ee2-907e-6b014c61b09b\") " Dec 07 16:21:02 crc kubenswrapper[4716]: I1207 16:21:02.763543 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a587b1d-a86e-4ee2-907e-6b014c61b09b-scripts\") pod \"7a587b1d-a86e-4ee2-907e-6b014c61b09b\" (UID: \"7a587b1d-a86e-4ee2-907e-6b014c61b09b\") " Dec 07 16:21:02 crc kubenswrapper[4716]: I1207 16:21:02.763650 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a587b1d-a86e-4ee2-907e-6b014c61b09b-combined-ca-bundle\") pod \"7a587b1d-a86e-4ee2-907e-6b014c61b09b\" (UID: \"7a587b1d-a86e-4ee2-907e-6b014c61b09b\") " Dec 07 16:21:02 crc kubenswrapper[4716]: I1207 16:21:02.774312 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a587b1d-a86e-4ee2-907e-6b014c61b09b-kube-api-access-d6j4h" (OuterVolumeSpecName: "kube-api-access-d6j4h") pod "7a587b1d-a86e-4ee2-907e-6b014c61b09b" (UID: "7a587b1d-a86e-4ee2-907e-6b014c61b09b"). InnerVolumeSpecName "kube-api-access-d6j4h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:21:02 crc kubenswrapper[4716]: I1207 16:21:02.787243 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a587b1d-a86e-4ee2-907e-6b014c61b09b-scripts" (OuterVolumeSpecName: "scripts") pod "7a587b1d-a86e-4ee2-907e-6b014c61b09b" (UID: "7a587b1d-a86e-4ee2-907e-6b014c61b09b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:21:02 crc kubenswrapper[4716]: I1207 16:21:02.816805 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a587b1d-a86e-4ee2-907e-6b014c61b09b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7a587b1d-a86e-4ee2-907e-6b014c61b09b" (UID: "7a587b1d-a86e-4ee2-907e-6b014c61b09b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:21:02 crc kubenswrapper[4716]: I1207 16:21:02.822603 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a587b1d-a86e-4ee2-907e-6b014c61b09b-config-data" (OuterVolumeSpecName: "config-data") pod "7a587b1d-a86e-4ee2-907e-6b014c61b09b" (UID: "7a587b1d-a86e-4ee2-907e-6b014c61b09b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:21:02 crc kubenswrapper[4716]: I1207 16:21:02.865688 4716 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a587b1d-a86e-4ee2-907e-6b014c61b09b-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 16:21:02 crc kubenswrapper[4716]: I1207 16:21:02.865852 4716 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a587b1d-a86e-4ee2-907e-6b014c61b09b-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 16:21:02 crc kubenswrapper[4716]: I1207 16:21:02.865909 4716 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a587b1d-a86e-4ee2-907e-6b014c61b09b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 16:21:02 crc kubenswrapper[4716]: I1207 16:21:02.865973 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6j4h\" (UniqueName: \"kubernetes.io/projected/7a587b1d-a86e-4ee2-907e-6b014c61b09b-kube-api-access-d6j4h\") on node \"crc\" DevicePath \"\"" Dec 07 16:21:03 crc kubenswrapper[4716]: I1207 16:21:03.361980 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-6kcdj" event={"ID":"7a587b1d-a86e-4ee2-907e-6b014c61b09b","Type":"ContainerDied","Data":"a86371ed238c219743c0204426f459d4c2eaaada8d805204592093b84261ccad"} Dec 07 16:21:03 crc kubenswrapper[4716]: I1207 16:21:03.362051 4716 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a86371ed238c219743c0204426f459d4c2eaaada8d805204592093b84261ccad" Dec 07 16:21:03 crc kubenswrapper[4716]: I1207 16:21:03.362228 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-6kcdj" Dec 07 16:21:03 crc kubenswrapper[4716]: I1207 16:21:03.532348 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 07 16:21:03 crc kubenswrapper[4716]: E1207 16:21:03.532801 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a587b1d-a86e-4ee2-907e-6b014c61b09b" containerName="nova-cell0-conductor-db-sync" Dec 07 16:21:03 crc kubenswrapper[4716]: I1207 16:21:03.532825 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a587b1d-a86e-4ee2-907e-6b014c61b09b" containerName="nova-cell0-conductor-db-sync" Dec 07 16:21:03 crc kubenswrapper[4716]: I1207 16:21:03.533120 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a587b1d-a86e-4ee2-907e-6b014c61b09b" containerName="nova-cell0-conductor-db-sync" Dec 07 16:21:03 crc kubenswrapper[4716]: I1207 16:21:03.533772 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 07 16:21:03 crc kubenswrapper[4716]: I1207 16:21:03.535702 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-j5z2b" Dec 07 16:21:03 crc kubenswrapper[4716]: I1207 16:21:03.535989 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 07 16:21:03 crc kubenswrapper[4716]: I1207 16:21:03.546948 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 07 16:21:03 crc kubenswrapper[4716]: I1207 16:21:03.580883 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3643cef-3c95-4ecc-9225-ebf929947127-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"b3643cef-3c95-4ecc-9225-ebf929947127\") " pod="openstack/nova-cell0-conductor-0" Dec 07 16:21:03 crc kubenswrapper[4716]: I1207 16:21:03.581133 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8bwr4\" (UniqueName: \"kubernetes.io/projected/b3643cef-3c95-4ecc-9225-ebf929947127-kube-api-access-8bwr4\") pod \"nova-cell0-conductor-0\" (UID: \"b3643cef-3c95-4ecc-9225-ebf929947127\") " pod="openstack/nova-cell0-conductor-0" Dec 07 16:21:03 crc kubenswrapper[4716]: I1207 16:21:03.581404 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3643cef-3c95-4ecc-9225-ebf929947127-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"b3643cef-3c95-4ecc-9225-ebf929947127\") " pod="openstack/nova-cell0-conductor-0" Dec 07 16:21:03 crc kubenswrapper[4716]: I1207 16:21:03.682190 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3643cef-3c95-4ecc-9225-ebf929947127-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"b3643cef-3c95-4ecc-9225-ebf929947127\") " pod="openstack/nova-cell0-conductor-0" Dec 07 16:21:03 crc kubenswrapper[4716]: I1207 16:21:03.682273 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3643cef-3c95-4ecc-9225-ebf929947127-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"b3643cef-3c95-4ecc-9225-ebf929947127\") " pod="openstack/nova-cell0-conductor-0" Dec 07 16:21:03 crc kubenswrapper[4716]: I1207 16:21:03.682358 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8bwr4\" (UniqueName: \"kubernetes.io/projected/b3643cef-3c95-4ecc-9225-ebf929947127-kube-api-access-8bwr4\") pod \"nova-cell0-conductor-0\" (UID: \"b3643cef-3c95-4ecc-9225-ebf929947127\") " pod="openstack/nova-cell0-conductor-0" Dec 07 16:21:03 crc kubenswrapper[4716]: I1207 16:21:03.689148 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3643cef-3c95-4ecc-9225-ebf929947127-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"b3643cef-3c95-4ecc-9225-ebf929947127\") " pod="openstack/nova-cell0-conductor-0" Dec 07 16:21:03 crc kubenswrapper[4716]: I1207 16:21:03.696362 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3643cef-3c95-4ecc-9225-ebf929947127-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"b3643cef-3c95-4ecc-9225-ebf929947127\") " pod="openstack/nova-cell0-conductor-0" Dec 07 16:21:03 crc kubenswrapper[4716]: I1207 16:21:03.701050 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8bwr4\" (UniqueName: \"kubernetes.io/projected/b3643cef-3c95-4ecc-9225-ebf929947127-kube-api-access-8bwr4\") pod \"nova-cell0-conductor-0\" (UID: \"b3643cef-3c95-4ecc-9225-ebf929947127\") " pod="openstack/nova-cell0-conductor-0" Dec 07 16:21:03 crc kubenswrapper[4716]: I1207 16:21:03.848912 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 07 16:21:04 crc kubenswrapper[4716]: I1207 16:21:04.332803 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 07 16:21:04 crc kubenswrapper[4716]: I1207 16:21:04.372284 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"b3643cef-3c95-4ecc-9225-ebf929947127","Type":"ContainerStarted","Data":"efb19ef3dbb70db77617900fa488dcb80860d2f52c8c15c6253788a2369b6179"} Dec 07 16:21:05 crc kubenswrapper[4716]: I1207 16:21:05.400052 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"b3643cef-3c95-4ecc-9225-ebf929947127","Type":"ContainerStarted","Data":"6bdc9d0733873bc6d7471d4214c8e71211371f650fbd01b2ec4494c87334b0ee"} Dec 07 16:21:05 crc kubenswrapper[4716]: I1207 16:21:05.400414 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 07 16:21:05 crc kubenswrapper[4716]: I1207 16:21:05.424657 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.4246320089999998 podStartE2EDuration="2.424632009s" podCreationTimestamp="2025-12-07 16:21:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:21:05.421984567 +0000 UTC m=+1128.112269549" watchObservedRunningTime="2025-12-07 16:21:05.424632009 +0000 UTC m=+1128.114916961" Dec 07 16:21:13 crc kubenswrapper[4716]: I1207 16:21:13.904247 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.442255 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-q9fx7"] Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.444357 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-q9fx7" Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.446468 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.447227 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.453902 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-q9fx7"] Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.519996 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/367f54c1-6d21-41fc-8c08-d6a7e9e637cf-scripts\") pod \"nova-cell0-cell-mapping-q9fx7\" (UID: \"367f54c1-6d21-41fc-8c08-d6a7e9e637cf\") " pod="openstack/nova-cell0-cell-mapping-q9fx7" Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.523406 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/367f54c1-6d21-41fc-8c08-d6a7e9e637cf-config-data\") pod \"nova-cell0-cell-mapping-q9fx7\" (UID: \"367f54c1-6d21-41fc-8c08-d6a7e9e637cf\") " pod="openstack/nova-cell0-cell-mapping-q9fx7" Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.524155 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/367f54c1-6d21-41fc-8c08-d6a7e9e637cf-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-q9fx7\" (UID: \"367f54c1-6d21-41fc-8c08-d6a7e9e637cf\") " pod="openstack/nova-cell0-cell-mapping-q9fx7" Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.524577 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66nfk\" (UniqueName: \"kubernetes.io/projected/367f54c1-6d21-41fc-8c08-d6a7e9e637cf-kube-api-access-66nfk\") pod \"nova-cell0-cell-mapping-q9fx7\" (UID: \"367f54c1-6d21-41fc-8c08-d6a7e9e637cf\") " pod="openstack/nova-cell0-cell-mapping-q9fx7" Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.637443 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66nfk\" (UniqueName: \"kubernetes.io/projected/367f54c1-6d21-41fc-8c08-d6a7e9e637cf-kube-api-access-66nfk\") pod \"nova-cell0-cell-mapping-q9fx7\" (UID: \"367f54c1-6d21-41fc-8c08-d6a7e9e637cf\") " pod="openstack/nova-cell0-cell-mapping-q9fx7" Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.638313 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/367f54c1-6d21-41fc-8c08-d6a7e9e637cf-scripts\") pod \"nova-cell0-cell-mapping-q9fx7\" (UID: \"367f54c1-6d21-41fc-8c08-d6a7e9e637cf\") " pod="openstack/nova-cell0-cell-mapping-q9fx7" Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.638495 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/367f54c1-6d21-41fc-8c08-d6a7e9e637cf-config-data\") pod \"nova-cell0-cell-mapping-q9fx7\" (UID: \"367f54c1-6d21-41fc-8c08-d6a7e9e637cf\") " pod="openstack/nova-cell0-cell-mapping-q9fx7" Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.638612 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/367f54c1-6d21-41fc-8c08-d6a7e9e637cf-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-q9fx7\" (UID: \"367f54c1-6d21-41fc-8c08-d6a7e9e637cf\") " pod="openstack/nova-cell0-cell-mapping-q9fx7" Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.649203 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/367f54c1-6d21-41fc-8c08-d6a7e9e637cf-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-q9fx7\" (UID: \"367f54c1-6d21-41fc-8c08-d6a7e9e637cf\") " pod="openstack/nova-cell0-cell-mapping-q9fx7" Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.653452 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/367f54c1-6d21-41fc-8c08-d6a7e9e637cf-scripts\") pod \"nova-cell0-cell-mapping-q9fx7\" (UID: \"367f54c1-6d21-41fc-8c08-d6a7e9e637cf\") " pod="openstack/nova-cell0-cell-mapping-q9fx7" Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.657267 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.662112 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.674518 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.678107 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.679038 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/367f54c1-6d21-41fc-8c08-d6a7e9e637cf-config-data\") pod \"nova-cell0-cell-mapping-q9fx7\" (UID: \"367f54c1-6d21-41fc-8c08-d6a7e9e637cf\") " pod="openstack/nova-cell0-cell-mapping-q9fx7" Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.689398 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66nfk\" (UniqueName: \"kubernetes.io/projected/367f54c1-6d21-41fc-8c08-d6a7e9e637cf-kube-api-access-66nfk\") pod \"nova-cell0-cell-mapping-q9fx7\" (UID: \"367f54c1-6d21-41fc-8c08-d6a7e9e637cf\") " pod="openstack/nova-cell0-cell-mapping-q9fx7" Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.730145 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.731825 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.734168 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.775982 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.816181 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-v4tqc"] Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.817854 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-v4tqc" Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.832392 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-q9fx7" Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.834789 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.836325 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.846467 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.848986 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/77330671-85cf-40c8-81b0-7a7fe1f480a6-logs\") pod \"nova-api-0\" (UID: \"77330671-85cf-40c8-81b0-7a7fe1f480a6\") " pod="openstack/nova-api-0" Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.849015 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/572389a3-395c-4626-b731-fda219d74c96-config-data\") pod \"nova-metadata-0\" (UID: \"572389a3-395c-4626-b731-fda219d74c96\") " pod="openstack/nova-metadata-0" Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.849038 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77330671-85cf-40c8-81b0-7a7fe1f480a6-config-data\") pod \"nova-api-0\" (UID: \"77330671-85cf-40c8-81b0-7a7fe1f480a6\") " pod="openstack/nova-api-0" Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.849096 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mc4tf\" (UniqueName: \"kubernetes.io/projected/77330671-85cf-40c8-81b0-7a7fe1f480a6-kube-api-access-mc4tf\") pod \"nova-api-0\" (UID: \"77330671-85cf-40c8-81b0-7a7fe1f480a6\") " pod="openstack/nova-api-0" Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.849114 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77330671-85cf-40c8-81b0-7a7fe1f480a6-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"77330671-85cf-40c8-81b0-7a7fe1f480a6\") " pod="openstack/nova-api-0" Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.849149 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37636129-36c0-4531-9a84-8b1d177a3e44-config\") pod \"dnsmasq-dns-757b4f8459-v4tqc\" (UID: \"37636129-36c0-4531-9a84-8b1d177a3e44\") " pod="openstack/dnsmasq-dns-757b4f8459-v4tqc" Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.849180 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wjmhp\" (UniqueName: \"kubernetes.io/projected/572389a3-395c-4626-b731-fda219d74c96-kube-api-access-wjmhp\") pod \"nova-metadata-0\" (UID: \"572389a3-395c-4626-b731-fda219d74c96\") " pod="openstack/nova-metadata-0" Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.849206 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/37636129-36c0-4531-9a84-8b1d177a3e44-dns-svc\") pod \"dnsmasq-dns-757b4f8459-v4tqc\" (UID: \"37636129-36c0-4531-9a84-8b1d177a3e44\") " pod="openstack/dnsmasq-dns-757b4f8459-v4tqc" Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.849236 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/37636129-36c0-4531-9a84-8b1d177a3e44-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-v4tqc\" (UID: \"37636129-36c0-4531-9a84-8b1d177a3e44\") " pod="openstack/dnsmasq-dns-757b4f8459-v4tqc" Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.849255 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/572389a3-395c-4626-b731-fda219d74c96-logs\") pod \"nova-metadata-0\" (UID: \"572389a3-395c-4626-b731-fda219d74c96\") " pod="openstack/nova-metadata-0" Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.849271 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a88434a-c17e-4de2-be0e-bf5c951012b2-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"3a88434a-c17e-4de2-be0e-bf5c951012b2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.849297 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/572389a3-395c-4626-b731-fda219d74c96-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"572389a3-395c-4626-b731-fda219d74c96\") " pod="openstack/nova-metadata-0" Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.849311 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/37636129-36c0-4531-9a84-8b1d177a3e44-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-v4tqc\" (UID: \"37636129-36c0-4531-9a84-8b1d177a3e44\") " pod="openstack/dnsmasq-dns-757b4f8459-v4tqc" Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.849350 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-brw67\" (UniqueName: \"kubernetes.io/projected/37636129-36c0-4531-9a84-8b1d177a3e44-kube-api-access-brw67\") pod \"dnsmasq-dns-757b4f8459-v4tqc\" (UID: \"37636129-36c0-4531-9a84-8b1d177a3e44\") " pod="openstack/dnsmasq-dns-757b4f8459-v4tqc" Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.849368 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a88434a-c17e-4de2-be0e-bf5c951012b2-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"3a88434a-c17e-4de2-be0e-bf5c951012b2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.849384 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/37636129-36c0-4531-9a84-8b1d177a3e44-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-v4tqc\" (UID: \"37636129-36c0-4531-9a84-8b1d177a3e44\") " pod="openstack/dnsmasq-dns-757b4f8459-v4tqc" Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.849405 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdpn8\" (UniqueName: \"kubernetes.io/projected/3a88434a-c17e-4de2-be0e-bf5c951012b2-kube-api-access-vdpn8\") pod \"nova-cell1-novncproxy-0\" (UID: \"3a88434a-c17e-4de2-be0e-bf5c951012b2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.888948 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-v4tqc"] Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.950505 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/572389a3-395c-4626-b731-fda219d74c96-logs\") pod \"nova-metadata-0\" (UID: \"572389a3-395c-4626-b731-fda219d74c96\") " pod="openstack/nova-metadata-0" Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.951487 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a88434a-c17e-4de2-be0e-bf5c951012b2-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"3a88434a-c17e-4de2-be0e-bf5c951012b2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.951579 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/572389a3-395c-4626-b731-fda219d74c96-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"572389a3-395c-4626-b731-fda219d74c96\") " pod="openstack/nova-metadata-0" Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.951671 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/37636129-36c0-4531-9a84-8b1d177a3e44-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-v4tqc\" (UID: \"37636129-36c0-4531-9a84-8b1d177a3e44\") " pod="openstack/dnsmasq-dns-757b4f8459-v4tqc" Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.951784 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-brw67\" (UniqueName: \"kubernetes.io/projected/37636129-36c0-4531-9a84-8b1d177a3e44-kube-api-access-brw67\") pod \"dnsmasq-dns-757b4f8459-v4tqc\" (UID: \"37636129-36c0-4531-9a84-8b1d177a3e44\") " pod="openstack/dnsmasq-dns-757b4f8459-v4tqc" Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.951873 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/37636129-36c0-4531-9a84-8b1d177a3e44-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-v4tqc\" (UID: \"37636129-36c0-4531-9a84-8b1d177a3e44\") " pod="openstack/dnsmasq-dns-757b4f8459-v4tqc" Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.951958 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a88434a-c17e-4de2-be0e-bf5c951012b2-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"3a88434a-c17e-4de2-be0e-bf5c951012b2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.952041 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdpn8\" (UniqueName: \"kubernetes.io/projected/3a88434a-c17e-4de2-be0e-bf5c951012b2-kube-api-access-vdpn8\") pod \"nova-cell1-novncproxy-0\" (UID: \"3a88434a-c17e-4de2-be0e-bf5c951012b2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.952148 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/77330671-85cf-40c8-81b0-7a7fe1f480a6-logs\") pod \"nova-api-0\" (UID: \"77330671-85cf-40c8-81b0-7a7fe1f480a6\") " pod="openstack/nova-api-0" Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.952222 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/572389a3-395c-4626-b731-fda219d74c96-config-data\") pod \"nova-metadata-0\" (UID: \"572389a3-395c-4626-b731-fda219d74c96\") " pod="openstack/nova-metadata-0" Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.952301 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77330671-85cf-40c8-81b0-7a7fe1f480a6-config-data\") pod \"nova-api-0\" (UID: \"77330671-85cf-40c8-81b0-7a7fe1f480a6\") " pod="openstack/nova-api-0" Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.952390 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mc4tf\" (UniqueName: \"kubernetes.io/projected/77330671-85cf-40c8-81b0-7a7fe1f480a6-kube-api-access-mc4tf\") pod \"nova-api-0\" (UID: \"77330671-85cf-40c8-81b0-7a7fe1f480a6\") " pod="openstack/nova-api-0" Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.952474 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77330671-85cf-40c8-81b0-7a7fe1f480a6-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"77330671-85cf-40c8-81b0-7a7fe1f480a6\") " pod="openstack/nova-api-0" Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.952571 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37636129-36c0-4531-9a84-8b1d177a3e44-config\") pod \"dnsmasq-dns-757b4f8459-v4tqc\" (UID: \"37636129-36c0-4531-9a84-8b1d177a3e44\") " pod="openstack/dnsmasq-dns-757b4f8459-v4tqc" Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.952657 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wjmhp\" (UniqueName: \"kubernetes.io/projected/572389a3-395c-4626-b731-fda219d74c96-kube-api-access-wjmhp\") pod \"nova-metadata-0\" (UID: \"572389a3-395c-4626-b731-fda219d74c96\") " pod="openstack/nova-metadata-0" Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.952736 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/37636129-36c0-4531-9a84-8b1d177a3e44-dns-svc\") pod \"dnsmasq-dns-757b4f8459-v4tqc\" (UID: \"37636129-36c0-4531-9a84-8b1d177a3e44\") " pod="openstack/dnsmasq-dns-757b4f8459-v4tqc" Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.952898 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/37636129-36c0-4531-9a84-8b1d177a3e44-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-v4tqc\" (UID: \"37636129-36c0-4531-9a84-8b1d177a3e44\") " pod="openstack/dnsmasq-dns-757b4f8459-v4tqc" Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.954831 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/77330671-85cf-40c8-81b0-7a7fe1f480a6-logs\") pod \"nova-api-0\" (UID: \"77330671-85cf-40c8-81b0-7a7fe1f480a6\") " pod="openstack/nova-api-0" Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.958107 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37636129-36c0-4531-9a84-8b1d177a3e44-config\") pod \"dnsmasq-dns-757b4f8459-v4tqc\" (UID: \"37636129-36c0-4531-9a84-8b1d177a3e44\") " pod="openstack/dnsmasq-dns-757b4f8459-v4tqc" Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.961880 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/37636129-36c0-4531-9a84-8b1d177a3e44-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-v4tqc\" (UID: \"37636129-36c0-4531-9a84-8b1d177a3e44\") " pod="openstack/dnsmasq-dns-757b4f8459-v4tqc" Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.963959 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/37636129-36c0-4531-9a84-8b1d177a3e44-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-v4tqc\" (UID: \"37636129-36c0-4531-9a84-8b1d177a3e44\") " pod="openstack/dnsmasq-dns-757b4f8459-v4tqc" Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.964038 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/572389a3-395c-4626-b731-fda219d74c96-config-data\") pod \"nova-metadata-0\" (UID: \"572389a3-395c-4626-b731-fda219d74c96\") " pod="openstack/nova-metadata-0" Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.964374 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/572389a3-395c-4626-b731-fda219d74c96-logs\") pod \"nova-metadata-0\" (UID: \"572389a3-395c-4626-b731-fda219d74c96\") " pod="openstack/nova-metadata-0" Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.965160 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/37636129-36c0-4531-9a84-8b1d177a3e44-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-v4tqc\" (UID: \"37636129-36c0-4531-9a84-8b1d177a3e44\") " pod="openstack/dnsmasq-dns-757b4f8459-v4tqc" Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.967065 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/572389a3-395c-4626-b731-fda219d74c96-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"572389a3-395c-4626-b731-fda219d74c96\") " pod="openstack/nova-metadata-0" Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.977716 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.981302 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/37636129-36c0-4531-9a84-8b1d177a3e44-dns-svc\") pod \"dnsmasq-dns-757b4f8459-v4tqc\" (UID: \"37636129-36c0-4531-9a84-8b1d177a3e44\") " pod="openstack/dnsmasq-dns-757b4f8459-v4tqc" Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.981595 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77330671-85cf-40c8-81b0-7a7fe1f480a6-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"77330671-85cf-40c8-81b0-7a7fe1f480a6\") " pod="openstack/nova-api-0" Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.981905 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a88434a-c17e-4de2-be0e-bf5c951012b2-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"3a88434a-c17e-4de2-be0e-bf5c951012b2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.992063 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77330671-85cf-40c8-81b0-7a7fe1f480a6-config-data\") pod \"nova-api-0\" (UID: \"77330671-85cf-40c8-81b0-7a7fe1f480a6\") " pod="openstack/nova-api-0" Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.995647 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 07 16:21:14 crc kubenswrapper[4716]: I1207 16:21:14.996971 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 07 16:21:15 crc kubenswrapper[4716]: I1207 16:21:15.007862 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 07 16:21:15 crc kubenswrapper[4716]: I1207 16:21:15.008041 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-brw67\" (UniqueName: \"kubernetes.io/projected/37636129-36c0-4531-9a84-8b1d177a3e44-kube-api-access-brw67\") pod \"dnsmasq-dns-757b4f8459-v4tqc\" (UID: \"37636129-36c0-4531-9a84-8b1d177a3e44\") " pod="openstack/dnsmasq-dns-757b4f8459-v4tqc" Dec 07 16:21:15 crc kubenswrapper[4716]: I1207 16:21:15.008042 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 07 16:21:15 crc kubenswrapper[4716]: I1207 16:21:15.011249 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a88434a-c17e-4de2-be0e-bf5c951012b2-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"3a88434a-c17e-4de2-be0e-bf5c951012b2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 07 16:21:15 crc kubenswrapper[4716]: I1207 16:21:15.011809 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wjmhp\" (UniqueName: \"kubernetes.io/projected/572389a3-395c-4626-b731-fda219d74c96-kube-api-access-wjmhp\") pod \"nova-metadata-0\" (UID: \"572389a3-395c-4626-b731-fda219d74c96\") " pod="openstack/nova-metadata-0" Dec 07 16:21:15 crc kubenswrapper[4716]: I1207 16:21:15.015619 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdpn8\" (UniqueName: \"kubernetes.io/projected/3a88434a-c17e-4de2-be0e-bf5c951012b2-kube-api-access-vdpn8\") pod \"nova-cell1-novncproxy-0\" (UID: \"3a88434a-c17e-4de2-be0e-bf5c951012b2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 07 16:21:15 crc kubenswrapper[4716]: I1207 16:21:15.032677 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mc4tf\" (UniqueName: \"kubernetes.io/projected/77330671-85cf-40c8-81b0-7a7fe1f480a6-kube-api-access-mc4tf\") pod \"nova-api-0\" (UID: \"77330671-85cf-40c8-81b0-7a7fe1f480a6\") " pod="openstack/nova-api-0" Dec 07 16:21:15 crc kubenswrapper[4716]: I1207 16:21:15.065881 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 07 16:21:15 crc kubenswrapper[4716]: I1207 16:21:15.083989 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 07 16:21:15 crc kubenswrapper[4716]: I1207 16:21:15.117827 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 07 16:21:15 crc kubenswrapper[4716]: I1207 16:21:15.138661 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-v4tqc" Dec 07 16:21:15 crc kubenswrapper[4716]: I1207 16:21:15.158836 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bb8zt\" (UniqueName: \"kubernetes.io/projected/c9ffe38d-2cbd-40fa-ae1d-12b42e721bb4-kube-api-access-bb8zt\") pod \"nova-scheduler-0\" (UID: \"c9ffe38d-2cbd-40fa-ae1d-12b42e721bb4\") " pod="openstack/nova-scheduler-0" Dec 07 16:21:15 crc kubenswrapper[4716]: I1207 16:21:15.158899 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9ffe38d-2cbd-40fa-ae1d-12b42e721bb4-config-data\") pod \"nova-scheduler-0\" (UID: \"c9ffe38d-2cbd-40fa-ae1d-12b42e721bb4\") " pod="openstack/nova-scheduler-0" Dec 07 16:21:15 crc kubenswrapper[4716]: I1207 16:21:15.158943 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9ffe38d-2cbd-40fa-ae1d-12b42e721bb4-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c9ffe38d-2cbd-40fa-ae1d-12b42e721bb4\") " pod="openstack/nova-scheduler-0" Dec 07 16:21:15 crc kubenswrapper[4716]: I1207 16:21:15.260399 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bb8zt\" (UniqueName: \"kubernetes.io/projected/c9ffe38d-2cbd-40fa-ae1d-12b42e721bb4-kube-api-access-bb8zt\") pod \"nova-scheduler-0\" (UID: \"c9ffe38d-2cbd-40fa-ae1d-12b42e721bb4\") " pod="openstack/nova-scheduler-0" Dec 07 16:21:15 crc kubenswrapper[4716]: I1207 16:21:15.260463 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9ffe38d-2cbd-40fa-ae1d-12b42e721bb4-config-data\") pod \"nova-scheduler-0\" (UID: \"c9ffe38d-2cbd-40fa-ae1d-12b42e721bb4\") " pod="openstack/nova-scheduler-0" Dec 07 16:21:15 crc kubenswrapper[4716]: I1207 16:21:15.260509 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9ffe38d-2cbd-40fa-ae1d-12b42e721bb4-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c9ffe38d-2cbd-40fa-ae1d-12b42e721bb4\") " pod="openstack/nova-scheduler-0" Dec 07 16:21:15 crc kubenswrapper[4716]: I1207 16:21:15.268025 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9ffe38d-2cbd-40fa-ae1d-12b42e721bb4-config-data\") pod \"nova-scheduler-0\" (UID: \"c9ffe38d-2cbd-40fa-ae1d-12b42e721bb4\") " pod="openstack/nova-scheduler-0" Dec 07 16:21:15 crc kubenswrapper[4716]: I1207 16:21:15.278175 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9ffe38d-2cbd-40fa-ae1d-12b42e721bb4-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c9ffe38d-2cbd-40fa-ae1d-12b42e721bb4\") " pod="openstack/nova-scheduler-0" Dec 07 16:21:15 crc kubenswrapper[4716]: I1207 16:21:15.286125 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bb8zt\" (UniqueName: \"kubernetes.io/projected/c9ffe38d-2cbd-40fa-ae1d-12b42e721bb4-kube-api-access-bb8zt\") pod \"nova-scheduler-0\" (UID: \"c9ffe38d-2cbd-40fa-ae1d-12b42e721bb4\") " pod="openstack/nova-scheduler-0" Dec 07 16:21:15 crc kubenswrapper[4716]: I1207 16:21:15.383806 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 07 16:21:15 crc kubenswrapper[4716]: I1207 16:21:15.487112 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-q9fx7"] Dec 07 16:21:15 crc kubenswrapper[4716]: I1207 16:21:15.549797 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-q9fx7" event={"ID":"367f54c1-6d21-41fc-8c08-d6a7e9e637cf","Type":"ContainerStarted","Data":"1b53dde7bac526b271c3852b6a52649fb6944a1753214a8dc2a3388051b27610"} Dec 07 16:21:15 crc kubenswrapper[4716]: I1207 16:21:15.695608 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 07 16:21:15 crc kubenswrapper[4716]: I1207 16:21:15.729371 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-x9skx"] Dec 07 16:21:15 crc kubenswrapper[4716]: I1207 16:21:15.731125 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-x9skx" Dec 07 16:21:15 crc kubenswrapper[4716]: I1207 16:21:15.735567 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 07 16:21:15 crc kubenswrapper[4716]: I1207 16:21:15.735706 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Dec 07 16:21:15 crc kubenswrapper[4716]: I1207 16:21:15.751974 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 07 16:21:15 crc kubenswrapper[4716]: I1207 16:21:15.768099 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-x9skx"] Dec 07 16:21:15 crc kubenswrapper[4716]: W1207 16:21:15.819586 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod37636129_36c0_4531_9a84_8b1d177a3e44.slice/crio-76e0a50916b702e9b7f979a5c7ace73c30a7371585dcbf0fe578b859fd1c86f7 WatchSource:0}: Error finding container 76e0a50916b702e9b7f979a5c7ace73c30a7371585dcbf0fe578b859fd1c86f7: Status 404 returned error can't find the container with id 76e0a50916b702e9b7f979a5c7ace73c30a7371585dcbf0fe578b859fd1c86f7 Dec 07 16:21:15 crc kubenswrapper[4716]: I1207 16:21:15.820445 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-v4tqc"] Dec 07 16:21:15 crc kubenswrapper[4716]: I1207 16:21:15.875899 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a950343-f5b9-4a42-857b-7de8ef106453-config-data\") pod \"nova-cell1-conductor-db-sync-x9skx\" (UID: \"9a950343-f5b9-4a42-857b-7de8ef106453\") " pod="openstack/nova-cell1-conductor-db-sync-x9skx" Dec 07 16:21:15 crc kubenswrapper[4716]: I1207 16:21:15.875983 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a950343-f5b9-4a42-857b-7de8ef106453-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-x9skx\" (UID: \"9a950343-f5b9-4a42-857b-7de8ef106453\") " pod="openstack/nova-cell1-conductor-db-sync-x9skx" Dec 07 16:21:15 crc kubenswrapper[4716]: I1207 16:21:15.876039 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j8dtp\" (UniqueName: \"kubernetes.io/projected/9a950343-f5b9-4a42-857b-7de8ef106453-kube-api-access-j8dtp\") pod \"nova-cell1-conductor-db-sync-x9skx\" (UID: \"9a950343-f5b9-4a42-857b-7de8ef106453\") " pod="openstack/nova-cell1-conductor-db-sync-x9skx" Dec 07 16:21:15 crc kubenswrapper[4716]: I1207 16:21:15.876073 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a950343-f5b9-4a42-857b-7de8ef106453-scripts\") pod \"nova-cell1-conductor-db-sync-x9skx\" (UID: \"9a950343-f5b9-4a42-857b-7de8ef106453\") " pod="openstack/nova-cell1-conductor-db-sync-x9skx" Dec 07 16:21:15 crc kubenswrapper[4716]: I1207 16:21:15.905718 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 07 16:21:15 crc kubenswrapper[4716]: I1207 16:21:15.977456 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j8dtp\" (UniqueName: \"kubernetes.io/projected/9a950343-f5b9-4a42-857b-7de8ef106453-kube-api-access-j8dtp\") pod \"nova-cell1-conductor-db-sync-x9skx\" (UID: \"9a950343-f5b9-4a42-857b-7de8ef106453\") " pod="openstack/nova-cell1-conductor-db-sync-x9skx" Dec 07 16:21:15 crc kubenswrapper[4716]: I1207 16:21:15.978892 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a950343-f5b9-4a42-857b-7de8ef106453-scripts\") pod \"nova-cell1-conductor-db-sync-x9skx\" (UID: \"9a950343-f5b9-4a42-857b-7de8ef106453\") " pod="openstack/nova-cell1-conductor-db-sync-x9skx" Dec 07 16:21:15 crc kubenswrapper[4716]: I1207 16:21:15.979155 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a950343-f5b9-4a42-857b-7de8ef106453-config-data\") pod \"nova-cell1-conductor-db-sync-x9skx\" (UID: \"9a950343-f5b9-4a42-857b-7de8ef106453\") " pod="openstack/nova-cell1-conductor-db-sync-x9skx" Dec 07 16:21:15 crc kubenswrapper[4716]: I1207 16:21:15.979327 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a950343-f5b9-4a42-857b-7de8ef106453-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-x9skx\" (UID: \"9a950343-f5b9-4a42-857b-7de8ef106453\") " pod="openstack/nova-cell1-conductor-db-sync-x9skx" Dec 07 16:21:15 crc kubenswrapper[4716]: I1207 16:21:15.984005 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a950343-f5b9-4a42-857b-7de8ef106453-config-data\") pod \"nova-cell1-conductor-db-sync-x9skx\" (UID: \"9a950343-f5b9-4a42-857b-7de8ef106453\") " pod="openstack/nova-cell1-conductor-db-sync-x9skx" Dec 07 16:21:15 crc kubenswrapper[4716]: I1207 16:21:15.984471 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a950343-f5b9-4a42-857b-7de8ef106453-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-x9skx\" (UID: \"9a950343-f5b9-4a42-857b-7de8ef106453\") " pod="openstack/nova-cell1-conductor-db-sync-x9skx" Dec 07 16:21:15 crc kubenswrapper[4716]: I1207 16:21:15.984564 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a950343-f5b9-4a42-857b-7de8ef106453-scripts\") pod \"nova-cell1-conductor-db-sync-x9skx\" (UID: \"9a950343-f5b9-4a42-857b-7de8ef106453\") " pod="openstack/nova-cell1-conductor-db-sync-x9skx" Dec 07 16:21:16 crc kubenswrapper[4716]: I1207 16:21:16.000983 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j8dtp\" (UniqueName: \"kubernetes.io/projected/9a950343-f5b9-4a42-857b-7de8ef106453-kube-api-access-j8dtp\") pod \"nova-cell1-conductor-db-sync-x9skx\" (UID: \"9a950343-f5b9-4a42-857b-7de8ef106453\") " pod="openstack/nova-cell1-conductor-db-sync-x9skx" Dec 07 16:21:16 crc kubenswrapper[4716]: W1207 16:21:16.008484 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc9ffe38d_2cbd_40fa_ae1d_12b42e721bb4.slice/crio-7586c0dc54847acf81822090ab9fb6276beb12198e1da69556617403e0bdf093 WatchSource:0}: Error finding container 7586c0dc54847acf81822090ab9fb6276beb12198e1da69556617403e0bdf093: Status 404 returned error can't find the container with id 7586c0dc54847acf81822090ab9fb6276beb12198e1da69556617403e0bdf093 Dec 07 16:21:16 crc kubenswrapper[4716]: I1207 16:21:16.010486 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 07 16:21:16 crc kubenswrapper[4716]: I1207 16:21:16.204393 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-x9skx" Dec 07 16:21:16 crc kubenswrapper[4716]: I1207 16:21:16.586391 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"572389a3-395c-4626-b731-fda219d74c96","Type":"ContainerStarted","Data":"d7ba4c73dc04e9c1bc2f6c93abf17ef83bc468f155424122a7ce641c8d9cdc4c"} Dec 07 16:21:16 crc kubenswrapper[4716]: I1207 16:21:16.589347 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c9ffe38d-2cbd-40fa-ae1d-12b42e721bb4","Type":"ContainerStarted","Data":"7586c0dc54847acf81822090ab9fb6276beb12198e1da69556617403e0bdf093"} Dec 07 16:21:16 crc kubenswrapper[4716]: I1207 16:21:16.595364 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"77330671-85cf-40c8-81b0-7a7fe1f480a6","Type":"ContainerStarted","Data":"857f961b190f662b0adfc7243385fad57fb1d5aea733c82c637787b9daca6c51"} Dec 07 16:21:16 crc kubenswrapper[4716]: I1207 16:21:16.597382 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"3a88434a-c17e-4de2-be0e-bf5c951012b2","Type":"ContainerStarted","Data":"0414a171402f251ee3e00b77c595c00f52269d16e904f154f3a1627639cb68a7"} Dec 07 16:21:16 crc kubenswrapper[4716]: I1207 16:21:16.613628 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-q9fx7" event={"ID":"367f54c1-6d21-41fc-8c08-d6a7e9e637cf","Type":"ContainerStarted","Data":"f918233e5ff893608dff230031ceae9e1dfe9739e05fd2ab94b483e2a33df54f"} Dec 07 16:21:16 crc kubenswrapper[4716]: I1207 16:21:16.616695 4716 generic.go:334] "Generic (PLEG): container finished" podID="37636129-36c0-4531-9a84-8b1d177a3e44" containerID="3b7d6afaf096e65d78714050058efe812609c9f66136801dc763beb063a783b9" exitCode=0 Dec 07 16:21:16 crc kubenswrapper[4716]: I1207 16:21:16.616787 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-v4tqc" event={"ID":"37636129-36c0-4531-9a84-8b1d177a3e44","Type":"ContainerDied","Data":"3b7d6afaf096e65d78714050058efe812609c9f66136801dc763beb063a783b9"} Dec 07 16:21:16 crc kubenswrapper[4716]: I1207 16:21:16.617623 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-v4tqc" event={"ID":"37636129-36c0-4531-9a84-8b1d177a3e44","Type":"ContainerStarted","Data":"76e0a50916b702e9b7f979a5c7ace73c30a7371585dcbf0fe578b859fd1c86f7"} Dec 07 16:21:16 crc kubenswrapper[4716]: I1207 16:21:16.630838 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-q9fx7" podStartSLOduration=2.630818746 podStartE2EDuration="2.630818746s" podCreationTimestamp="2025-12-07 16:21:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:21:16.627778244 +0000 UTC m=+1139.318063156" watchObservedRunningTime="2025-12-07 16:21:16.630818746 +0000 UTC m=+1139.321103658" Dec 07 16:21:16 crc kubenswrapper[4716]: I1207 16:21:16.735046 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-x9skx"] Dec 07 16:21:17 crc kubenswrapper[4716]: I1207 16:21:17.706876 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-x9skx" event={"ID":"9a950343-f5b9-4a42-857b-7de8ef106453","Type":"ContainerStarted","Data":"6368cc2ab6c45969b8edd9d2e09cdeb711278ef76446cd516bd195c0d856f570"} Dec 07 16:21:17 crc kubenswrapper[4716]: I1207 16:21:17.707298 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-x9skx" event={"ID":"9a950343-f5b9-4a42-857b-7de8ef106453","Type":"ContainerStarted","Data":"6613c95297620627e30d0fdb5ed6d868bf89ba99de35c7f670d1757ca345ac97"} Dec 07 16:21:17 crc kubenswrapper[4716]: I1207 16:21:17.709521 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-v4tqc" event={"ID":"37636129-36c0-4531-9a84-8b1d177a3e44","Type":"ContainerStarted","Data":"f29b859032b292ca3ab15aecab24a7a3d23e155abccef6904cc68386cf80b02d"} Dec 07 16:21:17 crc kubenswrapper[4716]: I1207 16:21:17.709908 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-757b4f8459-v4tqc" Dec 07 16:21:17 crc kubenswrapper[4716]: I1207 16:21:17.794239 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-x9skx" podStartSLOduration=2.794220268 podStartE2EDuration="2.794220268s" podCreationTimestamp="2025-12-07 16:21:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:21:17.784684942 +0000 UTC m=+1140.474969854" watchObservedRunningTime="2025-12-07 16:21:17.794220268 +0000 UTC m=+1140.484505180" Dec 07 16:21:17 crc kubenswrapper[4716]: I1207 16:21:17.804496 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-757b4f8459-v4tqc" podStartSLOduration=3.804476865 podStartE2EDuration="3.804476865s" podCreationTimestamp="2025-12-07 16:21:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:21:17.802625275 +0000 UTC m=+1140.492910187" watchObservedRunningTime="2025-12-07 16:21:17.804476865 +0000 UTC m=+1140.494761777" Dec 07 16:21:18 crc kubenswrapper[4716]: I1207 16:21:18.320022 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 07 16:21:18 crc kubenswrapper[4716]: I1207 16:21:18.333629 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 07 16:21:19 crc kubenswrapper[4716]: I1207 16:21:19.727367 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"572389a3-395c-4626-b731-fda219d74c96","Type":"ContainerStarted","Data":"d2f2180238a1c794b6d0ab4bc199979f41c72940340f9c578e0cd46d1784dfee"} Dec 07 16:21:19 crc kubenswrapper[4716]: I1207 16:21:19.727931 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"572389a3-395c-4626-b731-fda219d74c96","Type":"ContainerStarted","Data":"667a06ce95001f73ad3dba6c78a5a8cb2f34ff586efb54823be406a44a5a19ce"} Dec 07 16:21:19 crc kubenswrapper[4716]: I1207 16:21:19.727521 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="572389a3-395c-4626-b731-fda219d74c96" containerName="nova-metadata-log" containerID="cri-o://667a06ce95001f73ad3dba6c78a5a8cb2f34ff586efb54823be406a44a5a19ce" gracePeriod=30 Dec 07 16:21:19 crc kubenswrapper[4716]: I1207 16:21:19.727496 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="572389a3-395c-4626-b731-fda219d74c96" containerName="nova-metadata-metadata" containerID="cri-o://d2f2180238a1c794b6d0ab4bc199979f41c72940340f9c578e0cd46d1784dfee" gracePeriod=30 Dec 07 16:21:19 crc kubenswrapper[4716]: I1207 16:21:19.729294 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c9ffe38d-2cbd-40fa-ae1d-12b42e721bb4","Type":"ContainerStarted","Data":"793fc688ad69e9d6f7b90a700e5b82626010662968c1cb60c6a290fd528be52f"} Dec 07 16:21:19 crc kubenswrapper[4716]: I1207 16:21:19.734019 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"77330671-85cf-40c8-81b0-7a7fe1f480a6","Type":"ContainerStarted","Data":"00c23e67a27f2b5f91cde6ffa7187c805202a9c4f2f7ad2f5c22239560158ca7"} Dec 07 16:21:19 crc kubenswrapper[4716]: I1207 16:21:19.734059 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"77330671-85cf-40c8-81b0-7a7fe1f480a6","Type":"ContainerStarted","Data":"9b05dea586ef734718ea6ec7ea96d56fa62614c09c073e3925b8fdaca52a7b63"} Dec 07 16:21:19 crc kubenswrapper[4716]: I1207 16:21:19.737628 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"3a88434a-c17e-4de2-be0e-bf5c951012b2","Type":"ContainerStarted","Data":"08d95b491dfc6c45ca2b0f45ca175db851104e371ff2993f44bc54dd1187a719"} Dec 07 16:21:19 crc kubenswrapper[4716]: I1207 16:21:19.737747 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="3a88434a-c17e-4de2-be0e-bf5c951012b2" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://08d95b491dfc6c45ca2b0f45ca175db851104e371ff2993f44bc54dd1187a719" gracePeriod=30 Dec 07 16:21:19 crc kubenswrapper[4716]: I1207 16:21:19.752617 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.7426648609999997 podStartE2EDuration="5.752597208s" podCreationTimestamp="2025-12-07 16:21:14 +0000 UTC" firstStartedPulling="2025-12-07 16:21:15.909215767 +0000 UTC m=+1138.599500679" lastFinishedPulling="2025-12-07 16:21:18.919148114 +0000 UTC m=+1141.609433026" observedRunningTime="2025-12-07 16:21:19.748919199 +0000 UTC m=+1142.439204111" watchObservedRunningTime="2025-12-07 16:21:19.752597208 +0000 UTC m=+1142.442882120" Dec 07 16:21:19 crc kubenswrapper[4716]: I1207 16:21:19.768943 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.570599593 podStartE2EDuration="5.768922919s" podCreationTimestamp="2025-12-07 16:21:14 +0000 UTC" firstStartedPulling="2025-12-07 16:21:15.720017606 +0000 UTC m=+1138.410302518" lastFinishedPulling="2025-12-07 16:21:18.918340932 +0000 UTC m=+1141.608625844" observedRunningTime="2025-12-07 16:21:19.764987283 +0000 UTC m=+1142.455272195" watchObservedRunningTime="2025-12-07 16:21:19.768922919 +0000 UTC m=+1142.459207841" Dec 07 16:21:19 crc kubenswrapper[4716]: I1207 16:21:19.791454 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.624328551 podStartE2EDuration="5.791435786s" podCreationTimestamp="2025-12-07 16:21:14 +0000 UTC" firstStartedPulling="2025-12-07 16:21:15.750999481 +0000 UTC m=+1138.441284393" lastFinishedPulling="2025-12-07 16:21:18.918106716 +0000 UTC m=+1141.608391628" observedRunningTime="2025-12-07 16:21:19.787974343 +0000 UTC m=+1142.478259255" watchObservedRunningTime="2025-12-07 16:21:19.791435786 +0000 UTC m=+1142.481720698" Dec 07 16:21:19 crc kubenswrapper[4716]: I1207 16:21:19.812630 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.89833517 podStartE2EDuration="5.812603547s" podCreationTimestamp="2025-12-07 16:21:14 +0000 UTC" firstStartedPulling="2025-12-07 16:21:16.01129481 +0000 UTC m=+1138.701579722" lastFinishedPulling="2025-12-07 16:21:18.925563187 +0000 UTC m=+1141.615848099" observedRunningTime="2025-12-07 16:21:19.803164822 +0000 UTC m=+1142.493449744" watchObservedRunningTime="2025-12-07 16:21:19.812603547 +0000 UTC m=+1142.502888459" Dec 07 16:21:20 crc kubenswrapper[4716]: I1207 16:21:20.084978 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 07 16:21:20 crc kubenswrapper[4716]: I1207 16:21:20.143550 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 07 16:21:20 crc kubenswrapper[4716]: I1207 16:21:20.143624 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 07 16:21:20 crc kubenswrapper[4716]: I1207 16:21:20.383891 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 07 16:21:20 crc kubenswrapper[4716]: I1207 16:21:20.384962 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 07 16:21:20 crc kubenswrapper[4716]: I1207 16:21:20.484347 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/572389a3-395c-4626-b731-fda219d74c96-combined-ca-bundle\") pod \"572389a3-395c-4626-b731-fda219d74c96\" (UID: \"572389a3-395c-4626-b731-fda219d74c96\") " Dec 07 16:21:20 crc kubenswrapper[4716]: I1207 16:21:20.484391 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/572389a3-395c-4626-b731-fda219d74c96-logs\") pod \"572389a3-395c-4626-b731-fda219d74c96\" (UID: \"572389a3-395c-4626-b731-fda219d74c96\") " Dec 07 16:21:20 crc kubenswrapper[4716]: I1207 16:21:20.484492 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wjmhp\" (UniqueName: \"kubernetes.io/projected/572389a3-395c-4626-b731-fda219d74c96-kube-api-access-wjmhp\") pod \"572389a3-395c-4626-b731-fda219d74c96\" (UID: \"572389a3-395c-4626-b731-fda219d74c96\") " Dec 07 16:21:20 crc kubenswrapper[4716]: I1207 16:21:20.484577 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/572389a3-395c-4626-b731-fda219d74c96-config-data\") pod \"572389a3-395c-4626-b731-fda219d74c96\" (UID: \"572389a3-395c-4626-b731-fda219d74c96\") " Dec 07 16:21:20 crc kubenswrapper[4716]: I1207 16:21:20.484837 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/572389a3-395c-4626-b731-fda219d74c96-logs" (OuterVolumeSpecName: "logs") pod "572389a3-395c-4626-b731-fda219d74c96" (UID: "572389a3-395c-4626-b731-fda219d74c96"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:21:20 crc kubenswrapper[4716]: I1207 16:21:20.485477 4716 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/572389a3-395c-4626-b731-fda219d74c96-logs\") on node \"crc\" DevicePath \"\"" Dec 07 16:21:20 crc kubenswrapper[4716]: I1207 16:21:20.492278 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/572389a3-395c-4626-b731-fda219d74c96-kube-api-access-wjmhp" (OuterVolumeSpecName: "kube-api-access-wjmhp") pod "572389a3-395c-4626-b731-fda219d74c96" (UID: "572389a3-395c-4626-b731-fda219d74c96"). InnerVolumeSpecName "kube-api-access-wjmhp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:21:20 crc kubenswrapper[4716]: I1207 16:21:20.513372 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/572389a3-395c-4626-b731-fda219d74c96-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "572389a3-395c-4626-b731-fda219d74c96" (UID: "572389a3-395c-4626-b731-fda219d74c96"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:21:20 crc kubenswrapper[4716]: I1207 16:21:20.526870 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/572389a3-395c-4626-b731-fda219d74c96-config-data" (OuterVolumeSpecName: "config-data") pod "572389a3-395c-4626-b731-fda219d74c96" (UID: "572389a3-395c-4626-b731-fda219d74c96"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:21:20 crc kubenswrapper[4716]: I1207 16:21:20.587521 4716 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/572389a3-395c-4626-b731-fda219d74c96-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 16:21:20 crc kubenswrapper[4716]: I1207 16:21:20.587561 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wjmhp\" (UniqueName: \"kubernetes.io/projected/572389a3-395c-4626-b731-fda219d74c96-kube-api-access-wjmhp\") on node \"crc\" DevicePath \"\"" Dec 07 16:21:20 crc kubenswrapper[4716]: I1207 16:21:20.587574 4716 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/572389a3-395c-4626-b731-fda219d74c96-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 16:21:20 crc kubenswrapper[4716]: I1207 16:21:20.750386 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 07 16:21:20 crc kubenswrapper[4716]: I1207 16:21:20.750790 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"572389a3-395c-4626-b731-fda219d74c96","Type":"ContainerDied","Data":"d2f2180238a1c794b6d0ab4bc199979f41c72940340f9c578e0cd46d1784dfee"} Dec 07 16:21:20 crc kubenswrapper[4716]: I1207 16:21:20.750948 4716 generic.go:334] "Generic (PLEG): container finished" podID="572389a3-395c-4626-b731-fda219d74c96" containerID="d2f2180238a1c794b6d0ab4bc199979f41c72940340f9c578e0cd46d1784dfee" exitCode=0 Dec 07 16:21:20 crc kubenswrapper[4716]: I1207 16:21:20.750982 4716 generic.go:334] "Generic (PLEG): container finished" podID="572389a3-395c-4626-b731-fda219d74c96" containerID="667a06ce95001f73ad3dba6c78a5a8cb2f34ff586efb54823be406a44a5a19ce" exitCode=143 Dec 07 16:21:20 crc kubenswrapper[4716]: I1207 16:21:20.751115 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"572389a3-395c-4626-b731-fda219d74c96","Type":"ContainerDied","Data":"667a06ce95001f73ad3dba6c78a5a8cb2f34ff586efb54823be406a44a5a19ce"} Dec 07 16:21:20 crc kubenswrapper[4716]: I1207 16:21:20.751197 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"572389a3-395c-4626-b731-fda219d74c96","Type":"ContainerDied","Data":"d7ba4c73dc04e9c1bc2f6c93abf17ef83bc468f155424122a7ce641c8d9cdc4c"} Dec 07 16:21:20 crc kubenswrapper[4716]: I1207 16:21:20.751512 4716 scope.go:117] "RemoveContainer" containerID="d2f2180238a1c794b6d0ab4bc199979f41c72940340f9c578e0cd46d1784dfee" Dec 07 16:21:20 crc kubenswrapper[4716]: I1207 16:21:20.795841 4716 scope.go:117] "RemoveContainer" containerID="667a06ce95001f73ad3dba6c78a5a8cb2f34ff586efb54823be406a44a5a19ce" Dec 07 16:21:20 crc kubenswrapper[4716]: I1207 16:21:20.813179 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 07 16:21:20 crc kubenswrapper[4716]: I1207 16:21:20.827578 4716 scope.go:117] "RemoveContainer" containerID="d2f2180238a1c794b6d0ab4bc199979f41c72940340f9c578e0cd46d1784dfee" Dec 07 16:21:20 crc kubenswrapper[4716]: E1207 16:21:20.828328 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d2f2180238a1c794b6d0ab4bc199979f41c72940340f9c578e0cd46d1784dfee\": container with ID starting with d2f2180238a1c794b6d0ab4bc199979f41c72940340f9c578e0cd46d1784dfee not found: ID does not exist" containerID="d2f2180238a1c794b6d0ab4bc199979f41c72940340f9c578e0cd46d1784dfee" Dec 07 16:21:20 crc kubenswrapper[4716]: I1207 16:21:20.828366 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2f2180238a1c794b6d0ab4bc199979f41c72940340f9c578e0cd46d1784dfee"} err="failed to get container status \"d2f2180238a1c794b6d0ab4bc199979f41c72940340f9c578e0cd46d1784dfee\": rpc error: code = NotFound desc = could not find container \"d2f2180238a1c794b6d0ab4bc199979f41c72940340f9c578e0cd46d1784dfee\": container with ID starting with d2f2180238a1c794b6d0ab4bc199979f41c72940340f9c578e0cd46d1784dfee not found: ID does not exist" Dec 07 16:21:20 crc kubenswrapper[4716]: I1207 16:21:20.828388 4716 scope.go:117] "RemoveContainer" containerID="667a06ce95001f73ad3dba6c78a5a8cb2f34ff586efb54823be406a44a5a19ce" Dec 07 16:21:20 crc kubenswrapper[4716]: E1207 16:21:20.837689 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"667a06ce95001f73ad3dba6c78a5a8cb2f34ff586efb54823be406a44a5a19ce\": container with ID starting with 667a06ce95001f73ad3dba6c78a5a8cb2f34ff586efb54823be406a44a5a19ce not found: ID does not exist" containerID="667a06ce95001f73ad3dba6c78a5a8cb2f34ff586efb54823be406a44a5a19ce" Dec 07 16:21:20 crc kubenswrapper[4716]: I1207 16:21:20.837748 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"667a06ce95001f73ad3dba6c78a5a8cb2f34ff586efb54823be406a44a5a19ce"} err="failed to get container status \"667a06ce95001f73ad3dba6c78a5a8cb2f34ff586efb54823be406a44a5a19ce\": rpc error: code = NotFound desc = could not find container \"667a06ce95001f73ad3dba6c78a5a8cb2f34ff586efb54823be406a44a5a19ce\": container with ID starting with 667a06ce95001f73ad3dba6c78a5a8cb2f34ff586efb54823be406a44a5a19ce not found: ID does not exist" Dec 07 16:21:20 crc kubenswrapper[4716]: I1207 16:21:20.837781 4716 scope.go:117] "RemoveContainer" containerID="d2f2180238a1c794b6d0ab4bc199979f41c72940340f9c578e0cd46d1784dfee" Dec 07 16:21:20 crc kubenswrapper[4716]: I1207 16:21:20.837919 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 07 16:21:20 crc kubenswrapper[4716]: I1207 16:21:20.838151 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2f2180238a1c794b6d0ab4bc199979f41c72940340f9c578e0cd46d1784dfee"} err="failed to get container status \"d2f2180238a1c794b6d0ab4bc199979f41c72940340f9c578e0cd46d1784dfee\": rpc error: code = NotFound desc = could not find container \"d2f2180238a1c794b6d0ab4bc199979f41c72940340f9c578e0cd46d1784dfee\": container with ID starting with d2f2180238a1c794b6d0ab4bc199979f41c72940340f9c578e0cd46d1784dfee not found: ID does not exist" Dec 07 16:21:20 crc kubenswrapper[4716]: I1207 16:21:20.838175 4716 scope.go:117] "RemoveContainer" containerID="667a06ce95001f73ad3dba6c78a5a8cb2f34ff586efb54823be406a44a5a19ce" Dec 07 16:21:20 crc kubenswrapper[4716]: I1207 16:21:20.839602 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"667a06ce95001f73ad3dba6c78a5a8cb2f34ff586efb54823be406a44a5a19ce"} err="failed to get container status \"667a06ce95001f73ad3dba6c78a5a8cb2f34ff586efb54823be406a44a5a19ce\": rpc error: code = NotFound desc = could not find container \"667a06ce95001f73ad3dba6c78a5a8cb2f34ff586efb54823be406a44a5a19ce\": container with ID starting with 667a06ce95001f73ad3dba6c78a5a8cb2f34ff586efb54823be406a44a5a19ce not found: ID does not exist" Dec 07 16:21:20 crc kubenswrapper[4716]: I1207 16:21:20.844782 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 07 16:21:20 crc kubenswrapper[4716]: E1207 16:21:20.845441 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="572389a3-395c-4626-b731-fda219d74c96" containerName="nova-metadata-metadata" Dec 07 16:21:20 crc kubenswrapper[4716]: I1207 16:21:20.845461 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="572389a3-395c-4626-b731-fda219d74c96" containerName="nova-metadata-metadata" Dec 07 16:21:20 crc kubenswrapper[4716]: E1207 16:21:20.845479 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="572389a3-395c-4626-b731-fda219d74c96" containerName="nova-metadata-log" Dec 07 16:21:20 crc kubenswrapper[4716]: I1207 16:21:20.845486 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="572389a3-395c-4626-b731-fda219d74c96" containerName="nova-metadata-log" Dec 07 16:21:20 crc kubenswrapper[4716]: I1207 16:21:20.845683 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="572389a3-395c-4626-b731-fda219d74c96" containerName="nova-metadata-log" Dec 07 16:21:20 crc kubenswrapper[4716]: I1207 16:21:20.845709 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="572389a3-395c-4626-b731-fda219d74c96" containerName="nova-metadata-metadata" Dec 07 16:21:20 crc kubenswrapper[4716]: I1207 16:21:20.846799 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 07 16:21:20 crc kubenswrapper[4716]: I1207 16:21:20.851836 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 07 16:21:20 crc kubenswrapper[4716]: I1207 16:21:20.852115 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 07 16:21:20 crc kubenswrapper[4716]: I1207 16:21:20.854108 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 07 16:21:20 crc kubenswrapper[4716]: I1207 16:21:20.996405 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-494gn\" (UniqueName: \"kubernetes.io/projected/a1ae3465-8479-4587-a052-facebe0676d8-kube-api-access-494gn\") pod \"nova-metadata-0\" (UID: \"a1ae3465-8479-4587-a052-facebe0676d8\") " pod="openstack/nova-metadata-0" Dec 07 16:21:20 crc kubenswrapper[4716]: I1207 16:21:20.996465 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1ae3465-8479-4587-a052-facebe0676d8-logs\") pod \"nova-metadata-0\" (UID: \"a1ae3465-8479-4587-a052-facebe0676d8\") " pod="openstack/nova-metadata-0" Dec 07 16:21:20 crc kubenswrapper[4716]: I1207 16:21:20.996487 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1ae3465-8479-4587-a052-facebe0676d8-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a1ae3465-8479-4587-a052-facebe0676d8\") " pod="openstack/nova-metadata-0" Dec 07 16:21:20 crc kubenswrapper[4716]: I1207 16:21:20.996552 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1ae3465-8479-4587-a052-facebe0676d8-config-data\") pod \"nova-metadata-0\" (UID: \"a1ae3465-8479-4587-a052-facebe0676d8\") " pod="openstack/nova-metadata-0" Dec 07 16:21:20 crc kubenswrapper[4716]: I1207 16:21:20.996609 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1ae3465-8479-4587-a052-facebe0676d8-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"a1ae3465-8479-4587-a052-facebe0676d8\") " pod="openstack/nova-metadata-0" Dec 07 16:21:21 crc kubenswrapper[4716]: I1207 16:21:21.098657 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-494gn\" (UniqueName: \"kubernetes.io/projected/a1ae3465-8479-4587-a052-facebe0676d8-kube-api-access-494gn\") pod \"nova-metadata-0\" (UID: \"a1ae3465-8479-4587-a052-facebe0676d8\") " pod="openstack/nova-metadata-0" Dec 07 16:21:21 crc kubenswrapper[4716]: I1207 16:21:21.098719 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1ae3465-8479-4587-a052-facebe0676d8-logs\") pod \"nova-metadata-0\" (UID: \"a1ae3465-8479-4587-a052-facebe0676d8\") " pod="openstack/nova-metadata-0" Dec 07 16:21:21 crc kubenswrapper[4716]: I1207 16:21:21.098735 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1ae3465-8479-4587-a052-facebe0676d8-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a1ae3465-8479-4587-a052-facebe0676d8\") " pod="openstack/nova-metadata-0" Dec 07 16:21:21 crc kubenswrapper[4716]: I1207 16:21:21.098765 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1ae3465-8479-4587-a052-facebe0676d8-config-data\") pod \"nova-metadata-0\" (UID: \"a1ae3465-8479-4587-a052-facebe0676d8\") " pod="openstack/nova-metadata-0" Dec 07 16:21:21 crc kubenswrapper[4716]: I1207 16:21:21.098804 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1ae3465-8479-4587-a052-facebe0676d8-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"a1ae3465-8479-4587-a052-facebe0676d8\") " pod="openstack/nova-metadata-0" Dec 07 16:21:21 crc kubenswrapper[4716]: I1207 16:21:21.099504 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1ae3465-8479-4587-a052-facebe0676d8-logs\") pod \"nova-metadata-0\" (UID: \"a1ae3465-8479-4587-a052-facebe0676d8\") " pod="openstack/nova-metadata-0" Dec 07 16:21:21 crc kubenswrapper[4716]: I1207 16:21:21.103186 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1ae3465-8479-4587-a052-facebe0676d8-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"a1ae3465-8479-4587-a052-facebe0676d8\") " pod="openstack/nova-metadata-0" Dec 07 16:21:21 crc kubenswrapper[4716]: I1207 16:21:21.103286 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1ae3465-8479-4587-a052-facebe0676d8-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a1ae3465-8479-4587-a052-facebe0676d8\") " pod="openstack/nova-metadata-0" Dec 07 16:21:21 crc kubenswrapper[4716]: I1207 16:21:21.103304 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1ae3465-8479-4587-a052-facebe0676d8-config-data\") pod \"nova-metadata-0\" (UID: \"a1ae3465-8479-4587-a052-facebe0676d8\") " pod="openstack/nova-metadata-0" Dec 07 16:21:21 crc kubenswrapper[4716]: I1207 16:21:21.124581 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-494gn\" (UniqueName: \"kubernetes.io/projected/a1ae3465-8479-4587-a052-facebe0676d8-kube-api-access-494gn\") pod \"nova-metadata-0\" (UID: \"a1ae3465-8479-4587-a052-facebe0676d8\") " pod="openstack/nova-metadata-0" Dec 07 16:21:21 crc kubenswrapper[4716]: I1207 16:21:21.178491 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 07 16:21:21 crc kubenswrapper[4716]: I1207 16:21:21.672721 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="572389a3-395c-4626-b731-fda219d74c96" path="/var/lib/kubelet/pods/572389a3-395c-4626-b731-fda219d74c96/volumes" Dec 07 16:21:21 crc kubenswrapper[4716]: W1207 16:21:21.693801 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda1ae3465_8479_4587_a052_facebe0676d8.slice/crio-cc69e67433b5d9da81193226cb4c15b1879393fef98c7ec940758636c3c7d7b3 WatchSource:0}: Error finding container cc69e67433b5d9da81193226cb4c15b1879393fef98c7ec940758636c3c7d7b3: Status 404 returned error can't find the container with id cc69e67433b5d9da81193226cb4c15b1879393fef98c7ec940758636c3c7d7b3 Dec 07 16:21:21 crc kubenswrapper[4716]: I1207 16:21:21.717931 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 07 16:21:21 crc kubenswrapper[4716]: I1207 16:21:21.763149 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a1ae3465-8479-4587-a052-facebe0676d8","Type":"ContainerStarted","Data":"cc69e67433b5d9da81193226cb4c15b1879393fef98c7ec940758636c3c7d7b3"} Dec 07 16:21:22 crc kubenswrapper[4716]: I1207 16:21:22.781817 4716 generic.go:334] "Generic (PLEG): container finished" podID="367f54c1-6d21-41fc-8c08-d6a7e9e637cf" containerID="f918233e5ff893608dff230031ceae9e1dfe9739e05fd2ab94b483e2a33df54f" exitCode=0 Dec 07 16:21:22 crc kubenswrapper[4716]: I1207 16:21:22.782294 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-q9fx7" event={"ID":"367f54c1-6d21-41fc-8c08-d6a7e9e637cf","Type":"ContainerDied","Data":"f918233e5ff893608dff230031ceae9e1dfe9739e05fd2ab94b483e2a33df54f"} Dec 07 16:21:22 crc kubenswrapper[4716]: I1207 16:21:22.786315 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a1ae3465-8479-4587-a052-facebe0676d8","Type":"ContainerStarted","Data":"1a96bbd12163053b45ed73d921c0c03dc39f791c81cf56afc208090b9173227e"} Dec 07 16:21:22 crc kubenswrapper[4716]: I1207 16:21:22.786367 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a1ae3465-8479-4587-a052-facebe0676d8","Type":"ContainerStarted","Data":"3446f69c3a8acc400eba0b44ddc9d1e5e377d963615145383855417a5167a0fe"} Dec 07 16:21:22 crc kubenswrapper[4716]: I1207 16:21:22.842000 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.841970167 podStartE2EDuration="2.841970167s" podCreationTimestamp="2025-12-07 16:21:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:21:22.827872887 +0000 UTC m=+1145.518157799" watchObservedRunningTime="2025-12-07 16:21:22.841970167 +0000 UTC m=+1145.532255119" Dec 07 16:21:23 crc kubenswrapper[4716]: I1207 16:21:23.799115 4716 generic.go:334] "Generic (PLEG): container finished" podID="9a950343-f5b9-4a42-857b-7de8ef106453" containerID="6368cc2ab6c45969b8edd9d2e09cdeb711278ef76446cd516bd195c0d856f570" exitCode=0 Dec 07 16:21:23 crc kubenswrapper[4716]: I1207 16:21:23.799205 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-x9skx" event={"ID":"9a950343-f5b9-4a42-857b-7de8ef106453","Type":"ContainerDied","Data":"6368cc2ab6c45969b8edd9d2e09cdeb711278ef76446cd516bd195c0d856f570"} Dec 07 16:21:23 crc kubenswrapper[4716]: I1207 16:21:23.923897 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 07 16:21:24 crc kubenswrapper[4716]: I1207 16:21:24.216389 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-q9fx7" Dec 07 16:21:24 crc kubenswrapper[4716]: I1207 16:21:24.365669 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/367f54c1-6d21-41fc-8c08-d6a7e9e637cf-config-data\") pod \"367f54c1-6d21-41fc-8c08-d6a7e9e637cf\" (UID: \"367f54c1-6d21-41fc-8c08-d6a7e9e637cf\") " Dec 07 16:21:24 crc kubenswrapper[4716]: I1207 16:21:24.366061 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/367f54c1-6d21-41fc-8c08-d6a7e9e637cf-combined-ca-bundle\") pod \"367f54c1-6d21-41fc-8c08-d6a7e9e637cf\" (UID: \"367f54c1-6d21-41fc-8c08-d6a7e9e637cf\") " Dec 07 16:21:24 crc kubenswrapper[4716]: I1207 16:21:24.366209 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/367f54c1-6d21-41fc-8c08-d6a7e9e637cf-scripts\") pod \"367f54c1-6d21-41fc-8c08-d6a7e9e637cf\" (UID: \"367f54c1-6d21-41fc-8c08-d6a7e9e637cf\") " Dec 07 16:21:24 crc kubenswrapper[4716]: I1207 16:21:24.366342 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-66nfk\" (UniqueName: \"kubernetes.io/projected/367f54c1-6d21-41fc-8c08-d6a7e9e637cf-kube-api-access-66nfk\") pod \"367f54c1-6d21-41fc-8c08-d6a7e9e637cf\" (UID: \"367f54c1-6d21-41fc-8c08-d6a7e9e637cf\") " Dec 07 16:21:24 crc kubenswrapper[4716]: I1207 16:21:24.371459 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/367f54c1-6d21-41fc-8c08-d6a7e9e637cf-kube-api-access-66nfk" (OuterVolumeSpecName: "kube-api-access-66nfk") pod "367f54c1-6d21-41fc-8c08-d6a7e9e637cf" (UID: "367f54c1-6d21-41fc-8c08-d6a7e9e637cf"). InnerVolumeSpecName "kube-api-access-66nfk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:21:24 crc kubenswrapper[4716]: I1207 16:21:24.379902 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/367f54c1-6d21-41fc-8c08-d6a7e9e637cf-scripts" (OuterVolumeSpecName: "scripts") pod "367f54c1-6d21-41fc-8c08-d6a7e9e637cf" (UID: "367f54c1-6d21-41fc-8c08-d6a7e9e637cf"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:21:24 crc kubenswrapper[4716]: I1207 16:21:24.394219 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/367f54c1-6d21-41fc-8c08-d6a7e9e637cf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "367f54c1-6d21-41fc-8c08-d6a7e9e637cf" (UID: "367f54c1-6d21-41fc-8c08-d6a7e9e637cf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:21:24 crc kubenswrapper[4716]: I1207 16:21:24.401700 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/367f54c1-6d21-41fc-8c08-d6a7e9e637cf-config-data" (OuterVolumeSpecName: "config-data") pod "367f54c1-6d21-41fc-8c08-d6a7e9e637cf" (UID: "367f54c1-6d21-41fc-8c08-d6a7e9e637cf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:21:24 crc kubenswrapper[4716]: I1207 16:21:24.468708 4716 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/367f54c1-6d21-41fc-8c08-d6a7e9e637cf-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 16:21:24 crc kubenswrapper[4716]: I1207 16:21:24.468744 4716 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/367f54c1-6d21-41fc-8c08-d6a7e9e637cf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 16:21:24 crc kubenswrapper[4716]: I1207 16:21:24.468756 4716 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/367f54c1-6d21-41fc-8c08-d6a7e9e637cf-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 16:21:24 crc kubenswrapper[4716]: I1207 16:21:24.468765 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-66nfk\" (UniqueName: \"kubernetes.io/projected/367f54c1-6d21-41fc-8c08-d6a7e9e637cf-kube-api-access-66nfk\") on node \"crc\" DevicePath \"\"" Dec 07 16:21:24 crc kubenswrapper[4716]: I1207 16:21:24.819293 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-q9fx7" Dec 07 16:21:24 crc kubenswrapper[4716]: I1207 16:21:24.819319 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-q9fx7" event={"ID":"367f54c1-6d21-41fc-8c08-d6a7e9e637cf","Type":"ContainerDied","Data":"1b53dde7bac526b271c3852b6a52649fb6944a1753214a8dc2a3388051b27610"} Dec 07 16:21:24 crc kubenswrapper[4716]: I1207 16:21:24.819369 4716 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1b53dde7bac526b271c3852b6a52649fb6944a1753214a8dc2a3388051b27610" Dec 07 16:21:25 crc kubenswrapper[4716]: I1207 16:21:25.001363 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 07 16:21:25 crc kubenswrapper[4716]: I1207 16:21:25.001671 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="77330671-85cf-40c8-81b0-7a7fe1f480a6" containerName="nova-api-log" containerID="cri-o://9b05dea586ef734718ea6ec7ea96d56fa62614c09c073e3925b8fdaca52a7b63" gracePeriod=30 Dec 07 16:21:25 crc kubenswrapper[4716]: I1207 16:21:25.002261 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="77330671-85cf-40c8-81b0-7a7fe1f480a6" containerName="nova-api-api" containerID="cri-o://00c23e67a27f2b5f91cde6ffa7187c805202a9c4f2f7ad2f5c22239560158ca7" gracePeriod=30 Dec 07 16:21:25 crc kubenswrapper[4716]: I1207 16:21:25.010330 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 07 16:21:25 crc kubenswrapper[4716]: I1207 16:21:25.010559 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="c9ffe38d-2cbd-40fa-ae1d-12b42e721bb4" containerName="nova-scheduler-scheduler" containerID="cri-o://793fc688ad69e9d6f7b90a700e5b82626010662968c1cb60c6a290fd528be52f" gracePeriod=30 Dec 07 16:21:25 crc kubenswrapper[4716]: I1207 16:21:25.081318 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 07 16:21:25 crc kubenswrapper[4716]: I1207 16:21:25.081621 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="a1ae3465-8479-4587-a052-facebe0676d8" containerName="nova-metadata-log" containerID="cri-o://3446f69c3a8acc400eba0b44ddc9d1e5e377d963615145383855417a5167a0fe" gracePeriod=30 Dec 07 16:21:25 crc kubenswrapper[4716]: I1207 16:21:25.082207 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="a1ae3465-8479-4587-a052-facebe0676d8" containerName="nova-metadata-metadata" containerID="cri-o://1a96bbd12163053b45ed73d921c0c03dc39f791c81cf56afc208090b9173227e" gracePeriod=30 Dec 07 16:21:25 crc kubenswrapper[4716]: I1207 16:21:25.141264 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-757b4f8459-v4tqc" Dec 07 16:21:25 crc kubenswrapper[4716]: I1207 16:21:25.202653 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-vb6n9"] Dec 07 16:21:25 crc kubenswrapper[4716]: I1207 16:21:25.202936 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c9776ccc5-vb6n9" podUID="f9195c58-0e28-495d-9b97-cce7c292b7ac" containerName="dnsmasq-dns" containerID="cri-o://d55ffbf42228f71084be488104e19559e1a1a664a13930078576be29b62e11c6" gracePeriod=10 Dec 07 16:21:25 crc kubenswrapper[4716]: I1207 16:21:25.519851 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-x9skx" Dec 07 16:21:25 crc kubenswrapper[4716]: I1207 16:21:25.698558 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a950343-f5b9-4a42-857b-7de8ef106453-combined-ca-bundle\") pod \"9a950343-f5b9-4a42-857b-7de8ef106453\" (UID: \"9a950343-f5b9-4a42-857b-7de8ef106453\") " Dec 07 16:21:25 crc kubenswrapper[4716]: I1207 16:21:25.698620 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a950343-f5b9-4a42-857b-7de8ef106453-scripts\") pod \"9a950343-f5b9-4a42-857b-7de8ef106453\" (UID: \"9a950343-f5b9-4a42-857b-7de8ef106453\") " Dec 07 16:21:25 crc kubenswrapper[4716]: I1207 16:21:25.698741 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a950343-f5b9-4a42-857b-7de8ef106453-config-data\") pod \"9a950343-f5b9-4a42-857b-7de8ef106453\" (UID: \"9a950343-f5b9-4a42-857b-7de8ef106453\") " Dec 07 16:21:25 crc kubenswrapper[4716]: I1207 16:21:25.698838 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j8dtp\" (UniqueName: \"kubernetes.io/projected/9a950343-f5b9-4a42-857b-7de8ef106453-kube-api-access-j8dtp\") pod \"9a950343-f5b9-4a42-857b-7de8ef106453\" (UID: \"9a950343-f5b9-4a42-857b-7de8ef106453\") " Dec 07 16:21:25 crc kubenswrapper[4716]: I1207 16:21:25.702313 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a950343-f5b9-4a42-857b-7de8ef106453-scripts" (OuterVolumeSpecName: "scripts") pod "9a950343-f5b9-4a42-857b-7de8ef106453" (UID: "9a950343-f5b9-4a42-857b-7de8ef106453"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:21:25 crc kubenswrapper[4716]: I1207 16:21:25.703997 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a950343-f5b9-4a42-857b-7de8ef106453-kube-api-access-j8dtp" (OuterVolumeSpecName: "kube-api-access-j8dtp") pod "9a950343-f5b9-4a42-857b-7de8ef106453" (UID: "9a950343-f5b9-4a42-857b-7de8ef106453"). InnerVolumeSpecName "kube-api-access-j8dtp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:21:25 crc kubenswrapper[4716]: I1207 16:21:25.728681 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a950343-f5b9-4a42-857b-7de8ef106453-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9a950343-f5b9-4a42-857b-7de8ef106453" (UID: "9a950343-f5b9-4a42-857b-7de8ef106453"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:21:25 crc kubenswrapper[4716]: I1207 16:21:25.729279 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a950343-f5b9-4a42-857b-7de8ef106453-config-data" (OuterVolumeSpecName: "config-data") pod "9a950343-f5b9-4a42-857b-7de8ef106453" (UID: "9a950343-f5b9-4a42-857b-7de8ef106453"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:21:25 crc kubenswrapper[4716]: I1207 16:21:25.805965 4716 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a950343-f5b9-4a42-857b-7de8ef106453-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 16:21:25 crc kubenswrapper[4716]: I1207 16:21:25.806028 4716 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a950343-f5b9-4a42-857b-7de8ef106453-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 16:21:25 crc kubenswrapper[4716]: I1207 16:21:25.806043 4716 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a950343-f5b9-4a42-857b-7de8ef106453-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 16:21:25 crc kubenswrapper[4716]: I1207 16:21:25.806057 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j8dtp\" (UniqueName: \"kubernetes.io/projected/9a950343-f5b9-4a42-857b-7de8ef106453-kube-api-access-j8dtp\") on node \"crc\" DevicePath \"\"" Dec 07 16:21:25 crc kubenswrapper[4716]: I1207 16:21:25.833364 4716 generic.go:334] "Generic (PLEG): container finished" podID="a1ae3465-8479-4587-a052-facebe0676d8" containerID="1a96bbd12163053b45ed73d921c0c03dc39f791c81cf56afc208090b9173227e" exitCode=0 Dec 07 16:21:25 crc kubenswrapper[4716]: I1207 16:21:25.833407 4716 generic.go:334] "Generic (PLEG): container finished" podID="a1ae3465-8479-4587-a052-facebe0676d8" containerID="3446f69c3a8acc400eba0b44ddc9d1e5e377d963615145383855417a5167a0fe" exitCode=143 Dec 07 16:21:25 crc kubenswrapper[4716]: I1207 16:21:25.833430 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a1ae3465-8479-4587-a052-facebe0676d8","Type":"ContainerDied","Data":"1a96bbd12163053b45ed73d921c0c03dc39f791c81cf56afc208090b9173227e"} Dec 07 16:21:25 crc kubenswrapper[4716]: I1207 16:21:25.833510 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a1ae3465-8479-4587-a052-facebe0676d8","Type":"ContainerDied","Data":"3446f69c3a8acc400eba0b44ddc9d1e5e377d963615145383855417a5167a0fe"} Dec 07 16:21:25 crc kubenswrapper[4716]: I1207 16:21:25.853541 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-x9skx" event={"ID":"9a950343-f5b9-4a42-857b-7de8ef106453","Type":"ContainerDied","Data":"6613c95297620627e30d0fdb5ed6d868bf89ba99de35c7f670d1757ca345ac97"} Dec 07 16:21:25 crc kubenswrapper[4716]: I1207 16:21:25.853584 4716 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6613c95297620627e30d0fdb5ed6d868bf89ba99de35c7f670d1757ca345ac97" Dec 07 16:21:25 crc kubenswrapper[4716]: I1207 16:21:25.853666 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-x9skx" Dec 07 16:21:25 crc kubenswrapper[4716]: I1207 16:21:25.860136 4716 generic.go:334] "Generic (PLEG): container finished" podID="77330671-85cf-40c8-81b0-7a7fe1f480a6" containerID="00c23e67a27f2b5f91cde6ffa7187c805202a9c4f2f7ad2f5c22239560158ca7" exitCode=0 Dec 07 16:21:25 crc kubenswrapper[4716]: I1207 16:21:25.860175 4716 generic.go:334] "Generic (PLEG): container finished" podID="77330671-85cf-40c8-81b0-7a7fe1f480a6" containerID="9b05dea586ef734718ea6ec7ea96d56fa62614c09c073e3925b8fdaca52a7b63" exitCode=143 Dec 07 16:21:25 crc kubenswrapper[4716]: I1207 16:21:25.860208 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"77330671-85cf-40c8-81b0-7a7fe1f480a6","Type":"ContainerDied","Data":"00c23e67a27f2b5f91cde6ffa7187c805202a9c4f2f7ad2f5c22239560158ca7"} Dec 07 16:21:25 crc kubenswrapper[4716]: I1207 16:21:25.860251 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"77330671-85cf-40c8-81b0-7a7fe1f480a6","Type":"ContainerDied","Data":"9b05dea586ef734718ea6ec7ea96d56fa62614c09c073e3925b8fdaca52a7b63"} Dec 07 16:21:25 crc kubenswrapper[4716]: I1207 16:21:25.870852 4716 generic.go:334] "Generic (PLEG): container finished" podID="f9195c58-0e28-495d-9b97-cce7c292b7ac" containerID="d55ffbf42228f71084be488104e19559e1a1a664a13930078576be29b62e11c6" exitCode=0 Dec 07 16:21:25 crc kubenswrapper[4716]: I1207 16:21:25.870899 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-vb6n9" event={"ID":"f9195c58-0e28-495d-9b97-cce7c292b7ac","Type":"ContainerDied","Data":"d55ffbf42228f71084be488104e19559e1a1a664a13930078576be29b62e11c6"} Dec 07 16:21:25 crc kubenswrapper[4716]: I1207 16:21:25.908149 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 07 16:21:25 crc kubenswrapper[4716]: E1207 16:21:25.908904 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="367f54c1-6d21-41fc-8c08-d6a7e9e637cf" containerName="nova-manage" Dec 07 16:21:25 crc kubenswrapper[4716]: I1207 16:21:25.908919 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="367f54c1-6d21-41fc-8c08-d6a7e9e637cf" containerName="nova-manage" Dec 07 16:21:25 crc kubenswrapper[4716]: E1207 16:21:25.908933 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a950343-f5b9-4a42-857b-7de8ef106453" containerName="nova-cell1-conductor-db-sync" Dec 07 16:21:25 crc kubenswrapper[4716]: I1207 16:21:25.908940 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a950343-f5b9-4a42-857b-7de8ef106453" containerName="nova-cell1-conductor-db-sync" Dec 07 16:21:25 crc kubenswrapper[4716]: I1207 16:21:25.909212 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a950343-f5b9-4a42-857b-7de8ef106453" containerName="nova-cell1-conductor-db-sync" Dec 07 16:21:25 crc kubenswrapper[4716]: I1207 16:21:25.909235 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="367f54c1-6d21-41fc-8c08-d6a7e9e637cf" containerName="nova-manage" Dec 07 16:21:25 crc kubenswrapper[4716]: I1207 16:21:25.910023 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 07 16:21:25 crc kubenswrapper[4716]: I1207 16:21:25.912535 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 07 16:21:25 crc kubenswrapper[4716]: I1207 16:21:25.938103 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.021145 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-vb6n9" Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.024892 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pdzgz\" (UniqueName: \"kubernetes.io/projected/098f28a4-4852-4c3a-ada9-7b8b9272c3ae-kube-api-access-pdzgz\") pod \"nova-cell1-conductor-0\" (UID: \"098f28a4-4852-4c3a-ada9-7b8b9272c3ae\") " pod="openstack/nova-cell1-conductor-0" Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.024932 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/098f28a4-4852-4c3a-ada9-7b8b9272c3ae-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"098f28a4-4852-4c3a-ada9-7b8b9272c3ae\") " pod="openstack/nova-cell1-conductor-0" Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.024985 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/098f28a4-4852-4c3a-ada9-7b8b9272c3ae-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"098f28a4-4852-4c3a-ada9-7b8b9272c3ae\") " pod="openstack/nova-cell1-conductor-0" Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.126679 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9195c58-0e28-495d-9b97-cce7c292b7ac-config\") pod \"f9195c58-0e28-495d-9b97-cce7c292b7ac\" (UID: \"f9195c58-0e28-495d-9b97-cce7c292b7ac\") " Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.126746 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f9195c58-0e28-495d-9b97-cce7c292b7ac-dns-svc\") pod \"f9195c58-0e28-495d-9b97-cce7c292b7ac\" (UID: \"f9195c58-0e28-495d-9b97-cce7c292b7ac\") " Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.126777 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f9195c58-0e28-495d-9b97-cce7c292b7ac-ovsdbserver-sb\") pod \"f9195c58-0e28-495d-9b97-cce7c292b7ac\" (UID: \"f9195c58-0e28-495d-9b97-cce7c292b7ac\") " Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.126811 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k58zw\" (UniqueName: \"kubernetes.io/projected/f9195c58-0e28-495d-9b97-cce7c292b7ac-kube-api-access-k58zw\") pod \"f9195c58-0e28-495d-9b97-cce7c292b7ac\" (UID: \"f9195c58-0e28-495d-9b97-cce7c292b7ac\") " Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.126833 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f9195c58-0e28-495d-9b97-cce7c292b7ac-ovsdbserver-nb\") pod \"f9195c58-0e28-495d-9b97-cce7c292b7ac\" (UID: \"f9195c58-0e28-495d-9b97-cce7c292b7ac\") " Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.126972 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f9195c58-0e28-495d-9b97-cce7c292b7ac-dns-swift-storage-0\") pod \"f9195c58-0e28-495d-9b97-cce7c292b7ac\" (UID: \"f9195c58-0e28-495d-9b97-cce7c292b7ac\") " Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.127287 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pdzgz\" (UniqueName: \"kubernetes.io/projected/098f28a4-4852-4c3a-ada9-7b8b9272c3ae-kube-api-access-pdzgz\") pod \"nova-cell1-conductor-0\" (UID: \"098f28a4-4852-4c3a-ada9-7b8b9272c3ae\") " pod="openstack/nova-cell1-conductor-0" Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.127323 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/098f28a4-4852-4c3a-ada9-7b8b9272c3ae-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"098f28a4-4852-4c3a-ada9-7b8b9272c3ae\") " pod="openstack/nova-cell1-conductor-0" Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.127376 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/098f28a4-4852-4c3a-ada9-7b8b9272c3ae-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"098f28a4-4852-4c3a-ada9-7b8b9272c3ae\") " pod="openstack/nova-cell1-conductor-0" Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.132817 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9195c58-0e28-495d-9b97-cce7c292b7ac-kube-api-access-k58zw" (OuterVolumeSpecName: "kube-api-access-k58zw") pod "f9195c58-0e28-495d-9b97-cce7c292b7ac" (UID: "f9195c58-0e28-495d-9b97-cce7c292b7ac"). InnerVolumeSpecName "kube-api-access-k58zw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.133312 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/098f28a4-4852-4c3a-ada9-7b8b9272c3ae-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"098f28a4-4852-4c3a-ada9-7b8b9272c3ae\") " pod="openstack/nova-cell1-conductor-0" Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.136866 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/098f28a4-4852-4c3a-ada9-7b8b9272c3ae-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"098f28a4-4852-4c3a-ada9-7b8b9272c3ae\") " pod="openstack/nova-cell1-conductor-0" Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.147752 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pdzgz\" (UniqueName: \"kubernetes.io/projected/098f28a4-4852-4c3a-ada9-7b8b9272c3ae-kube-api-access-pdzgz\") pod \"nova-cell1-conductor-0\" (UID: \"098f28a4-4852-4c3a-ada9-7b8b9272c3ae\") " pod="openstack/nova-cell1-conductor-0" Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.180503 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.180549 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.181935 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f9195c58-0e28-495d-9b97-cce7c292b7ac-config" (OuterVolumeSpecName: "config") pod "f9195c58-0e28-495d-9b97-cce7c292b7ac" (UID: "f9195c58-0e28-495d-9b97-cce7c292b7ac"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.188583 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f9195c58-0e28-495d-9b97-cce7c292b7ac-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f9195c58-0e28-495d-9b97-cce7c292b7ac" (UID: "f9195c58-0e28-495d-9b97-cce7c292b7ac"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.190053 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f9195c58-0e28-495d-9b97-cce7c292b7ac-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f9195c58-0e28-495d-9b97-cce7c292b7ac" (UID: "f9195c58-0e28-495d-9b97-cce7c292b7ac"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.208671 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f9195c58-0e28-495d-9b97-cce7c292b7ac-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f9195c58-0e28-495d-9b97-cce7c292b7ac" (UID: "f9195c58-0e28-495d-9b97-cce7c292b7ac"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.212484 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f9195c58-0e28-495d-9b97-cce7c292b7ac-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "f9195c58-0e28-495d-9b97-cce7c292b7ac" (UID: "f9195c58-0e28-495d-9b97-cce7c292b7ac"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.229391 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.229535 4716 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f9195c58-0e28-495d-9b97-cce7c292b7ac-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.229878 4716 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9195c58-0e28-495d-9b97-cce7c292b7ac-config\") on node \"crc\" DevicePath \"\"" Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.229890 4716 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f9195c58-0e28-495d-9b97-cce7c292b7ac-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.229900 4716 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f9195c58-0e28-495d-9b97-cce7c292b7ac-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.229909 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k58zw\" (UniqueName: \"kubernetes.io/projected/f9195c58-0e28-495d-9b97-cce7c292b7ac-kube-api-access-k58zw\") on node \"crc\" DevicePath \"\"" Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.229917 4716 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f9195c58-0e28-495d-9b97-cce7c292b7ac-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.327265 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.401695 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.432848 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1ae3465-8479-4587-a052-facebe0676d8-nova-metadata-tls-certs\") pod \"a1ae3465-8479-4587-a052-facebe0676d8\" (UID: \"a1ae3465-8479-4587-a052-facebe0676d8\") " Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.432911 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1ae3465-8479-4587-a052-facebe0676d8-config-data\") pod \"a1ae3465-8479-4587-a052-facebe0676d8\" (UID: \"a1ae3465-8479-4587-a052-facebe0676d8\") " Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.432992 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1ae3465-8479-4587-a052-facebe0676d8-logs\") pod \"a1ae3465-8479-4587-a052-facebe0676d8\" (UID: \"a1ae3465-8479-4587-a052-facebe0676d8\") " Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.433062 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-494gn\" (UniqueName: \"kubernetes.io/projected/a1ae3465-8479-4587-a052-facebe0676d8-kube-api-access-494gn\") pod \"a1ae3465-8479-4587-a052-facebe0676d8\" (UID: \"a1ae3465-8479-4587-a052-facebe0676d8\") " Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.433118 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1ae3465-8479-4587-a052-facebe0676d8-combined-ca-bundle\") pod \"a1ae3465-8479-4587-a052-facebe0676d8\" (UID: \"a1ae3465-8479-4587-a052-facebe0676d8\") " Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.433448 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a1ae3465-8479-4587-a052-facebe0676d8-logs" (OuterVolumeSpecName: "logs") pod "a1ae3465-8479-4587-a052-facebe0676d8" (UID: "a1ae3465-8479-4587-a052-facebe0676d8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.433658 4716 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1ae3465-8479-4587-a052-facebe0676d8-logs\") on node \"crc\" DevicePath \"\"" Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.438797 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1ae3465-8479-4587-a052-facebe0676d8-kube-api-access-494gn" (OuterVolumeSpecName: "kube-api-access-494gn") pod "a1ae3465-8479-4587-a052-facebe0676d8" (UID: "a1ae3465-8479-4587-a052-facebe0676d8"). InnerVolumeSpecName "kube-api-access-494gn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.458386 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1ae3465-8479-4587-a052-facebe0676d8-config-data" (OuterVolumeSpecName: "config-data") pod "a1ae3465-8479-4587-a052-facebe0676d8" (UID: "a1ae3465-8479-4587-a052-facebe0676d8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.492278 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1ae3465-8479-4587-a052-facebe0676d8-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "a1ae3465-8479-4587-a052-facebe0676d8" (UID: "a1ae3465-8479-4587-a052-facebe0676d8"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.494098 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1ae3465-8479-4587-a052-facebe0676d8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a1ae3465-8479-4587-a052-facebe0676d8" (UID: "a1ae3465-8479-4587-a052-facebe0676d8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.535141 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77330671-85cf-40c8-81b0-7a7fe1f480a6-config-data\") pod \"77330671-85cf-40c8-81b0-7a7fe1f480a6\" (UID: \"77330671-85cf-40c8-81b0-7a7fe1f480a6\") " Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.535234 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77330671-85cf-40c8-81b0-7a7fe1f480a6-combined-ca-bundle\") pod \"77330671-85cf-40c8-81b0-7a7fe1f480a6\" (UID: \"77330671-85cf-40c8-81b0-7a7fe1f480a6\") " Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.535267 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/77330671-85cf-40c8-81b0-7a7fe1f480a6-logs\") pod \"77330671-85cf-40c8-81b0-7a7fe1f480a6\" (UID: \"77330671-85cf-40c8-81b0-7a7fe1f480a6\") " Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.535393 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mc4tf\" (UniqueName: \"kubernetes.io/projected/77330671-85cf-40c8-81b0-7a7fe1f480a6-kube-api-access-mc4tf\") pod \"77330671-85cf-40c8-81b0-7a7fe1f480a6\" (UID: \"77330671-85cf-40c8-81b0-7a7fe1f480a6\") " Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.535701 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/77330671-85cf-40c8-81b0-7a7fe1f480a6-logs" (OuterVolumeSpecName: "logs") pod "77330671-85cf-40c8-81b0-7a7fe1f480a6" (UID: "77330671-85cf-40c8-81b0-7a7fe1f480a6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.536137 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-494gn\" (UniqueName: \"kubernetes.io/projected/a1ae3465-8479-4587-a052-facebe0676d8-kube-api-access-494gn\") on node \"crc\" DevicePath \"\"" Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.536165 4716 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1ae3465-8479-4587-a052-facebe0676d8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.536178 4716 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1ae3465-8479-4587-a052-facebe0676d8-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.536192 4716 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/77330671-85cf-40c8-81b0-7a7fe1f480a6-logs\") on node \"crc\" DevicePath \"\"" Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.536203 4716 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1ae3465-8479-4587-a052-facebe0676d8-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.539919 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77330671-85cf-40c8-81b0-7a7fe1f480a6-kube-api-access-mc4tf" (OuterVolumeSpecName: "kube-api-access-mc4tf") pod "77330671-85cf-40c8-81b0-7a7fe1f480a6" (UID: "77330671-85cf-40c8-81b0-7a7fe1f480a6"). InnerVolumeSpecName "kube-api-access-mc4tf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.565289 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77330671-85cf-40c8-81b0-7a7fe1f480a6-config-data" (OuterVolumeSpecName: "config-data") pod "77330671-85cf-40c8-81b0-7a7fe1f480a6" (UID: "77330671-85cf-40c8-81b0-7a7fe1f480a6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.573323 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77330671-85cf-40c8-81b0-7a7fe1f480a6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "77330671-85cf-40c8-81b0-7a7fe1f480a6" (UID: "77330671-85cf-40c8-81b0-7a7fe1f480a6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.637821 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mc4tf\" (UniqueName: \"kubernetes.io/projected/77330671-85cf-40c8-81b0-7a7fe1f480a6-kube-api-access-mc4tf\") on node \"crc\" DevicePath \"\"" Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.637847 4716 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77330671-85cf-40c8-81b0-7a7fe1f480a6-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.637858 4716 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77330671-85cf-40c8-81b0-7a7fe1f480a6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.721503 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.730345 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.752850 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bb8zt\" (UniqueName: \"kubernetes.io/projected/c9ffe38d-2cbd-40fa-ae1d-12b42e721bb4-kube-api-access-bb8zt\") pod \"c9ffe38d-2cbd-40fa-ae1d-12b42e721bb4\" (UID: \"c9ffe38d-2cbd-40fa-ae1d-12b42e721bb4\") " Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.752940 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9ffe38d-2cbd-40fa-ae1d-12b42e721bb4-config-data\") pod \"c9ffe38d-2cbd-40fa-ae1d-12b42e721bb4\" (UID: \"c9ffe38d-2cbd-40fa-ae1d-12b42e721bb4\") " Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.753144 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9ffe38d-2cbd-40fa-ae1d-12b42e721bb4-combined-ca-bundle\") pod \"c9ffe38d-2cbd-40fa-ae1d-12b42e721bb4\" (UID: \"c9ffe38d-2cbd-40fa-ae1d-12b42e721bb4\") " Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.766548 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9ffe38d-2cbd-40fa-ae1d-12b42e721bb4-kube-api-access-bb8zt" (OuterVolumeSpecName: "kube-api-access-bb8zt") pod "c9ffe38d-2cbd-40fa-ae1d-12b42e721bb4" (UID: "c9ffe38d-2cbd-40fa-ae1d-12b42e721bb4"). InnerVolumeSpecName "kube-api-access-bb8zt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.804444 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9ffe38d-2cbd-40fa-ae1d-12b42e721bb4-config-data" (OuterVolumeSpecName: "config-data") pod "c9ffe38d-2cbd-40fa-ae1d-12b42e721bb4" (UID: "c9ffe38d-2cbd-40fa-ae1d-12b42e721bb4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.836467 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9ffe38d-2cbd-40fa-ae1d-12b42e721bb4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c9ffe38d-2cbd-40fa-ae1d-12b42e721bb4" (UID: "c9ffe38d-2cbd-40fa-ae1d-12b42e721bb4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.855444 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bb8zt\" (UniqueName: \"kubernetes.io/projected/c9ffe38d-2cbd-40fa-ae1d-12b42e721bb4-kube-api-access-bb8zt\") on node \"crc\" DevicePath \"\"" Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.855480 4716 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9ffe38d-2cbd-40fa-ae1d-12b42e721bb4-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.855489 4716 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9ffe38d-2cbd-40fa-ae1d-12b42e721bb4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.900973 4716 generic.go:334] "Generic (PLEG): container finished" podID="c9ffe38d-2cbd-40fa-ae1d-12b42e721bb4" containerID="793fc688ad69e9d6f7b90a700e5b82626010662968c1cb60c6a290fd528be52f" exitCode=0 Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.901104 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.902419 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c9ffe38d-2cbd-40fa-ae1d-12b42e721bb4","Type":"ContainerDied","Data":"793fc688ad69e9d6f7b90a700e5b82626010662968c1cb60c6a290fd528be52f"} Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.902471 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c9ffe38d-2cbd-40fa-ae1d-12b42e721bb4","Type":"ContainerDied","Data":"7586c0dc54847acf81822090ab9fb6276beb12198e1da69556617403e0bdf093"} Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.902492 4716 scope.go:117] "RemoveContainer" containerID="793fc688ad69e9d6f7b90a700e5b82626010662968c1cb60c6a290fd528be52f" Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.909020 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"77330671-85cf-40c8-81b0-7a7fe1f480a6","Type":"ContainerDied","Data":"857f961b190f662b0adfc7243385fad57fb1d5aea733c82c637787b9daca6c51"} Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.909120 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.912877 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-vb6n9" event={"ID":"f9195c58-0e28-495d-9b97-cce7c292b7ac","Type":"ContainerDied","Data":"02f0c3164652010b3ae9c510cbde541b845559b46876f389528a76f1c0e93832"} Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.912965 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-vb6n9" Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.931888 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a1ae3465-8479-4587-a052-facebe0676d8","Type":"ContainerDied","Data":"cc69e67433b5d9da81193226cb4c15b1879393fef98c7ec940758636c3c7d7b3"} Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.932051 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.932760 4716 scope.go:117] "RemoveContainer" containerID="793fc688ad69e9d6f7b90a700e5b82626010662968c1cb60c6a290fd528be52f" Dec 07 16:21:26 crc kubenswrapper[4716]: E1207 16:21:26.933211 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"793fc688ad69e9d6f7b90a700e5b82626010662968c1cb60c6a290fd528be52f\": container with ID starting with 793fc688ad69e9d6f7b90a700e5b82626010662968c1cb60c6a290fd528be52f not found: ID does not exist" containerID="793fc688ad69e9d6f7b90a700e5b82626010662968c1cb60c6a290fd528be52f" Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.933255 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"793fc688ad69e9d6f7b90a700e5b82626010662968c1cb60c6a290fd528be52f"} err="failed to get container status \"793fc688ad69e9d6f7b90a700e5b82626010662968c1cb60c6a290fd528be52f\": rpc error: code = NotFound desc = could not find container \"793fc688ad69e9d6f7b90a700e5b82626010662968c1cb60c6a290fd528be52f\": container with ID starting with 793fc688ad69e9d6f7b90a700e5b82626010662968c1cb60c6a290fd528be52f not found: ID does not exist" Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.933286 4716 scope.go:117] "RemoveContainer" containerID="00c23e67a27f2b5f91cde6ffa7187c805202a9c4f2f7ad2f5c22239560158ca7" Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.936596 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"098f28a4-4852-4c3a-ada9-7b8b9272c3ae","Type":"ContainerStarted","Data":"9b27f81fd9d2a59d78c5b36d6f4cffbfc18e309b82d3659b547ed9b97a40f193"} Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.959721 4716 scope.go:117] "RemoveContainer" containerID="9b05dea586ef734718ea6ec7ea96d56fa62614c09c073e3925b8fdaca52a7b63" Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.974454 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 07 16:21:26 crc kubenswrapper[4716]: I1207 16:21:26.986027 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.006175 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 07 16:21:27 crc kubenswrapper[4716]: E1207 16:21:27.006651 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9ffe38d-2cbd-40fa-ae1d-12b42e721bb4" containerName="nova-scheduler-scheduler" Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.006674 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9ffe38d-2cbd-40fa-ae1d-12b42e721bb4" containerName="nova-scheduler-scheduler" Dec 07 16:21:27 crc kubenswrapper[4716]: E1207 16:21:27.006696 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9195c58-0e28-495d-9b97-cce7c292b7ac" containerName="dnsmasq-dns" Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.006706 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9195c58-0e28-495d-9b97-cce7c292b7ac" containerName="dnsmasq-dns" Dec 07 16:21:27 crc kubenswrapper[4716]: E1207 16:21:27.006727 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1ae3465-8479-4587-a052-facebe0676d8" containerName="nova-metadata-metadata" Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.006735 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1ae3465-8479-4587-a052-facebe0676d8" containerName="nova-metadata-metadata" Dec 07 16:21:27 crc kubenswrapper[4716]: E1207 16:21:27.006750 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9195c58-0e28-495d-9b97-cce7c292b7ac" containerName="init" Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.006758 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9195c58-0e28-495d-9b97-cce7c292b7ac" containerName="init" Dec 07 16:21:27 crc kubenswrapper[4716]: E1207 16:21:27.006771 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1ae3465-8479-4587-a052-facebe0676d8" containerName="nova-metadata-log" Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.006779 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1ae3465-8479-4587-a052-facebe0676d8" containerName="nova-metadata-log" Dec 07 16:21:27 crc kubenswrapper[4716]: E1207 16:21:27.006788 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77330671-85cf-40c8-81b0-7a7fe1f480a6" containerName="nova-api-api" Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.006797 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="77330671-85cf-40c8-81b0-7a7fe1f480a6" containerName="nova-api-api" Dec 07 16:21:27 crc kubenswrapper[4716]: E1207 16:21:27.006831 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77330671-85cf-40c8-81b0-7a7fe1f480a6" containerName="nova-api-log" Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.006840 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="77330671-85cf-40c8-81b0-7a7fe1f480a6" containerName="nova-api-log" Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.007151 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1ae3465-8479-4587-a052-facebe0676d8" containerName="nova-metadata-log" Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.007180 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9ffe38d-2cbd-40fa-ae1d-12b42e721bb4" containerName="nova-scheduler-scheduler" Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.007196 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1ae3465-8479-4587-a052-facebe0676d8" containerName="nova-metadata-metadata" Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.007214 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9195c58-0e28-495d-9b97-cce7c292b7ac" containerName="dnsmasq-dns" Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.007233 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="77330671-85cf-40c8-81b0-7a7fe1f480a6" containerName="nova-api-log" Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.007251 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="77330671-85cf-40c8-81b0-7a7fe1f480a6" containerName="nova-api-api" Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.008289 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.029151 4716 scope.go:117] "RemoveContainer" containerID="d55ffbf42228f71084be488104e19559e1a1a664a13930078576be29b62e11c6" Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.034478 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.041576 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.055945 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.073150 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.079731 4716 scope.go:117] "RemoveContainer" containerID="1d0c3f6589d32291dc7ce091aaf5272b16e51ae7368ca4e1dce1771c26b7b79c" Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.098225 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.106707 4716 scope.go:117] "RemoveContainer" containerID="1a96bbd12163053b45ed73d921c0c03dc39f791c81cf56afc208090b9173227e" Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.108091 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.119784 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.121146 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.122974 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.136704 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-vb6n9"] Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.144680 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-vb6n9"] Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.153298 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.155289 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.158486 4716 scope.go:117] "RemoveContainer" containerID="3446f69c3a8acc400eba0b44ddc9d1e5e377d963615145383855417a5167a0fe" Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.158979 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.159024 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.165304 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.170490 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/156ae83a-3fe2-4947-8ae5-5e665f19325a-logs\") pod \"nova-api-0\" (UID: \"156ae83a-3fe2-4947-8ae5-5e665f19325a\") " pod="openstack/nova-api-0" Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.170537 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/156ae83a-3fe2-4947-8ae5-5e665f19325a-config-data\") pod \"nova-api-0\" (UID: \"156ae83a-3fe2-4947-8ae5-5e665f19325a\") " pod="openstack/nova-api-0" Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.170572 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jplqw\" (UniqueName: \"kubernetes.io/projected/156ae83a-3fe2-4947-8ae5-5e665f19325a-kube-api-access-jplqw\") pod \"nova-api-0\" (UID: \"156ae83a-3fe2-4947-8ae5-5e665f19325a\") " pod="openstack/nova-api-0" Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.170599 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/156ae83a-3fe2-4947-8ae5-5e665f19325a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"156ae83a-3fe2-4947-8ae5-5e665f19325a\") " pod="openstack/nova-api-0" Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.177206 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.272270 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4jqpm\" (UniqueName: \"kubernetes.io/projected/fdcb1fee-b3e5-4ed1-b868-875220d969be-kube-api-access-4jqpm\") pod \"nova-metadata-0\" (UID: \"fdcb1fee-b3e5-4ed1-b868-875220d969be\") " pod="openstack/nova-metadata-0" Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.272335 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/156ae83a-3fe2-4947-8ae5-5e665f19325a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"156ae83a-3fe2-4947-8ae5-5e665f19325a\") " pod="openstack/nova-api-0" Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.272416 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed6ed6ec-40bd-4628-91c8-2410b68f7384-config-data\") pod \"nova-scheduler-0\" (UID: \"ed6ed6ec-40bd-4628-91c8-2410b68f7384\") " pod="openstack/nova-scheduler-0" Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.272483 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mlh98\" (UniqueName: \"kubernetes.io/projected/ed6ed6ec-40bd-4628-91c8-2410b68f7384-kube-api-access-mlh98\") pod \"nova-scheduler-0\" (UID: \"ed6ed6ec-40bd-4628-91c8-2410b68f7384\") " pod="openstack/nova-scheduler-0" Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.272528 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdcb1fee-b3e5-4ed1-b868-875220d969be-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"fdcb1fee-b3e5-4ed1-b868-875220d969be\") " pod="openstack/nova-metadata-0" Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.272563 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fdcb1fee-b3e5-4ed1-b868-875220d969be-logs\") pod \"nova-metadata-0\" (UID: \"fdcb1fee-b3e5-4ed1-b868-875220d969be\") " pod="openstack/nova-metadata-0" Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.272582 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed6ed6ec-40bd-4628-91c8-2410b68f7384-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ed6ed6ec-40bd-4628-91c8-2410b68f7384\") " pod="openstack/nova-scheduler-0" Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.272789 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fdcb1fee-b3e5-4ed1-b868-875220d969be-config-data\") pod \"nova-metadata-0\" (UID: \"fdcb1fee-b3e5-4ed1-b868-875220d969be\") " pod="openstack/nova-metadata-0" Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.273002 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/156ae83a-3fe2-4947-8ae5-5e665f19325a-logs\") pod \"nova-api-0\" (UID: \"156ae83a-3fe2-4947-8ae5-5e665f19325a\") " pod="openstack/nova-api-0" Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.273049 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/156ae83a-3fe2-4947-8ae5-5e665f19325a-config-data\") pod \"nova-api-0\" (UID: \"156ae83a-3fe2-4947-8ae5-5e665f19325a\") " pod="openstack/nova-api-0" Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.273184 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jplqw\" (UniqueName: \"kubernetes.io/projected/156ae83a-3fe2-4947-8ae5-5e665f19325a-kube-api-access-jplqw\") pod \"nova-api-0\" (UID: \"156ae83a-3fe2-4947-8ae5-5e665f19325a\") " pod="openstack/nova-api-0" Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.273223 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/fdcb1fee-b3e5-4ed1-b868-875220d969be-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"fdcb1fee-b3e5-4ed1-b868-875220d969be\") " pod="openstack/nova-metadata-0" Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.273407 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/156ae83a-3fe2-4947-8ae5-5e665f19325a-logs\") pod \"nova-api-0\" (UID: \"156ae83a-3fe2-4947-8ae5-5e665f19325a\") " pod="openstack/nova-api-0" Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.276782 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/156ae83a-3fe2-4947-8ae5-5e665f19325a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"156ae83a-3fe2-4947-8ae5-5e665f19325a\") " pod="openstack/nova-api-0" Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.277247 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/156ae83a-3fe2-4947-8ae5-5e665f19325a-config-data\") pod \"nova-api-0\" (UID: \"156ae83a-3fe2-4947-8ae5-5e665f19325a\") " pod="openstack/nova-api-0" Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.288696 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jplqw\" (UniqueName: \"kubernetes.io/projected/156ae83a-3fe2-4947-8ae5-5e665f19325a-kube-api-access-jplqw\") pod \"nova-api-0\" (UID: \"156ae83a-3fe2-4947-8ae5-5e665f19325a\") " pod="openstack/nova-api-0" Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.360066 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.374685 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdcb1fee-b3e5-4ed1-b868-875220d969be-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"fdcb1fee-b3e5-4ed1-b868-875220d969be\") " pod="openstack/nova-metadata-0" Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.374919 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fdcb1fee-b3e5-4ed1-b868-875220d969be-logs\") pod \"nova-metadata-0\" (UID: \"fdcb1fee-b3e5-4ed1-b868-875220d969be\") " pod="openstack/nova-metadata-0" Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.375061 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed6ed6ec-40bd-4628-91c8-2410b68f7384-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ed6ed6ec-40bd-4628-91c8-2410b68f7384\") " pod="openstack/nova-scheduler-0" Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.375212 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fdcb1fee-b3e5-4ed1-b868-875220d969be-config-data\") pod \"nova-metadata-0\" (UID: \"fdcb1fee-b3e5-4ed1-b868-875220d969be\") " pod="openstack/nova-metadata-0" Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.375366 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/fdcb1fee-b3e5-4ed1-b868-875220d969be-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"fdcb1fee-b3e5-4ed1-b868-875220d969be\") " pod="openstack/nova-metadata-0" Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.375465 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4jqpm\" (UniqueName: \"kubernetes.io/projected/fdcb1fee-b3e5-4ed1-b868-875220d969be-kube-api-access-4jqpm\") pod \"nova-metadata-0\" (UID: \"fdcb1fee-b3e5-4ed1-b868-875220d969be\") " pod="openstack/nova-metadata-0" Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.375583 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed6ed6ec-40bd-4628-91c8-2410b68f7384-config-data\") pod \"nova-scheduler-0\" (UID: \"ed6ed6ec-40bd-4628-91c8-2410b68f7384\") " pod="openstack/nova-scheduler-0" Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.375694 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mlh98\" (UniqueName: \"kubernetes.io/projected/ed6ed6ec-40bd-4628-91c8-2410b68f7384-kube-api-access-mlh98\") pod \"nova-scheduler-0\" (UID: \"ed6ed6ec-40bd-4628-91c8-2410b68f7384\") " pod="openstack/nova-scheduler-0" Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.375521 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fdcb1fee-b3e5-4ed1-b868-875220d969be-logs\") pod \"nova-metadata-0\" (UID: \"fdcb1fee-b3e5-4ed1-b868-875220d969be\") " pod="openstack/nova-metadata-0" Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.380482 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed6ed6ec-40bd-4628-91c8-2410b68f7384-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ed6ed6ec-40bd-4628-91c8-2410b68f7384\") " pod="openstack/nova-scheduler-0" Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.382069 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/fdcb1fee-b3e5-4ed1-b868-875220d969be-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"fdcb1fee-b3e5-4ed1-b868-875220d969be\") " pod="openstack/nova-metadata-0" Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.382235 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fdcb1fee-b3e5-4ed1-b868-875220d969be-config-data\") pod \"nova-metadata-0\" (UID: \"fdcb1fee-b3e5-4ed1-b868-875220d969be\") " pod="openstack/nova-metadata-0" Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.382315 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdcb1fee-b3e5-4ed1-b868-875220d969be-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"fdcb1fee-b3e5-4ed1-b868-875220d969be\") " pod="openstack/nova-metadata-0" Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.384271 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed6ed6ec-40bd-4628-91c8-2410b68f7384-config-data\") pod \"nova-scheduler-0\" (UID: \"ed6ed6ec-40bd-4628-91c8-2410b68f7384\") " pod="openstack/nova-scheduler-0" Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.398903 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mlh98\" (UniqueName: \"kubernetes.io/projected/ed6ed6ec-40bd-4628-91c8-2410b68f7384-kube-api-access-mlh98\") pod \"nova-scheduler-0\" (UID: \"ed6ed6ec-40bd-4628-91c8-2410b68f7384\") " pod="openstack/nova-scheduler-0" Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.400950 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4jqpm\" (UniqueName: \"kubernetes.io/projected/fdcb1fee-b3e5-4ed1-b868-875220d969be-kube-api-access-4jqpm\") pod \"nova-metadata-0\" (UID: \"fdcb1fee-b3e5-4ed1-b868-875220d969be\") " pod="openstack/nova-metadata-0" Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.496328 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.507272 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.678300 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="77330671-85cf-40c8-81b0-7a7fe1f480a6" path="/var/lib/kubelet/pods/77330671-85cf-40c8-81b0-7a7fe1f480a6/volumes" Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.679027 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1ae3465-8479-4587-a052-facebe0676d8" path="/var/lib/kubelet/pods/a1ae3465-8479-4587-a052-facebe0676d8/volumes" Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.679792 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c9ffe38d-2cbd-40fa-ae1d-12b42e721bb4" path="/var/lib/kubelet/pods/c9ffe38d-2cbd-40fa-ae1d-12b42e721bb4/volumes" Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.681068 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f9195c58-0e28-495d-9b97-cce7c292b7ac" path="/var/lib/kubelet/pods/f9195c58-0e28-495d-9b97-cce7c292b7ac/volumes" Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.858759 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.948098 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"156ae83a-3fe2-4947-8ae5-5e665f19325a","Type":"ContainerStarted","Data":"414c1d2ac8b747511a2e99394e45be9443459487e6fad6faba39dfb474d8fc7d"} Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.952875 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"098f28a4-4852-4c3a-ada9-7b8b9272c3ae","Type":"ContainerStarted","Data":"27178a7fdd8abdc5d39859ecccf3ea8552c04ea50f3e470f3de2dc02d130d463"} Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.954106 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.979992 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 07 16:21:27 crc kubenswrapper[4716]: I1207 16:21:27.987850 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.9878247399999998 podStartE2EDuration="2.98782474s" podCreationTimestamp="2025-12-07 16:21:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:21:27.972635571 +0000 UTC m=+1150.662920493" watchObservedRunningTime="2025-12-07 16:21:27.98782474 +0000 UTC m=+1150.678109672" Dec 07 16:21:27 crc kubenswrapper[4716]: W1207 16:21:27.990805 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfdcb1fee_b3e5_4ed1_b868_875220d969be.slice/crio-8888e46d76c01eb12230bcdefa7ef3662a8bdf028b1170cfe2cc088ffd8fe61b WatchSource:0}: Error finding container 8888e46d76c01eb12230bcdefa7ef3662a8bdf028b1170cfe2cc088ffd8fe61b: Status 404 returned error can't find the container with id 8888e46d76c01eb12230bcdefa7ef3662a8bdf028b1170cfe2cc088ffd8fe61b Dec 07 16:21:28 crc kubenswrapper[4716]: I1207 16:21:28.079304 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 07 16:21:28 crc kubenswrapper[4716]: I1207 16:21:28.964851 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ed6ed6ec-40bd-4628-91c8-2410b68f7384","Type":"ContainerStarted","Data":"57fcac8ed81c947a24d209d04c59bfeb69b36b49d18fa729d309902f2e7b5e08"} Dec 07 16:21:28 crc kubenswrapper[4716]: I1207 16:21:28.965236 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ed6ed6ec-40bd-4628-91c8-2410b68f7384","Type":"ContainerStarted","Data":"b5767352f18a00cb5ba3b22faaca38a55b64a41cb97249d96fd0f9fcf4ef684b"} Dec 07 16:21:28 crc kubenswrapper[4716]: I1207 16:21:28.968528 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fdcb1fee-b3e5-4ed1-b868-875220d969be","Type":"ContainerStarted","Data":"906d4a87c6ec35f764939dacb31e576c109400d4b65951469b0bde9c6098df07"} Dec 07 16:21:28 crc kubenswrapper[4716]: I1207 16:21:28.968564 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fdcb1fee-b3e5-4ed1-b868-875220d969be","Type":"ContainerStarted","Data":"ef9a5cfc9371486feec9d5c023ed78c6fd30243c2a69c617f4dcb9287c1dac96"} Dec 07 16:21:28 crc kubenswrapper[4716]: I1207 16:21:28.968575 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fdcb1fee-b3e5-4ed1-b868-875220d969be","Type":"ContainerStarted","Data":"8888e46d76c01eb12230bcdefa7ef3662a8bdf028b1170cfe2cc088ffd8fe61b"} Dec 07 16:21:28 crc kubenswrapper[4716]: I1207 16:21:28.972021 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"156ae83a-3fe2-4947-8ae5-5e665f19325a","Type":"ContainerStarted","Data":"9e5593d595481b95bf77d0188463dd99be293fdb876a9b8eb29fbc597473f419"} Dec 07 16:21:28 crc kubenswrapper[4716]: I1207 16:21:28.972065 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"156ae83a-3fe2-4947-8ae5-5e665f19325a","Type":"ContainerStarted","Data":"a57d656873ef4fdd9843cc133b204d2f76a4ad3bda5244b6df1d075a4c7bd148"} Dec 07 16:21:28 crc kubenswrapper[4716]: I1207 16:21:28.991346 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.991327111 podStartE2EDuration="2.991327111s" podCreationTimestamp="2025-12-07 16:21:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:21:28.984383044 +0000 UTC m=+1151.674667956" watchObservedRunningTime="2025-12-07 16:21:28.991327111 +0000 UTC m=+1151.681612023" Dec 07 16:21:29 crc kubenswrapper[4716]: I1207 16:21:29.020594 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.02056989 podStartE2EDuration="2.02056989s" podCreationTimestamp="2025-12-07 16:21:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:21:29.019680335 +0000 UTC m=+1151.709965267" watchObservedRunningTime="2025-12-07 16:21:29.02056989 +0000 UTC m=+1151.710854792" Dec 07 16:21:29 crc kubenswrapper[4716]: I1207 16:21:29.054027 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.054002322 podStartE2EDuration="3.054002322s" podCreationTimestamp="2025-12-07 16:21:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:21:29.04317997 +0000 UTC m=+1151.733464882" watchObservedRunningTime="2025-12-07 16:21:29.054002322 +0000 UTC m=+1151.744287244" Dec 07 16:21:32 crc kubenswrapper[4716]: I1207 16:21:32.496836 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 07 16:21:32 crc kubenswrapper[4716]: I1207 16:21:32.508142 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 07 16:21:32 crc kubenswrapper[4716]: I1207 16:21:32.508222 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 07 16:21:36 crc kubenswrapper[4716]: I1207 16:21:36.260572 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 07 16:21:37 crc kubenswrapper[4716]: I1207 16:21:37.360451 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 07 16:21:37 crc kubenswrapper[4716]: I1207 16:21:37.361149 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 07 16:21:37 crc kubenswrapper[4716]: I1207 16:21:37.496858 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 07 16:21:37 crc kubenswrapper[4716]: I1207 16:21:37.507719 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 07 16:21:37 crc kubenswrapper[4716]: I1207 16:21:37.507762 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 07 16:21:37 crc kubenswrapper[4716]: I1207 16:21:37.558930 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 07 16:21:38 crc kubenswrapper[4716]: I1207 16:21:38.099643 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 07 16:21:38 crc kubenswrapper[4716]: I1207 16:21:38.443490 4716 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="156ae83a-3fe2-4947-8ae5-5e665f19325a" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.193:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 07 16:21:38 crc kubenswrapper[4716]: I1207 16:21:38.443540 4716 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="156ae83a-3fe2-4947-8ae5-5e665f19325a" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.193:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 07 16:21:38 crc kubenswrapper[4716]: I1207 16:21:38.523435 4716 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="fdcb1fee-b3e5-4ed1-b868-875220d969be" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.195:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 07 16:21:38 crc kubenswrapper[4716]: I1207 16:21:38.523454 4716 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="fdcb1fee-b3e5-4ed1-b868-875220d969be" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.195:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 07 16:21:47 crc kubenswrapper[4716]: I1207 16:21:47.365172 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 07 16:21:47 crc kubenswrapper[4716]: I1207 16:21:47.366282 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 07 16:21:47 crc kubenswrapper[4716]: I1207 16:21:47.367567 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 07 16:21:47 crc kubenswrapper[4716]: I1207 16:21:47.369288 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 07 16:21:47 crc kubenswrapper[4716]: I1207 16:21:47.371643 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 07 16:21:47 crc kubenswrapper[4716]: I1207 16:21:47.373860 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 07 16:21:47 crc kubenswrapper[4716]: I1207 16:21:47.521648 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 07 16:21:47 crc kubenswrapper[4716]: I1207 16:21:47.523927 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 07 16:21:47 crc kubenswrapper[4716]: I1207 16:21:47.543553 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 07 16:21:47 crc kubenswrapper[4716]: I1207 16:21:47.605637 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-ld7pc"] Dec 07 16:21:47 crc kubenswrapper[4716]: I1207 16:21:47.607808 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-ld7pc" Dec 07 16:21:47 crc kubenswrapper[4716]: I1207 16:21:47.632255 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-ld7pc"] Dec 07 16:21:47 crc kubenswrapper[4716]: I1207 16:21:47.638141 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bb9f2412-53a4-4e98-9c37-93b57e6df8a9-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-ld7pc\" (UID: \"bb9f2412-53a4-4e98-9c37-93b57e6df8a9\") " pod="openstack/dnsmasq-dns-89c5cd4d5-ld7pc" Dec 07 16:21:47 crc kubenswrapper[4716]: I1207 16:21:47.638232 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bb9f2412-53a4-4e98-9c37-93b57e6df8a9-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-ld7pc\" (UID: \"bb9f2412-53a4-4e98-9c37-93b57e6df8a9\") " pod="openstack/dnsmasq-dns-89c5cd4d5-ld7pc" Dec 07 16:21:47 crc kubenswrapper[4716]: I1207 16:21:47.638521 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bb9f2412-53a4-4e98-9c37-93b57e6df8a9-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-ld7pc\" (UID: \"bb9f2412-53a4-4e98-9c37-93b57e6df8a9\") " pod="openstack/dnsmasq-dns-89c5cd4d5-ld7pc" Dec 07 16:21:47 crc kubenswrapper[4716]: I1207 16:21:47.638567 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb9f2412-53a4-4e98-9c37-93b57e6df8a9-config\") pod \"dnsmasq-dns-89c5cd4d5-ld7pc\" (UID: \"bb9f2412-53a4-4e98-9c37-93b57e6df8a9\") " pod="openstack/dnsmasq-dns-89c5cd4d5-ld7pc" Dec 07 16:21:47 crc kubenswrapper[4716]: I1207 16:21:47.638714 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w6ptf\" (UniqueName: \"kubernetes.io/projected/bb9f2412-53a4-4e98-9c37-93b57e6df8a9-kube-api-access-w6ptf\") pod \"dnsmasq-dns-89c5cd4d5-ld7pc\" (UID: \"bb9f2412-53a4-4e98-9c37-93b57e6df8a9\") " pod="openstack/dnsmasq-dns-89c5cd4d5-ld7pc" Dec 07 16:21:47 crc kubenswrapper[4716]: I1207 16:21:47.638762 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bb9f2412-53a4-4e98-9c37-93b57e6df8a9-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-ld7pc\" (UID: \"bb9f2412-53a4-4e98-9c37-93b57e6df8a9\") " pod="openstack/dnsmasq-dns-89c5cd4d5-ld7pc" Dec 07 16:21:47 crc kubenswrapper[4716]: I1207 16:21:47.740357 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bb9f2412-53a4-4e98-9c37-93b57e6df8a9-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-ld7pc\" (UID: \"bb9f2412-53a4-4e98-9c37-93b57e6df8a9\") " pod="openstack/dnsmasq-dns-89c5cd4d5-ld7pc" Dec 07 16:21:47 crc kubenswrapper[4716]: I1207 16:21:47.740516 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bb9f2412-53a4-4e98-9c37-93b57e6df8a9-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-ld7pc\" (UID: \"bb9f2412-53a4-4e98-9c37-93b57e6df8a9\") " pod="openstack/dnsmasq-dns-89c5cd4d5-ld7pc" Dec 07 16:21:47 crc kubenswrapper[4716]: I1207 16:21:47.740570 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bb9f2412-53a4-4e98-9c37-93b57e6df8a9-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-ld7pc\" (UID: \"bb9f2412-53a4-4e98-9c37-93b57e6df8a9\") " pod="openstack/dnsmasq-dns-89c5cd4d5-ld7pc" Dec 07 16:21:47 crc kubenswrapper[4716]: I1207 16:21:47.740600 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bb9f2412-53a4-4e98-9c37-93b57e6df8a9-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-ld7pc\" (UID: \"bb9f2412-53a4-4e98-9c37-93b57e6df8a9\") " pod="openstack/dnsmasq-dns-89c5cd4d5-ld7pc" Dec 07 16:21:47 crc kubenswrapper[4716]: I1207 16:21:47.740642 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb9f2412-53a4-4e98-9c37-93b57e6df8a9-config\") pod \"dnsmasq-dns-89c5cd4d5-ld7pc\" (UID: \"bb9f2412-53a4-4e98-9c37-93b57e6df8a9\") " pod="openstack/dnsmasq-dns-89c5cd4d5-ld7pc" Dec 07 16:21:47 crc kubenswrapper[4716]: I1207 16:21:47.740763 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w6ptf\" (UniqueName: \"kubernetes.io/projected/bb9f2412-53a4-4e98-9c37-93b57e6df8a9-kube-api-access-w6ptf\") pod \"dnsmasq-dns-89c5cd4d5-ld7pc\" (UID: \"bb9f2412-53a4-4e98-9c37-93b57e6df8a9\") " pod="openstack/dnsmasq-dns-89c5cd4d5-ld7pc" Dec 07 16:21:47 crc kubenswrapper[4716]: I1207 16:21:47.742695 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb9f2412-53a4-4e98-9c37-93b57e6df8a9-config\") pod \"dnsmasq-dns-89c5cd4d5-ld7pc\" (UID: \"bb9f2412-53a4-4e98-9c37-93b57e6df8a9\") " pod="openstack/dnsmasq-dns-89c5cd4d5-ld7pc" Dec 07 16:21:47 crc kubenswrapper[4716]: I1207 16:21:47.743112 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bb9f2412-53a4-4e98-9c37-93b57e6df8a9-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-ld7pc\" (UID: \"bb9f2412-53a4-4e98-9c37-93b57e6df8a9\") " pod="openstack/dnsmasq-dns-89c5cd4d5-ld7pc" Dec 07 16:21:47 crc kubenswrapper[4716]: I1207 16:21:47.743325 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bb9f2412-53a4-4e98-9c37-93b57e6df8a9-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-ld7pc\" (UID: \"bb9f2412-53a4-4e98-9c37-93b57e6df8a9\") " pod="openstack/dnsmasq-dns-89c5cd4d5-ld7pc" Dec 07 16:21:47 crc kubenswrapper[4716]: I1207 16:21:47.743473 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bb9f2412-53a4-4e98-9c37-93b57e6df8a9-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-ld7pc\" (UID: \"bb9f2412-53a4-4e98-9c37-93b57e6df8a9\") " pod="openstack/dnsmasq-dns-89c5cd4d5-ld7pc" Dec 07 16:21:47 crc kubenswrapper[4716]: I1207 16:21:47.744324 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bb9f2412-53a4-4e98-9c37-93b57e6df8a9-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-ld7pc\" (UID: \"bb9f2412-53a4-4e98-9c37-93b57e6df8a9\") " pod="openstack/dnsmasq-dns-89c5cd4d5-ld7pc" Dec 07 16:21:47 crc kubenswrapper[4716]: I1207 16:21:47.760247 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w6ptf\" (UniqueName: \"kubernetes.io/projected/bb9f2412-53a4-4e98-9c37-93b57e6df8a9-kube-api-access-w6ptf\") pod \"dnsmasq-dns-89c5cd4d5-ld7pc\" (UID: \"bb9f2412-53a4-4e98-9c37-93b57e6df8a9\") " pod="openstack/dnsmasq-dns-89c5cd4d5-ld7pc" Dec 07 16:21:47 crc kubenswrapper[4716]: I1207 16:21:47.930846 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-ld7pc" Dec 07 16:21:48 crc kubenswrapper[4716]: I1207 16:21:48.178664 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 07 16:21:48 crc kubenswrapper[4716]: I1207 16:21:48.684467 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-ld7pc"] Dec 07 16:21:49 crc kubenswrapper[4716]: I1207 16:21:49.182618 4716 generic.go:334] "Generic (PLEG): container finished" podID="bb9f2412-53a4-4e98-9c37-93b57e6df8a9" containerID="13927b13b0cabeb3877b59bc7c647efd771bc28b6871f9c1cc5308712fa4d0e3" exitCode=0 Dec 07 16:21:49 crc kubenswrapper[4716]: I1207 16:21:49.182776 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-ld7pc" event={"ID":"bb9f2412-53a4-4e98-9c37-93b57e6df8a9","Type":"ContainerDied","Data":"13927b13b0cabeb3877b59bc7c647efd771bc28b6871f9c1cc5308712fa4d0e3"} Dec 07 16:21:49 crc kubenswrapper[4716]: I1207 16:21:49.182910 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-ld7pc" event={"ID":"bb9f2412-53a4-4e98-9c37-93b57e6df8a9","Type":"ContainerStarted","Data":"37db71dde61341ccec072e9a56d289580dfda9719d299b4f41110e55c929e643"} Dec 07 16:21:49 crc kubenswrapper[4716]: I1207 16:21:49.346318 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 07 16:21:49 crc kubenswrapper[4716]: I1207 16:21:49.346612 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="859fd7aa-6477-4335-a671-a760a1dd7cf8" containerName="ceilometer-central-agent" containerID="cri-o://11ff631224cb8de030199944568fa2d79ef30c81639510b9f243eec113ca4ca3" gracePeriod=30 Dec 07 16:21:49 crc kubenswrapper[4716]: I1207 16:21:49.346723 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="859fd7aa-6477-4335-a671-a760a1dd7cf8" containerName="proxy-httpd" containerID="cri-o://f7ee11dae27dd6d0d58a7912b0ae41e6c384cc2424cdf86625543cd75053e7c6" gracePeriod=30 Dec 07 16:21:49 crc kubenswrapper[4716]: I1207 16:21:49.346778 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="859fd7aa-6477-4335-a671-a760a1dd7cf8" containerName="sg-core" containerID="cri-o://01ea3429fecbd3394b8b9df1392f6e7969d731311de9f2029960bc25ad215ba0" gracePeriod=30 Dec 07 16:21:49 crc kubenswrapper[4716]: I1207 16:21:49.346813 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="859fd7aa-6477-4335-a671-a760a1dd7cf8" containerName="ceilometer-notification-agent" containerID="cri-o://f7b515d9c5e9dcc7f85ce5e46b76c9af49507ad42f2d3fb0fa16bcf2f017f206" gracePeriod=30 Dec 07 16:21:50 crc kubenswrapper[4716]: I1207 16:21:50.143094 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 07 16:21:50 crc kubenswrapper[4716]: I1207 16:21:50.208694 4716 generic.go:334] "Generic (PLEG): container finished" podID="859fd7aa-6477-4335-a671-a760a1dd7cf8" containerID="f7ee11dae27dd6d0d58a7912b0ae41e6c384cc2424cdf86625543cd75053e7c6" exitCode=0 Dec 07 16:21:50 crc kubenswrapper[4716]: I1207 16:21:50.208737 4716 generic.go:334] "Generic (PLEG): container finished" podID="859fd7aa-6477-4335-a671-a760a1dd7cf8" containerID="01ea3429fecbd3394b8b9df1392f6e7969d731311de9f2029960bc25ad215ba0" exitCode=2 Dec 07 16:21:50 crc kubenswrapper[4716]: I1207 16:21:50.208750 4716 generic.go:334] "Generic (PLEG): container finished" podID="859fd7aa-6477-4335-a671-a760a1dd7cf8" containerID="11ff631224cb8de030199944568fa2d79ef30c81639510b9f243eec113ca4ca3" exitCode=0 Dec 07 16:21:50 crc kubenswrapper[4716]: I1207 16:21:50.208803 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"859fd7aa-6477-4335-a671-a760a1dd7cf8","Type":"ContainerDied","Data":"f7ee11dae27dd6d0d58a7912b0ae41e6c384cc2424cdf86625543cd75053e7c6"} Dec 07 16:21:50 crc kubenswrapper[4716]: I1207 16:21:50.208837 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"859fd7aa-6477-4335-a671-a760a1dd7cf8","Type":"ContainerDied","Data":"01ea3429fecbd3394b8b9df1392f6e7969d731311de9f2029960bc25ad215ba0"} Dec 07 16:21:50 crc kubenswrapper[4716]: I1207 16:21:50.208851 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"859fd7aa-6477-4335-a671-a760a1dd7cf8","Type":"ContainerDied","Data":"11ff631224cb8de030199944568fa2d79ef30c81639510b9f243eec113ca4ca3"} Dec 07 16:21:50 crc kubenswrapper[4716]: I1207 16:21:50.210718 4716 generic.go:334] "Generic (PLEG): container finished" podID="3a88434a-c17e-4de2-be0e-bf5c951012b2" containerID="08d95b491dfc6c45ca2b0f45ca175db851104e371ff2993f44bc54dd1187a719" exitCode=137 Dec 07 16:21:50 crc kubenswrapper[4716]: I1207 16:21:50.210773 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"3a88434a-c17e-4de2-be0e-bf5c951012b2","Type":"ContainerDied","Data":"08d95b491dfc6c45ca2b0f45ca175db851104e371ff2993f44bc54dd1187a719"} Dec 07 16:21:50 crc kubenswrapper[4716]: I1207 16:21:50.210790 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"3a88434a-c17e-4de2-be0e-bf5c951012b2","Type":"ContainerDied","Data":"0414a171402f251ee3e00b77c595c00f52269d16e904f154f3a1627639cb68a7"} Dec 07 16:21:50 crc kubenswrapper[4716]: I1207 16:21:50.210801 4716 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0414a171402f251ee3e00b77c595c00f52269d16e904f154f3a1627639cb68a7" Dec 07 16:21:50 crc kubenswrapper[4716]: I1207 16:21:50.212973 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-ld7pc" event={"ID":"bb9f2412-53a4-4e98-9c37-93b57e6df8a9","Type":"ContainerStarted","Data":"96a44c2634479106efbff992e763b0c05a596d0ebc22f57a5e6740f60c32c468"} Dec 07 16:21:50 crc kubenswrapper[4716]: I1207 16:21:50.213134 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="156ae83a-3fe2-4947-8ae5-5e665f19325a" containerName="nova-api-log" containerID="cri-o://a57d656873ef4fdd9843cc133b204d2f76a4ad3bda5244b6df1d075a4c7bd148" gracePeriod=30 Dec 07 16:21:50 crc kubenswrapper[4716]: I1207 16:21:50.213455 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="156ae83a-3fe2-4947-8ae5-5e665f19325a" containerName="nova-api-api" containerID="cri-o://9e5593d595481b95bf77d0188463dd99be293fdb876a9b8eb29fbc597473f419" gracePeriod=30 Dec 07 16:21:50 crc kubenswrapper[4716]: I1207 16:21:50.240348 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-89c5cd4d5-ld7pc" podStartSLOduration=3.240066108 podStartE2EDuration="3.240066108s" podCreationTimestamp="2025-12-07 16:21:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:21:50.237618252 +0000 UTC m=+1172.927903164" watchObservedRunningTime="2025-12-07 16:21:50.240066108 +0000 UTC m=+1172.930351020" Dec 07 16:21:50 crc kubenswrapper[4716]: I1207 16:21:50.246941 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 07 16:21:50 crc kubenswrapper[4716]: I1207 16:21:50.391578 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a88434a-c17e-4de2-be0e-bf5c951012b2-combined-ca-bundle\") pod \"3a88434a-c17e-4de2-be0e-bf5c951012b2\" (UID: \"3a88434a-c17e-4de2-be0e-bf5c951012b2\") " Dec 07 16:21:50 crc kubenswrapper[4716]: I1207 16:21:50.391951 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vdpn8\" (UniqueName: \"kubernetes.io/projected/3a88434a-c17e-4de2-be0e-bf5c951012b2-kube-api-access-vdpn8\") pod \"3a88434a-c17e-4de2-be0e-bf5c951012b2\" (UID: \"3a88434a-c17e-4de2-be0e-bf5c951012b2\") " Dec 07 16:21:50 crc kubenswrapper[4716]: I1207 16:21:50.392208 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a88434a-c17e-4de2-be0e-bf5c951012b2-config-data\") pod \"3a88434a-c17e-4de2-be0e-bf5c951012b2\" (UID: \"3a88434a-c17e-4de2-be0e-bf5c951012b2\") " Dec 07 16:21:50 crc kubenswrapper[4716]: I1207 16:21:50.397038 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a88434a-c17e-4de2-be0e-bf5c951012b2-kube-api-access-vdpn8" (OuterVolumeSpecName: "kube-api-access-vdpn8") pod "3a88434a-c17e-4de2-be0e-bf5c951012b2" (UID: "3a88434a-c17e-4de2-be0e-bf5c951012b2"). InnerVolumeSpecName "kube-api-access-vdpn8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:21:50 crc kubenswrapper[4716]: I1207 16:21:50.419227 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a88434a-c17e-4de2-be0e-bf5c951012b2-config-data" (OuterVolumeSpecName: "config-data") pod "3a88434a-c17e-4de2-be0e-bf5c951012b2" (UID: "3a88434a-c17e-4de2-be0e-bf5c951012b2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:21:50 crc kubenswrapper[4716]: I1207 16:21:50.427151 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a88434a-c17e-4de2-be0e-bf5c951012b2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3a88434a-c17e-4de2-be0e-bf5c951012b2" (UID: "3a88434a-c17e-4de2-be0e-bf5c951012b2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:21:50 crc kubenswrapper[4716]: I1207 16:21:50.494392 4716 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a88434a-c17e-4de2-be0e-bf5c951012b2-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 16:21:50 crc kubenswrapper[4716]: I1207 16:21:50.494654 4716 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a88434a-c17e-4de2-be0e-bf5c951012b2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 16:21:50 crc kubenswrapper[4716]: I1207 16:21:50.494718 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vdpn8\" (UniqueName: \"kubernetes.io/projected/3a88434a-c17e-4de2-be0e-bf5c951012b2-kube-api-access-vdpn8\") on node \"crc\" DevicePath \"\"" Dec 07 16:21:51 crc kubenswrapper[4716]: I1207 16:21:51.232683 4716 generic.go:334] "Generic (PLEG): container finished" podID="156ae83a-3fe2-4947-8ae5-5e665f19325a" containerID="a57d656873ef4fdd9843cc133b204d2f76a4ad3bda5244b6df1d075a4c7bd148" exitCode=143 Dec 07 16:21:51 crc kubenswrapper[4716]: I1207 16:21:51.232773 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"156ae83a-3fe2-4947-8ae5-5e665f19325a","Type":"ContainerDied","Data":"a57d656873ef4fdd9843cc133b204d2f76a4ad3bda5244b6df1d075a4c7bd148"} Dec 07 16:21:51 crc kubenswrapper[4716]: I1207 16:21:51.233032 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 07 16:21:51 crc kubenswrapper[4716]: I1207 16:21:51.233064 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-89c5cd4d5-ld7pc" Dec 07 16:21:51 crc kubenswrapper[4716]: I1207 16:21:51.288679 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 07 16:21:51 crc kubenswrapper[4716]: I1207 16:21:51.310126 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 07 16:21:51 crc kubenswrapper[4716]: I1207 16:21:51.323130 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 07 16:21:51 crc kubenswrapper[4716]: E1207 16:21:51.323593 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a88434a-c17e-4de2-be0e-bf5c951012b2" containerName="nova-cell1-novncproxy-novncproxy" Dec 07 16:21:51 crc kubenswrapper[4716]: I1207 16:21:51.323607 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a88434a-c17e-4de2-be0e-bf5c951012b2" containerName="nova-cell1-novncproxy-novncproxy" Dec 07 16:21:51 crc kubenswrapper[4716]: I1207 16:21:51.323813 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a88434a-c17e-4de2-be0e-bf5c951012b2" containerName="nova-cell1-novncproxy-novncproxy" Dec 07 16:21:51 crc kubenswrapper[4716]: I1207 16:21:51.324557 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 07 16:21:51 crc kubenswrapper[4716]: I1207 16:21:51.330970 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 07 16:21:51 crc kubenswrapper[4716]: I1207 16:21:51.331192 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Dec 07 16:21:51 crc kubenswrapper[4716]: I1207 16:21:51.331312 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Dec 07 16:21:51 crc kubenswrapper[4716]: I1207 16:21:51.332112 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 07 16:21:51 crc kubenswrapper[4716]: I1207 16:21:51.516380 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0c9c313-94be-404d-849d-3eb23c5eb5e5-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"c0c9c313-94be-404d-849d-3eb23c5eb5e5\") " pod="openstack/nova-cell1-novncproxy-0" Dec 07 16:21:51 crc kubenswrapper[4716]: I1207 16:21:51.516600 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0c9c313-94be-404d-849d-3eb23c5eb5e5-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"c0c9c313-94be-404d-849d-3eb23c5eb5e5\") " pod="openstack/nova-cell1-novncproxy-0" Dec 07 16:21:51 crc kubenswrapper[4716]: I1207 16:21:51.516760 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/c0c9c313-94be-404d-849d-3eb23c5eb5e5-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"c0c9c313-94be-404d-849d-3eb23c5eb5e5\") " pod="openstack/nova-cell1-novncproxy-0" Dec 07 16:21:51 crc kubenswrapper[4716]: I1207 16:21:51.516790 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/c0c9c313-94be-404d-849d-3eb23c5eb5e5-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"c0c9c313-94be-404d-849d-3eb23c5eb5e5\") " pod="openstack/nova-cell1-novncproxy-0" Dec 07 16:21:51 crc kubenswrapper[4716]: I1207 16:21:51.516851 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79m6b\" (UniqueName: \"kubernetes.io/projected/c0c9c313-94be-404d-849d-3eb23c5eb5e5-kube-api-access-79m6b\") pod \"nova-cell1-novncproxy-0\" (UID: \"c0c9c313-94be-404d-849d-3eb23c5eb5e5\") " pod="openstack/nova-cell1-novncproxy-0" Dec 07 16:21:51 crc kubenswrapper[4716]: I1207 16:21:51.618485 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/c0c9c313-94be-404d-849d-3eb23c5eb5e5-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"c0c9c313-94be-404d-849d-3eb23c5eb5e5\") " pod="openstack/nova-cell1-novncproxy-0" Dec 07 16:21:51 crc kubenswrapper[4716]: I1207 16:21:51.618528 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/c0c9c313-94be-404d-849d-3eb23c5eb5e5-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"c0c9c313-94be-404d-849d-3eb23c5eb5e5\") " pod="openstack/nova-cell1-novncproxy-0" Dec 07 16:21:51 crc kubenswrapper[4716]: I1207 16:21:51.618557 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79m6b\" (UniqueName: \"kubernetes.io/projected/c0c9c313-94be-404d-849d-3eb23c5eb5e5-kube-api-access-79m6b\") pod \"nova-cell1-novncproxy-0\" (UID: \"c0c9c313-94be-404d-849d-3eb23c5eb5e5\") " pod="openstack/nova-cell1-novncproxy-0" Dec 07 16:21:51 crc kubenswrapper[4716]: I1207 16:21:51.618639 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0c9c313-94be-404d-849d-3eb23c5eb5e5-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"c0c9c313-94be-404d-849d-3eb23c5eb5e5\") " pod="openstack/nova-cell1-novncproxy-0" Dec 07 16:21:51 crc kubenswrapper[4716]: I1207 16:21:51.618730 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0c9c313-94be-404d-849d-3eb23c5eb5e5-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"c0c9c313-94be-404d-849d-3eb23c5eb5e5\") " pod="openstack/nova-cell1-novncproxy-0" Dec 07 16:21:51 crc kubenswrapper[4716]: I1207 16:21:51.624553 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0c9c313-94be-404d-849d-3eb23c5eb5e5-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"c0c9c313-94be-404d-849d-3eb23c5eb5e5\") " pod="openstack/nova-cell1-novncproxy-0" Dec 07 16:21:51 crc kubenswrapper[4716]: I1207 16:21:51.624812 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/c0c9c313-94be-404d-849d-3eb23c5eb5e5-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"c0c9c313-94be-404d-849d-3eb23c5eb5e5\") " pod="openstack/nova-cell1-novncproxy-0" Dec 07 16:21:51 crc kubenswrapper[4716]: I1207 16:21:51.626035 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/c0c9c313-94be-404d-849d-3eb23c5eb5e5-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"c0c9c313-94be-404d-849d-3eb23c5eb5e5\") " pod="openstack/nova-cell1-novncproxy-0" Dec 07 16:21:51 crc kubenswrapper[4716]: I1207 16:21:51.626460 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0c9c313-94be-404d-849d-3eb23c5eb5e5-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"c0c9c313-94be-404d-849d-3eb23c5eb5e5\") " pod="openstack/nova-cell1-novncproxy-0" Dec 07 16:21:51 crc kubenswrapper[4716]: I1207 16:21:51.644957 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79m6b\" (UniqueName: \"kubernetes.io/projected/c0c9c313-94be-404d-849d-3eb23c5eb5e5-kube-api-access-79m6b\") pod \"nova-cell1-novncproxy-0\" (UID: \"c0c9c313-94be-404d-849d-3eb23c5eb5e5\") " pod="openstack/nova-cell1-novncproxy-0" Dec 07 16:21:51 crc kubenswrapper[4716]: I1207 16:21:51.667307 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 07 16:21:51 crc kubenswrapper[4716]: I1207 16:21:51.689293 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a88434a-c17e-4de2-be0e-bf5c951012b2" path="/var/lib/kubelet/pods/3a88434a-c17e-4de2-be0e-bf5c951012b2/volumes" Dec 07 16:21:52 crc kubenswrapper[4716]: I1207 16:21:52.123036 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 07 16:21:52 crc kubenswrapper[4716]: I1207 16:21:52.244518 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"c0c9c313-94be-404d-849d-3eb23c5eb5e5","Type":"ContainerStarted","Data":"9c242a79389f614a981617075728a1b2c35a2866a168cefaa36bb466ece96fca"} Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.022521 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.143786 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/859fd7aa-6477-4335-a671-a760a1dd7cf8-config-data\") pod \"859fd7aa-6477-4335-a671-a760a1dd7cf8\" (UID: \"859fd7aa-6477-4335-a671-a760a1dd7cf8\") " Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.143840 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/859fd7aa-6477-4335-a671-a760a1dd7cf8-ceilometer-tls-certs\") pod \"859fd7aa-6477-4335-a671-a760a1dd7cf8\" (UID: \"859fd7aa-6477-4335-a671-a760a1dd7cf8\") " Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.143882 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/859fd7aa-6477-4335-a671-a760a1dd7cf8-combined-ca-bundle\") pod \"859fd7aa-6477-4335-a671-a760a1dd7cf8\" (UID: \"859fd7aa-6477-4335-a671-a760a1dd7cf8\") " Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.143907 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/859fd7aa-6477-4335-a671-a760a1dd7cf8-scripts\") pod \"859fd7aa-6477-4335-a671-a760a1dd7cf8\" (UID: \"859fd7aa-6477-4335-a671-a760a1dd7cf8\") " Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.143992 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/859fd7aa-6477-4335-a671-a760a1dd7cf8-log-httpd\") pod \"859fd7aa-6477-4335-a671-a760a1dd7cf8\" (UID: \"859fd7aa-6477-4335-a671-a760a1dd7cf8\") " Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.144034 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/859fd7aa-6477-4335-a671-a760a1dd7cf8-run-httpd\") pod \"859fd7aa-6477-4335-a671-a760a1dd7cf8\" (UID: \"859fd7aa-6477-4335-a671-a760a1dd7cf8\") " Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.144098 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/859fd7aa-6477-4335-a671-a760a1dd7cf8-sg-core-conf-yaml\") pod \"859fd7aa-6477-4335-a671-a760a1dd7cf8\" (UID: \"859fd7aa-6477-4335-a671-a760a1dd7cf8\") " Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.144151 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xqnkh\" (UniqueName: \"kubernetes.io/projected/859fd7aa-6477-4335-a671-a760a1dd7cf8-kube-api-access-xqnkh\") pod \"859fd7aa-6477-4335-a671-a760a1dd7cf8\" (UID: \"859fd7aa-6477-4335-a671-a760a1dd7cf8\") " Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.149086 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/859fd7aa-6477-4335-a671-a760a1dd7cf8-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "859fd7aa-6477-4335-a671-a760a1dd7cf8" (UID: "859fd7aa-6477-4335-a671-a760a1dd7cf8"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.149517 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/859fd7aa-6477-4335-a671-a760a1dd7cf8-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "859fd7aa-6477-4335-a671-a760a1dd7cf8" (UID: "859fd7aa-6477-4335-a671-a760a1dd7cf8"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.151685 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/859fd7aa-6477-4335-a671-a760a1dd7cf8-kube-api-access-xqnkh" (OuterVolumeSpecName: "kube-api-access-xqnkh") pod "859fd7aa-6477-4335-a671-a760a1dd7cf8" (UID: "859fd7aa-6477-4335-a671-a760a1dd7cf8"). InnerVolumeSpecName "kube-api-access-xqnkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.155201 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/859fd7aa-6477-4335-a671-a760a1dd7cf8-scripts" (OuterVolumeSpecName: "scripts") pod "859fd7aa-6477-4335-a671-a760a1dd7cf8" (UID: "859fd7aa-6477-4335-a671-a760a1dd7cf8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.182060 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/859fd7aa-6477-4335-a671-a760a1dd7cf8-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "859fd7aa-6477-4335-a671-a760a1dd7cf8" (UID: "859fd7aa-6477-4335-a671-a760a1dd7cf8"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.201540 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/859fd7aa-6477-4335-a671-a760a1dd7cf8-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "859fd7aa-6477-4335-a671-a760a1dd7cf8" (UID: "859fd7aa-6477-4335-a671-a760a1dd7cf8"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.239869 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/859fd7aa-6477-4335-a671-a760a1dd7cf8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "859fd7aa-6477-4335-a671-a760a1dd7cf8" (UID: "859fd7aa-6477-4335-a671-a760a1dd7cf8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.248134 4716 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/859fd7aa-6477-4335-a671-a760a1dd7cf8-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.248164 4716 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/859fd7aa-6477-4335-a671-a760a1dd7cf8-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.248174 4716 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/859fd7aa-6477-4335-a671-a760a1dd7cf8-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.248186 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xqnkh\" (UniqueName: \"kubernetes.io/projected/859fd7aa-6477-4335-a671-a760a1dd7cf8-kube-api-access-xqnkh\") on node \"crc\" DevicePath \"\"" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.248196 4716 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/859fd7aa-6477-4335-a671-a760a1dd7cf8-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.248408 4716 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/859fd7aa-6477-4335-a671-a760a1dd7cf8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.248417 4716 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/859fd7aa-6477-4335-a671-a760a1dd7cf8-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.256886 4716 generic.go:334] "Generic (PLEG): container finished" podID="859fd7aa-6477-4335-a671-a760a1dd7cf8" containerID="f7b515d9c5e9dcc7f85ce5e46b76c9af49507ad42f2d3fb0fa16bcf2f017f206" exitCode=0 Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.256925 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"859fd7aa-6477-4335-a671-a760a1dd7cf8","Type":"ContainerDied","Data":"f7b515d9c5e9dcc7f85ce5e46b76c9af49507ad42f2d3fb0fa16bcf2f017f206"} Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.256983 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"859fd7aa-6477-4335-a671-a760a1dd7cf8","Type":"ContainerDied","Data":"4b8b03b5a4bdc7bc29fc1ed279f7d918e015de09a8da939857e3cf9383be859e"} Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.257003 4716 scope.go:117] "RemoveContainer" containerID="f7ee11dae27dd6d0d58a7912b0ae41e6c384cc2424cdf86625543cd75053e7c6" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.257042 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.260322 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"c0c9c313-94be-404d-849d-3eb23c5eb5e5","Type":"ContainerStarted","Data":"2c9ab5324db7c174b7e566217b8a8af59e0013725396c62c74f4442c24b1f0d9"} Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.283992 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.2839712 podStartE2EDuration="2.2839712s" podCreationTimestamp="2025-12-07 16:21:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:21:53.275660666 +0000 UTC m=+1175.965945598" watchObservedRunningTime="2025-12-07 16:21:53.2839712 +0000 UTC m=+1175.974256112" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.297605 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/859fd7aa-6477-4335-a671-a760a1dd7cf8-config-data" (OuterVolumeSpecName: "config-data") pod "859fd7aa-6477-4335-a671-a760a1dd7cf8" (UID: "859fd7aa-6477-4335-a671-a760a1dd7cf8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.352569 4716 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/859fd7aa-6477-4335-a671-a760a1dd7cf8-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.367227 4716 scope.go:117] "RemoveContainer" containerID="01ea3429fecbd3394b8b9df1392f6e7969d731311de9f2029960bc25ad215ba0" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.395793 4716 scope.go:117] "RemoveContainer" containerID="f7b515d9c5e9dcc7f85ce5e46b76c9af49507ad42f2d3fb0fa16bcf2f017f206" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.463377 4716 scope.go:117] "RemoveContainer" containerID="11ff631224cb8de030199944568fa2d79ef30c81639510b9f243eec113ca4ca3" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.481698 4716 scope.go:117] "RemoveContainer" containerID="f7ee11dae27dd6d0d58a7912b0ae41e6c384cc2424cdf86625543cd75053e7c6" Dec 07 16:21:53 crc kubenswrapper[4716]: E1207 16:21:53.482148 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f7ee11dae27dd6d0d58a7912b0ae41e6c384cc2424cdf86625543cd75053e7c6\": container with ID starting with f7ee11dae27dd6d0d58a7912b0ae41e6c384cc2424cdf86625543cd75053e7c6 not found: ID does not exist" containerID="f7ee11dae27dd6d0d58a7912b0ae41e6c384cc2424cdf86625543cd75053e7c6" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.482187 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7ee11dae27dd6d0d58a7912b0ae41e6c384cc2424cdf86625543cd75053e7c6"} err="failed to get container status \"f7ee11dae27dd6d0d58a7912b0ae41e6c384cc2424cdf86625543cd75053e7c6\": rpc error: code = NotFound desc = could not find container \"f7ee11dae27dd6d0d58a7912b0ae41e6c384cc2424cdf86625543cd75053e7c6\": container with ID starting with f7ee11dae27dd6d0d58a7912b0ae41e6c384cc2424cdf86625543cd75053e7c6 not found: ID does not exist" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.482213 4716 scope.go:117] "RemoveContainer" containerID="01ea3429fecbd3394b8b9df1392f6e7969d731311de9f2029960bc25ad215ba0" Dec 07 16:21:53 crc kubenswrapper[4716]: E1207 16:21:53.482471 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01ea3429fecbd3394b8b9df1392f6e7969d731311de9f2029960bc25ad215ba0\": container with ID starting with 01ea3429fecbd3394b8b9df1392f6e7969d731311de9f2029960bc25ad215ba0 not found: ID does not exist" containerID="01ea3429fecbd3394b8b9df1392f6e7969d731311de9f2029960bc25ad215ba0" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.482497 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01ea3429fecbd3394b8b9df1392f6e7969d731311de9f2029960bc25ad215ba0"} err="failed to get container status \"01ea3429fecbd3394b8b9df1392f6e7969d731311de9f2029960bc25ad215ba0\": rpc error: code = NotFound desc = could not find container \"01ea3429fecbd3394b8b9df1392f6e7969d731311de9f2029960bc25ad215ba0\": container with ID starting with 01ea3429fecbd3394b8b9df1392f6e7969d731311de9f2029960bc25ad215ba0 not found: ID does not exist" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.482510 4716 scope.go:117] "RemoveContainer" containerID="f7b515d9c5e9dcc7f85ce5e46b76c9af49507ad42f2d3fb0fa16bcf2f017f206" Dec 07 16:21:53 crc kubenswrapper[4716]: E1207 16:21:53.482805 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f7b515d9c5e9dcc7f85ce5e46b76c9af49507ad42f2d3fb0fa16bcf2f017f206\": container with ID starting with f7b515d9c5e9dcc7f85ce5e46b76c9af49507ad42f2d3fb0fa16bcf2f017f206 not found: ID does not exist" containerID="f7b515d9c5e9dcc7f85ce5e46b76c9af49507ad42f2d3fb0fa16bcf2f017f206" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.482825 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7b515d9c5e9dcc7f85ce5e46b76c9af49507ad42f2d3fb0fa16bcf2f017f206"} err="failed to get container status \"f7b515d9c5e9dcc7f85ce5e46b76c9af49507ad42f2d3fb0fa16bcf2f017f206\": rpc error: code = NotFound desc = could not find container \"f7b515d9c5e9dcc7f85ce5e46b76c9af49507ad42f2d3fb0fa16bcf2f017f206\": container with ID starting with f7b515d9c5e9dcc7f85ce5e46b76c9af49507ad42f2d3fb0fa16bcf2f017f206 not found: ID does not exist" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.482836 4716 scope.go:117] "RemoveContainer" containerID="11ff631224cb8de030199944568fa2d79ef30c81639510b9f243eec113ca4ca3" Dec 07 16:21:53 crc kubenswrapper[4716]: E1207 16:21:53.483157 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11ff631224cb8de030199944568fa2d79ef30c81639510b9f243eec113ca4ca3\": container with ID starting with 11ff631224cb8de030199944568fa2d79ef30c81639510b9f243eec113ca4ca3 not found: ID does not exist" containerID="11ff631224cb8de030199944568fa2d79ef30c81639510b9f243eec113ca4ca3" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.483176 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11ff631224cb8de030199944568fa2d79ef30c81639510b9f243eec113ca4ca3"} err="failed to get container status \"11ff631224cb8de030199944568fa2d79ef30c81639510b9f243eec113ca4ca3\": rpc error: code = NotFound desc = could not find container \"11ff631224cb8de030199944568fa2d79ef30c81639510b9f243eec113ca4ca3\": container with ID starting with 11ff631224cb8de030199944568fa2d79ef30c81639510b9f243eec113ca4ca3 not found: ID does not exist" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.592337 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.614143 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.635046 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 07 16:21:53 crc kubenswrapper[4716]: E1207 16:21:53.636583 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="859fd7aa-6477-4335-a671-a760a1dd7cf8" containerName="proxy-httpd" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.636636 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="859fd7aa-6477-4335-a671-a760a1dd7cf8" containerName="proxy-httpd" Dec 07 16:21:53 crc kubenswrapper[4716]: E1207 16:21:53.636670 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="859fd7aa-6477-4335-a671-a760a1dd7cf8" containerName="sg-core" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.636840 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="859fd7aa-6477-4335-a671-a760a1dd7cf8" containerName="sg-core" Dec 07 16:21:53 crc kubenswrapper[4716]: E1207 16:21:53.636897 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="859fd7aa-6477-4335-a671-a760a1dd7cf8" containerName="ceilometer-central-agent" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.636906 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="859fd7aa-6477-4335-a671-a760a1dd7cf8" containerName="ceilometer-central-agent" Dec 07 16:21:53 crc kubenswrapper[4716]: E1207 16:21:53.636929 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="859fd7aa-6477-4335-a671-a760a1dd7cf8" containerName="ceilometer-notification-agent" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.636964 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="859fd7aa-6477-4335-a671-a760a1dd7cf8" containerName="ceilometer-notification-agent" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.637340 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="859fd7aa-6477-4335-a671-a760a1dd7cf8" containerName="proxy-httpd" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.637359 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="859fd7aa-6477-4335-a671-a760a1dd7cf8" containerName="ceilometer-central-agent" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.637375 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="859fd7aa-6477-4335-a671-a760a1dd7cf8" containerName="ceilometer-notification-agent" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.637396 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="859fd7aa-6477-4335-a671-a760a1dd7cf8" containerName="sg-core" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.640568 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.642932 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.646792 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.647944 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.649072 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.679055 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="859fd7aa-6477-4335-a671-a760a1dd7cf8" path="/var/lib/kubelet/pods/859fd7aa-6477-4335-a671-a760a1dd7cf8/volumes" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.746592 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.758821 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f86be30-3a49-4354-a6e9-059592ee5ebc-config-data\") pod \"ceilometer-0\" (UID: \"5f86be30-3a49-4354-a6e9-059592ee5ebc\") " pod="openstack/ceilometer-0" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.758900 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ssj8v\" (UniqueName: \"kubernetes.io/projected/5f86be30-3a49-4354-a6e9-059592ee5ebc-kube-api-access-ssj8v\") pod \"ceilometer-0\" (UID: \"5f86be30-3a49-4354-a6e9-059592ee5ebc\") " pod="openstack/ceilometer-0" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.758992 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/5f86be30-3a49-4354-a6e9-059592ee5ebc-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"5f86be30-3a49-4354-a6e9-059592ee5ebc\") " pod="openstack/ceilometer-0" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.759065 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5f86be30-3a49-4354-a6e9-059592ee5ebc-scripts\") pod \"ceilometer-0\" (UID: \"5f86be30-3a49-4354-a6e9-059592ee5ebc\") " pod="openstack/ceilometer-0" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.759122 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5f86be30-3a49-4354-a6e9-059592ee5ebc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5f86be30-3a49-4354-a6e9-059592ee5ebc\") " pod="openstack/ceilometer-0" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.759140 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5f86be30-3a49-4354-a6e9-059592ee5ebc-log-httpd\") pod \"ceilometer-0\" (UID: \"5f86be30-3a49-4354-a6e9-059592ee5ebc\") " pod="openstack/ceilometer-0" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.759165 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5f86be30-3a49-4354-a6e9-059592ee5ebc-run-httpd\") pod \"ceilometer-0\" (UID: \"5f86be30-3a49-4354-a6e9-059592ee5ebc\") " pod="openstack/ceilometer-0" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.759258 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f86be30-3a49-4354-a6e9-059592ee5ebc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5f86be30-3a49-4354-a6e9-059592ee5ebc\") " pod="openstack/ceilometer-0" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.860072 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/156ae83a-3fe2-4947-8ae5-5e665f19325a-config-data\") pod \"156ae83a-3fe2-4947-8ae5-5e665f19325a\" (UID: \"156ae83a-3fe2-4947-8ae5-5e665f19325a\") " Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.860197 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/156ae83a-3fe2-4947-8ae5-5e665f19325a-combined-ca-bundle\") pod \"156ae83a-3fe2-4947-8ae5-5e665f19325a\" (UID: \"156ae83a-3fe2-4947-8ae5-5e665f19325a\") " Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.860276 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/156ae83a-3fe2-4947-8ae5-5e665f19325a-logs\") pod \"156ae83a-3fe2-4947-8ae5-5e665f19325a\" (UID: \"156ae83a-3fe2-4947-8ae5-5e665f19325a\") " Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.860307 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jplqw\" (UniqueName: \"kubernetes.io/projected/156ae83a-3fe2-4947-8ae5-5e665f19325a-kube-api-access-jplqw\") pod \"156ae83a-3fe2-4947-8ae5-5e665f19325a\" (UID: \"156ae83a-3fe2-4947-8ae5-5e665f19325a\") " Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.860722 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/5f86be30-3a49-4354-a6e9-059592ee5ebc-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"5f86be30-3a49-4354-a6e9-059592ee5ebc\") " pod="openstack/ceilometer-0" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.860874 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5f86be30-3a49-4354-a6e9-059592ee5ebc-scripts\") pod \"ceilometer-0\" (UID: \"5f86be30-3a49-4354-a6e9-059592ee5ebc\") " pod="openstack/ceilometer-0" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.860953 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5f86be30-3a49-4354-a6e9-059592ee5ebc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5f86be30-3a49-4354-a6e9-059592ee5ebc\") " pod="openstack/ceilometer-0" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.860973 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5f86be30-3a49-4354-a6e9-059592ee5ebc-log-httpd\") pod \"ceilometer-0\" (UID: \"5f86be30-3a49-4354-a6e9-059592ee5ebc\") " pod="openstack/ceilometer-0" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.861028 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5f86be30-3a49-4354-a6e9-059592ee5ebc-run-httpd\") pod \"ceilometer-0\" (UID: \"5f86be30-3a49-4354-a6e9-059592ee5ebc\") " pod="openstack/ceilometer-0" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.861171 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f86be30-3a49-4354-a6e9-059592ee5ebc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5f86be30-3a49-4354-a6e9-059592ee5ebc\") " pod="openstack/ceilometer-0" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.861199 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f86be30-3a49-4354-a6e9-059592ee5ebc-config-data\") pod \"ceilometer-0\" (UID: \"5f86be30-3a49-4354-a6e9-059592ee5ebc\") " pod="openstack/ceilometer-0" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.861367 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ssj8v\" (UniqueName: \"kubernetes.io/projected/5f86be30-3a49-4354-a6e9-059592ee5ebc-kube-api-access-ssj8v\") pod \"ceilometer-0\" (UID: \"5f86be30-3a49-4354-a6e9-059592ee5ebc\") " pod="openstack/ceilometer-0" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.861434 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/156ae83a-3fe2-4947-8ae5-5e665f19325a-logs" (OuterVolumeSpecName: "logs") pod "156ae83a-3fe2-4947-8ae5-5e665f19325a" (UID: "156ae83a-3fe2-4947-8ae5-5e665f19325a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.861527 4716 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/156ae83a-3fe2-4947-8ae5-5e665f19325a-logs\") on node \"crc\" DevicePath \"\"" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.861796 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5f86be30-3a49-4354-a6e9-059592ee5ebc-log-httpd\") pod \"ceilometer-0\" (UID: \"5f86be30-3a49-4354-a6e9-059592ee5ebc\") " pod="openstack/ceilometer-0" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.861858 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5f86be30-3a49-4354-a6e9-059592ee5ebc-run-httpd\") pod \"ceilometer-0\" (UID: \"5f86be30-3a49-4354-a6e9-059592ee5ebc\") " pod="openstack/ceilometer-0" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.866313 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/156ae83a-3fe2-4947-8ae5-5e665f19325a-kube-api-access-jplqw" (OuterVolumeSpecName: "kube-api-access-jplqw") pod "156ae83a-3fe2-4947-8ae5-5e665f19325a" (UID: "156ae83a-3fe2-4947-8ae5-5e665f19325a"). InnerVolumeSpecName "kube-api-access-jplqw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.871746 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f86be30-3a49-4354-a6e9-059592ee5ebc-config-data\") pod \"ceilometer-0\" (UID: \"5f86be30-3a49-4354-a6e9-059592ee5ebc\") " pod="openstack/ceilometer-0" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.872328 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f86be30-3a49-4354-a6e9-059592ee5ebc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5f86be30-3a49-4354-a6e9-059592ee5ebc\") " pod="openstack/ceilometer-0" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.880906 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5f86be30-3a49-4354-a6e9-059592ee5ebc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5f86be30-3a49-4354-a6e9-059592ee5ebc\") " pod="openstack/ceilometer-0" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.881216 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/5f86be30-3a49-4354-a6e9-059592ee5ebc-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"5f86be30-3a49-4354-a6e9-059592ee5ebc\") " pod="openstack/ceilometer-0" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.881446 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5f86be30-3a49-4354-a6e9-059592ee5ebc-scripts\") pod \"ceilometer-0\" (UID: \"5f86be30-3a49-4354-a6e9-059592ee5ebc\") " pod="openstack/ceilometer-0" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.885322 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ssj8v\" (UniqueName: \"kubernetes.io/projected/5f86be30-3a49-4354-a6e9-059592ee5ebc-kube-api-access-ssj8v\") pod \"ceilometer-0\" (UID: \"5f86be30-3a49-4354-a6e9-059592ee5ebc\") " pod="openstack/ceilometer-0" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.913167 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/156ae83a-3fe2-4947-8ae5-5e665f19325a-config-data" (OuterVolumeSpecName: "config-data") pod "156ae83a-3fe2-4947-8ae5-5e665f19325a" (UID: "156ae83a-3fe2-4947-8ae5-5e665f19325a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.928799 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/156ae83a-3fe2-4947-8ae5-5e665f19325a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "156ae83a-3fe2-4947-8ae5-5e665f19325a" (UID: "156ae83a-3fe2-4947-8ae5-5e665f19325a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.961481 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.962975 4716 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/156ae83a-3fe2-4947-8ae5-5e665f19325a-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.962998 4716 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/156ae83a-3fe2-4947-8ae5-5e665f19325a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 16:21:53 crc kubenswrapper[4716]: I1207 16:21:53.963009 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jplqw\" (UniqueName: \"kubernetes.io/projected/156ae83a-3fe2-4947-8ae5-5e665f19325a-kube-api-access-jplqw\") on node \"crc\" DevicePath \"\"" Dec 07 16:21:54 crc kubenswrapper[4716]: I1207 16:21:54.272336 4716 generic.go:334] "Generic (PLEG): container finished" podID="156ae83a-3fe2-4947-8ae5-5e665f19325a" containerID="9e5593d595481b95bf77d0188463dd99be293fdb876a9b8eb29fbc597473f419" exitCode=0 Dec 07 16:21:54 crc kubenswrapper[4716]: I1207 16:21:54.272390 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"156ae83a-3fe2-4947-8ae5-5e665f19325a","Type":"ContainerDied","Data":"9e5593d595481b95bf77d0188463dd99be293fdb876a9b8eb29fbc597473f419"} Dec 07 16:21:54 crc kubenswrapper[4716]: I1207 16:21:54.272440 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"156ae83a-3fe2-4947-8ae5-5e665f19325a","Type":"ContainerDied","Data":"414c1d2ac8b747511a2e99394e45be9443459487e6fad6faba39dfb474d8fc7d"} Dec 07 16:21:54 crc kubenswrapper[4716]: I1207 16:21:54.272459 4716 scope.go:117] "RemoveContainer" containerID="9e5593d595481b95bf77d0188463dd99be293fdb876a9b8eb29fbc597473f419" Dec 07 16:21:54 crc kubenswrapper[4716]: I1207 16:21:54.274186 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 07 16:21:54 crc kubenswrapper[4716]: I1207 16:21:54.295179 4716 scope.go:117] "RemoveContainer" containerID="a57d656873ef4fdd9843cc133b204d2f76a4ad3bda5244b6df1d075a4c7bd148" Dec 07 16:21:54 crc kubenswrapper[4716]: I1207 16:21:54.314577 4716 scope.go:117] "RemoveContainer" containerID="9e5593d595481b95bf77d0188463dd99be293fdb876a9b8eb29fbc597473f419" Dec 07 16:21:54 crc kubenswrapper[4716]: E1207 16:21:54.315138 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e5593d595481b95bf77d0188463dd99be293fdb876a9b8eb29fbc597473f419\": container with ID starting with 9e5593d595481b95bf77d0188463dd99be293fdb876a9b8eb29fbc597473f419 not found: ID does not exist" containerID="9e5593d595481b95bf77d0188463dd99be293fdb876a9b8eb29fbc597473f419" Dec 07 16:21:54 crc kubenswrapper[4716]: I1207 16:21:54.315261 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e5593d595481b95bf77d0188463dd99be293fdb876a9b8eb29fbc597473f419"} err="failed to get container status \"9e5593d595481b95bf77d0188463dd99be293fdb876a9b8eb29fbc597473f419\": rpc error: code = NotFound desc = could not find container \"9e5593d595481b95bf77d0188463dd99be293fdb876a9b8eb29fbc597473f419\": container with ID starting with 9e5593d595481b95bf77d0188463dd99be293fdb876a9b8eb29fbc597473f419 not found: ID does not exist" Dec 07 16:21:54 crc kubenswrapper[4716]: I1207 16:21:54.315295 4716 scope.go:117] "RemoveContainer" containerID="a57d656873ef4fdd9843cc133b204d2f76a4ad3bda5244b6df1d075a4c7bd148" Dec 07 16:21:54 crc kubenswrapper[4716]: E1207 16:21:54.315795 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a57d656873ef4fdd9843cc133b204d2f76a4ad3bda5244b6df1d075a4c7bd148\": container with ID starting with a57d656873ef4fdd9843cc133b204d2f76a4ad3bda5244b6df1d075a4c7bd148 not found: ID does not exist" containerID="a57d656873ef4fdd9843cc133b204d2f76a4ad3bda5244b6df1d075a4c7bd148" Dec 07 16:21:54 crc kubenswrapper[4716]: I1207 16:21:54.315831 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a57d656873ef4fdd9843cc133b204d2f76a4ad3bda5244b6df1d075a4c7bd148"} err="failed to get container status \"a57d656873ef4fdd9843cc133b204d2f76a4ad3bda5244b6df1d075a4c7bd148\": rpc error: code = NotFound desc = could not find container \"a57d656873ef4fdd9843cc133b204d2f76a4ad3bda5244b6df1d075a4c7bd148\": container with ID starting with a57d656873ef4fdd9843cc133b204d2f76a4ad3bda5244b6df1d075a4c7bd148 not found: ID does not exist" Dec 07 16:21:54 crc kubenswrapper[4716]: I1207 16:21:54.333116 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 07 16:21:54 crc kubenswrapper[4716]: I1207 16:21:54.371418 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 07 16:21:54 crc kubenswrapper[4716]: I1207 16:21:54.392714 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 07 16:21:54 crc kubenswrapper[4716]: E1207 16:21:54.393450 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="156ae83a-3fe2-4947-8ae5-5e665f19325a" containerName="nova-api-api" Dec 07 16:21:54 crc kubenswrapper[4716]: I1207 16:21:54.393469 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="156ae83a-3fe2-4947-8ae5-5e665f19325a" containerName="nova-api-api" Dec 07 16:21:54 crc kubenswrapper[4716]: E1207 16:21:54.393488 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="156ae83a-3fe2-4947-8ae5-5e665f19325a" containerName="nova-api-log" Dec 07 16:21:54 crc kubenswrapper[4716]: I1207 16:21:54.393495 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="156ae83a-3fe2-4947-8ae5-5e665f19325a" containerName="nova-api-log" Dec 07 16:21:54 crc kubenswrapper[4716]: I1207 16:21:54.393807 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="156ae83a-3fe2-4947-8ae5-5e665f19325a" containerName="nova-api-log" Dec 07 16:21:54 crc kubenswrapper[4716]: I1207 16:21:54.393821 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="156ae83a-3fe2-4947-8ae5-5e665f19325a" containerName="nova-api-api" Dec 07 16:21:54 crc kubenswrapper[4716]: I1207 16:21:54.400064 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 07 16:21:54 crc kubenswrapper[4716]: I1207 16:21:54.402381 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 07 16:21:54 crc kubenswrapper[4716]: I1207 16:21:54.402905 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 07 16:21:54 crc kubenswrapper[4716]: I1207 16:21:54.405136 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 07 16:21:54 crc kubenswrapper[4716]: I1207 16:21:54.413928 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 07 16:21:54 crc kubenswrapper[4716]: I1207 16:21:54.427342 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 07 16:21:54 crc kubenswrapper[4716]: E1207 16:21:54.461681 4716 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod156ae83a_3fe2_4947_8ae5_5e665f19325a.slice/crio-414c1d2ac8b747511a2e99394e45be9443459487e6fad6faba39dfb474d8fc7d\": RecentStats: unable to find data in memory cache]" Dec 07 16:21:54 crc kubenswrapper[4716]: I1207 16:21:54.578454 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1-config-data\") pod \"nova-api-0\" (UID: \"19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1\") " pod="openstack/nova-api-0" Dec 07 16:21:54 crc kubenswrapper[4716]: I1207 16:21:54.578510 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1-logs\") pod \"nova-api-0\" (UID: \"19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1\") " pod="openstack/nova-api-0" Dec 07 16:21:54 crc kubenswrapper[4716]: I1207 16:21:54.578539 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1-internal-tls-certs\") pod \"nova-api-0\" (UID: \"19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1\") " pod="openstack/nova-api-0" Dec 07 16:21:54 crc kubenswrapper[4716]: I1207 16:21:54.578635 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1-public-tls-certs\") pod \"nova-api-0\" (UID: \"19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1\") " pod="openstack/nova-api-0" Dec 07 16:21:54 crc kubenswrapper[4716]: I1207 16:21:54.578657 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmt72\" (UniqueName: \"kubernetes.io/projected/19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1-kube-api-access-qmt72\") pod \"nova-api-0\" (UID: \"19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1\") " pod="openstack/nova-api-0" Dec 07 16:21:54 crc kubenswrapper[4716]: I1207 16:21:54.578676 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1\") " pod="openstack/nova-api-0" Dec 07 16:21:54 crc kubenswrapper[4716]: I1207 16:21:54.680796 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1-config-data\") pod \"nova-api-0\" (UID: \"19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1\") " pod="openstack/nova-api-0" Dec 07 16:21:54 crc kubenswrapper[4716]: I1207 16:21:54.680840 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1-logs\") pod \"nova-api-0\" (UID: \"19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1\") " pod="openstack/nova-api-0" Dec 07 16:21:54 crc kubenswrapper[4716]: I1207 16:21:54.680867 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1-internal-tls-certs\") pod \"nova-api-0\" (UID: \"19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1\") " pod="openstack/nova-api-0" Dec 07 16:21:54 crc kubenswrapper[4716]: I1207 16:21:54.680963 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1-public-tls-certs\") pod \"nova-api-0\" (UID: \"19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1\") " pod="openstack/nova-api-0" Dec 07 16:21:54 crc kubenswrapper[4716]: I1207 16:21:54.680987 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmt72\" (UniqueName: \"kubernetes.io/projected/19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1-kube-api-access-qmt72\") pod \"nova-api-0\" (UID: \"19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1\") " pod="openstack/nova-api-0" Dec 07 16:21:54 crc kubenswrapper[4716]: I1207 16:21:54.681009 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1\") " pod="openstack/nova-api-0" Dec 07 16:21:54 crc kubenswrapper[4716]: I1207 16:21:54.682183 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1-logs\") pod \"nova-api-0\" (UID: \"19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1\") " pod="openstack/nova-api-0" Dec 07 16:21:54 crc kubenswrapper[4716]: I1207 16:21:54.686488 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1\") " pod="openstack/nova-api-0" Dec 07 16:21:54 crc kubenswrapper[4716]: I1207 16:21:54.686980 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1-internal-tls-certs\") pod \"nova-api-0\" (UID: \"19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1\") " pod="openstack/nova-api-0" Dec 07 16:21:54 crc kubenswrapper[4716]: I1207 16:21:54.688024 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1-public-tls-certs\") pod \"nova-api-0\" (UID: \"19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1\") " pod="openstack/nova-api-0" Dec 07 16:21:54 crc kubenswrapper[4716]: I1207 16:21:54.691666 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1-config-data\") pod \"nova-api-0\" (UID: \"19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1\") " pod="openstack/nova-api-0" Dec 07 16:21:54 crc kubenswrapper[4716]: I1207 16:21:54.698280 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmt72\" (UniqueName: \"kubernetes.io/projected/19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1-kube-api-access-qmt72\") pod \"nova-api-0\" (UID: \"19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1\") " pod="openstack/nova-api-0" Dec 07 16:21:54 crc kubenswrapper[4716]: I1207 16:21:54.775979 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 07 16:21:55 crc kubenswrapper[4716]: I1207 16:21:55.207004 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 07 16:21:55 crc kubenswrapper[4716]: I1207 16:21:55.286042 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5f86be30-3a49-4354-a6e9-059592ee5ebc","Type":"ContainerStarted","Data":"d24b33a7d55111309aeb3fe887e212e44d9d06aed55fd83426146cf16d1f85bd"} Dec 07 16:21:55 crc kubenswrapper[4716]: I1207 16:21:55.286107 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5f86be30-3a49-4354-a6e9-059592ee5ebc","Type":"ContainerStarted","Data":"e73349c4522daa7c2350993c2a0631a1d399446be0edf9de13b23661772a5682"} Dec 07 16:21:55 crc kubenswrapper[4716]: I1207 16:21:55.287807 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1","Type":"ContainerStarted","Data":"607d5d91dea0f6ca5a576bf2e1c57bdfa42ce778f94f5b70df6c42aba2c00017"} Dec 07 16:21:55 crc kubenswrapper[4716]: I1207 16:21:55.671452 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="156ae83a-3fe2-4947-8ae5-5e665f19325a" path="/var/lib/kubelet/pods/156ae83a-3fe2-4947-8ae5-5e665f19325a/volumes" Dec 07 16:21:56 crc kubenswrapper[4716]: I1207 16:21:56.322656 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5f86be30-3a49-4354-a6e9-059592ee5ebc","Type":"ContainerStarted","Data":"b793b2d0ca19ae182e81ee764a2010c3c9a555ef59844fd853ff096a00c777d2"} Dec 07 16:21:56 crc kubenswrapper[4716]: I1207 16:21:56.325406 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1","Type":"ContainerStarted","Data":"448533af007b033c0997dd7c1efbf8576f5ef9f35642d4b1d15ed2db601ed66d"} Dec 07 16:21:56 crc kubenswrapper[4716]: I1207 16:21:56.325455 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1","Type":"ContainerStarted","Data":"f9c628f1f05b2f3810e5993338e5e68c94669cc9af38b9e58a18e886cb3108f1"} Dec 07 16:21:56 crc kubenswrapper[4716]: I1207 16:21:56.355580 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.355557669 podStartE2EDuration="2.355557669s" podCreationTimestamp="2025-12-07 16:21:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:21:56.352304431 +0000 UTC m=+1179.042589343" watchObservedRunningTime="2025-12-07 16:21:56.355557669 +0000 UTC m=+1179.045842591" Dec 07 16:21:56 crc kubenswrapper[4716]: I1207 16:21:56.669240 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 07 16:21:57 crc kubenswrapper[4716]: I1207 16:21:57.338756 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5f86be30-3a49-4354-a6e9-059592ee5ebc","Type":"ContainerStarted","Data":"4f433226df0c5be8772213be23c644d9e270f1e59f1114903e2a9a66604a21ae"} Dec 07 16:21:57 crc kubenswrapper[4716]: I1207 16:21:57.932304 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-89c5cd4d5-ld7pc" Dec 07 16:21:58 crc kubenswrapper[4716]: I1207 16:21:58.014818 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-v4tqc"] Dec 07 16:21:58 crc kubenswrapper[4716]: I1207 16:21:58.015090 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-757b4f8459-v4tqc" podUID="37636129-36c0-4531-9a84-8b1d177a3e44" containerName="dnsmasq-dns" containerID="cri-o://f29b859032b292ca3ab15aecab24a7a3d23e155abccef6904cc68386cf80b02d" gracePeriod=10 Dec 07 16:21:58 crc kubenswrapper[4716]: I1207 16:21:58.346500 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5f86be30-3a49-4354-a6e9-059592ee5ebc","Type":"ContainerStarted","Data":"c1d5d4fb68c704bd5e3e59fdc9b8878adff7087c7ea985b0bfe1f6d033a22754"} Dec 07 16:21:58 crc kubenswrapper[4716]: I1207 16:21:58.346947 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 07 16:21:58 crc kubenswrapper[4716]: I1207 16:21:58.348381 4716 generic.go:334] "Generic (PLEG): container finished" podID="37636129-36c0-4531-9a84-8b1d177a3e44" containerID="f29b859032b292ca3ab15aecab24a7a3d23e155abccef6904cc68386cf80b02d" exitCode=0 Dec 07 16:21:58 crc kubenswrapper[4716]: I1207 16:21:58.348424 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-v4tqc" event={"ID":"37636129-36c0-4531-9a84-8b1d177a3e44","Type":"ContainerDied","Data":"f29b859032b292ca3ab15aecab24a7a3d23e155abccef6904cc68386cf80b02d"} Dec 07 16:21:58 crc kubenswrapper[4716]: I1207 16:21:58.379241 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.067448474 podStartE2EDuration="5.37921963s" podCreationTimestamp="2025-12-07 16:21:53 +0000 UTC" firstStartedPulling="2025-12-07 16:21:54.40536755 +0000 UTC m=+1177.095652462" lastFinishedPulling="2025-12-07 16:21:57.717138666 +0000 UTC m=+1180.407423618" observedRunningTime="2025-12-07 16:21:58.364940625 +0000 UTC m=+1181.055225537" watchObservedRunningTime="2025-12-07 16:21:58.37921963 +0000 UTC m=+1181.069504532" Dec 07 16:21:58 crc kubenswrapper[4716]: I1207 16:21:58.516269 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-v4tqc" Dec 07 16:21:58 crc kubenswrapper[4716]: I1207 16:21:58.678609 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/37636129-36c0-4531-9a84-8b1d177a3e44-dns-swift-storage-0\") pod \"37636129-36c0-4531-9a84-8b1d177a3e44\" (UID: \"37636129-36c0-4531-9a84-8b1d177a3e44\") " Dec 07 16:21:58 crc kubenswrapper[4716]: I1207 16:21:58.678779 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37636129-36c0-4531-9a84-8b1d177a3e44-config\") pod \"37636129-36c0-4531-9a84-8b1d177a3e44\" (UID: \"37636129-36c0-4531-9a84-8b1d177a3e44\") " Dec 07 16:21:58 crc kubenswrapper[4716]: I1207 16:21:58.678825 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/37636129-36c0-4531-9a84-8b1d177a3e44-ovsdbserver-sb\") pod \"37636129-36c0-4531-9a84-8b1d177a3e44\" (UID: \"37636129-36c0-4531-9a84-8b1d177a3e44\") " Dec 07 16:21:58 crc kubenswrapper[4716]: I1207 16:21:58.678889 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-brw67\" (UniqueName: \"kubernetes.io/projected/37636129-36c0-4531-9a84-8b1d177a3e44-kube-api-access-brw67\") pod \"37636129-36c0-4531-9a84-8b1d177a3e44\" (UID: \"37636129-36c0-4531-9a84-8b1d177a3e44\") " Dec 07 16:21:58 crc kubenswrapper[4716]: I1207 16:21:58.678937 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/37636129-36c0-4531-9a84-8b1d177a3e44-dns-svc\") pod \"37636129-36c0-4531-9a84-8b1d177a3e44\" (UID: \"37636129-36c0-4531-9a84-8b1d177a3e44\") " Dec 07 16:21:58 crc kubenswrapper[4716]: I1207 16:21:58.678959 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/37636129-36c0-4531-9a84-8b1d177a3e44-ovsdbserver-nb\") pod \"37636129-36c0-4531-9a84-8b1d177a3e44\" (UID: \"37636129-36c0-4531-9a84-8b1d177a3e44\") " Dec 07 16:21:58 crc kubenswrapper[4716]: I1207 16:21:58.688453 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37636129-36c0-4531-9a84-8b1d177a3e44-kube-api-access-brw67" (OuterVolumeSpecName: "kube-api-access-brw67") pod "37636129-36c0-4531-9a84-8b1d177a3e44" (UID: "37636129-36c0-4531-9a84-8b1d177a3e44"). InnerVolumeSpecName "kube-api-access-brw67". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:21:58 crc kubenswrapper[4716]: I1207 16:21:58.757045 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37636129-36c0-4531-9a84-8b1d177a3e44-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "37636129-36c0-4531-9a84-8b1d177a3e44" (UID: "37636129-36c0-4531-9a84-8b1d177a3e44"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:21:58 crc kubenswrapper[4716]: I1207 16:21:58.767286 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37636129-36c0-4531-9a84-8b1d177a3e44-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "37636129-36c0-4531-9a84-8b1d177a3e44" (UID: "37636129-36c0-4531-9a84-8b1d177a3e44"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:21:58 crc kubenswrapper[4716]: I1207 16:21:58.773134 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37636129-36c0-4531-9a84-8b1d177a3e44-config" (OuterVolumeSpecName: "config") pod "37636129-36c0-4531-9a84-8b1d177a3e44" (UID: "37636129-36c0-4531-9a84-8b1d177a3e44"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:21:58 crc kubenswrapper[4716]: I1207 16:21:58.779582 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37636129-36c0-4531-9a84-8b1d177a3e44-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "37636129-36c0-4531-9a84-8b1d177a3e44" (UID: "37636129-36c0-4531-9a84-8b1d177a3e44"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:21:58 crc kubenswrapper[4716]: I1207 16:21:58.781664 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-brw67\" (UniqueName: \"kubernetes.io/projected/37636129-36c0-4531-9a84-8b1d177a3e44-kube-api-access-brw67\") on node \"crc\" DevicePath \"\"" Dec 07 16:21:58 crc kubenswrapper[4716]: I1207 16:21:58.781699 4716 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/37636129-36c0-4531-9a84-8b1d177a3e44-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 07 16:21:58 crc kubenswrapper[4716]: I1207 16:21:58.781710 4716 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/37636129-36c0-4531-9a84-8b1d177a3e44-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 07 16:21:58 crc kubenswrapper[4716]: I1207 16:21:58.781721 4716 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/37636129-36c0-4531-9a84-8b1d177a3e44-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 07 16:21:58 crc kubenswrapper[4716]: I1207 16:21:58.781730 4716 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37636129-36c0-4531-9a84-8b1d177a3e44-config\") on node \"crc\" DevicePath \"\"" Dec 07 16:21:58 crc kubenswrapper[4716]: I1207 16:21:58.787884 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37636129-36c0-4531-9a84-8b1d177a3e44-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "37636129-36c0-4531-9a84-8b1d177a3e44" (UID: "37636129-36c0-4531-9a84-8b1d177a3e44"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:21:58 crc kubenswrapper[4716]: I1207 16:21:58.882900 4716 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/37636129-36c0-4531-9a84-8b1d177a3e44-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 07 16:21:59 crc kubenswrapper[4716]: I1207 16:21:59.359328 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-v4tqc" event={"ID":"37636129-36c0-4531-9a84-8b1d177a3e44","Type":"ContainerDied","Data":"76e0a50916b702e9b7f979a5c7ace73c30a7371585dcbf0fe578b859fd1c86f7"} Dec 07 16:21:59 crc kubenswrapper[4716]: I1207 16:21:59.359389 4716 scope.go:117] "RemoveContainer" containerID="f29b859032b292ca3ab15aecab24a7a3d23e155abccef6904cc68386cf80b02d" Dec 07 16:21:59 crc kubenswrapper[4716]: I1207 16:21:59.359497 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-v4tqc" Dec 07 16:21:59 crc kubenswrapper[4716]: I1207 16:21:59.399694 4716 scope.go:117] "RemoveContainer" containerID="3b7d6afaf096e65d78714050058efe812609c9f66136801dc763beb063a783b9" Dec 07 16:21:59 crc kubenswrapper[4716]: I1207 16:21:59.410671 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-v4tqc"] Dec 07 16:21:59 crc kubenswrapper[4716]: I1207 16:21:59.422691 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-v4tqc"] Dec 07 16:21:59 crc kubenswrapper[4716]: I1207 16:21:59.667145 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37636129-36c0-4531-9a84-8b1d177a3e44" path="/var/lib/kubelet/pods/37636129-36c0-4531-9a84-8b1d177a3e44/volumes" Dec 07 16:22:01 crc kubenswrapper[4716]: I1207 16:22:01.681340 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 07 16:22:01 crc kubenswrapper[4716]: I1207 16:22:01.696550 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Dec 07 16:22:02 crc kubenswrapper[4716]: I1207 16:22:02.413353 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Dec 07 16:22:02 crc kubenswrapper[4716]: I1207 16:22:02.633354 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-cbchp"] Dec 07 16:22:02 crc kubenswrapper[4716]: E1207 16:22:02.633970 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37636129-36c0-4531-9a84-8b1d177a3e44" containerName="init" Dec 07 16:22:02 crc kubenswrapper[4716]: I1207 16:22:02.634049 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="37636129-36c0-4531-9a84-8b1d177a3e44" containerName="init" Dec 07 16:22:02 crc kubenswrapper[4716]: E1207 16:22:02.634146 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37636129-36c0-4531-9a84-8b1d177a3e44" containerName="dnsmasq-dns" Dec 07 16:22:02 crc kubenswrapper[4716]: I1207 16:22:02.634204 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="37636129-36c0-4531-9a84-8b1d177a3e44" containerName="dnsmasq-dns" Dec 07 16:22:02 crc kubenswrapper[4716]: I1207 16:22:02.634464 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="37636129-36c0-4531-9a84-8b1d177a3e44" containerName="dnsmasq-dns" Dec 07 16:22:02 crc kubenswrapper[4716]: I1207 16:22:02.635170 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-cbchp" Dec 07 16:22:02 crc kubenswrapper[4716]: I1207 16:22:02.637105 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Dec 07 16:22:02 crc kubenswrapper[4716]: I1207 16:22:02.637512 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Dec 07 16:22:02 crc kubenswrapper[4716]: I1207 16:22:02.655811 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-cbchp"] Dec 07 16:22:02 crc kubenswrapper[4716]: I1207 16:22:02.760505 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aef53883-32c3-44d7-9d3e-dd15928778a9-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-cbchp\" (UID: \"aef53883-32c3-44d7-9d3e-dd15928778a9\") " pod="openstack/nova-cell1-cell-mapping-cbchp" Dec 07 16:22:02 crc kubenswrapper[4716]: I1207 16:22:02.760558 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aef53883-32c3-44d7-9d3e-dd15928778a9-scripts\") pod \"nova-cell1-cell-mapping-cbchp\" (UID: \"aef53883-32c3-44d7-9d3e-dd15928778a9\") " pod="openstack/nova-cell1-cell-mapping-cbchp" Dec 07 16:22:02 crc kubenswrapper[4716]: I1207 16:22:02.760622 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hv4wv\" (UniqueName: \"kubernetes.io/projected/aef53883-32c3-44d7-9d3e-dd15928778a9-kube-api-access-hv4wv\") pod \"nova-cell1-cell-mapping-cbchp\" (UID: \"aef53883-32c3-44d7-9d3e-dd15928778a9\") " pod="openstack/nova-cell1-cell-mapping-cbchp" Dec 07 16:22:02 crc kubenswrapper[4716]: I1207 16:22:02.760679 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aef53883-32c3-44d7-9d3e-dd15928778a9-config-data\") pod \"nova-cell1-cell-mapping-cbchp\" (UID: \"aef53883-32c3-44d7-9d3e-dd15928778a9\") " pod="openstack/nova-cell1-cell-mapping-cbchp" Dec 07 16:22:02 crc kubenswrapper[4716]: I1207 16:22:02.863380 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aef53883-32c3-44d7-9d3e-dd15928778a9-config-data\") pod \"nova-cell1-cell-mapping-cbchp\" (UID: \"aef53883-32c3-44d7-9d3e-dd15928778a9\") " pod="openstack/nova-cell1-cell-mapping-cbchp" Dec 07 16:22:02 crc kubenswrapper[4716]: I1207 16:22:02.863584 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aef53883-32c3-44d7-9d3e-dd15928778a9-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-cbchp\" (UID: \"aef53883-32c3-44d7-9d3e-dd15928778a9\") " pod="openstack/nova-cell1-cell-mapping-cbchp" Dec 07 16:22:02 crc kubenswrapper[4716]: I1207 16:22:02.863603 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aef53883-32c3-44d7-9d3e-dd15928778a9-scripts\") pod \"nova-cell1-cell-mapping-cbchp\" (UID: \"aef53883-32c3-44d7-9d3e-dd15928778a9\") " pod="openstack/nova-cell1-cell-mapping-cbchp" Dec 07 16:22:02 crc kubenswrapper[4716]: I1207 16:22:02.863630 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hv4wv\" (UniqueName: \"kubernetes.io/projected/aef53883-32c3-44d7-9d3e-dd15928778a9-kube-api-access-hv4wv\") pod \"nova-cell1-cell-mapping-cbchp\" (UID: \"aef53883-32c3-44d7-9d3e-dd15928778a9\") " pod="openstack/nova-cell1-cell-mapping-cbchp" Dec 07 16:22:02 crc kubenswrapper[4716]: I1207 16:22:02.870111 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aef53883-32c3-44d7-9d3e-dd15928778a9-scripts\") pod \"nova-cell1-cell-mapping-cbchp\" (UID: \"aef53883-32c3-44d7-9d3e-dd15928778a9\") " pod="openstack/nova-cell1-cell-mapping-cbchp" Dec 07 16:22:02 crc kubenswrapper[4716]: I1207 16:22:02.872640 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aef53883-32c3-44d7-9d3e-dd15928778a9-config-data\") pod \"nova-cell1-cell-mapping-cbchp\" (UID: \"aef53883-32c3-44d7-9d3e-dd15928778a9\") " pod="openstack/nova-cell1-cell-mapping-cbchp" Dec 07 16:22:02 crc kubenswrapper[4716]: I1207 16:22:02.880429 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hv4wv\" (UniqueName: \"kubernetes.io/projected/aef53883-32c3-44d7-9d3e-dd15928778a9-kube-api-access-hv4wv\") pod \"nova-cell1-cell-mapping-cbchp\" (UID: \"aef53883-32c3-44d7-9d3e-dd15928778a9\") " pod="openstack/nova-cell1-cell-mapping-cbchp" Dec 07 16:22:02 crc kubenswrapper[4716]: I1207 16:22:02.880472 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aef53883-32c3-44d7-9d3e-dd15928778a9-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-cbchp\" (UID: \"aef53883-32c3-44d7-9d3e-dd15928778a9\") " pod="openstack/nova-cell1-cell-mapping-cbchp" Dec 07 16:22:02 crc kubenswrapper[4716]: I1207 16:22:02.963070 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-cbchp" Dec 07 16:22:03 crc kubenswrapper[4716]: I1207 16:22:03.430176 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-cbchp"] Dec 07 16:22:03 crc kubenswrapper[4716]: W1207 16:22:03.439402 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaef53883_32c3_44d7_9d3e_dd15928778a9.slice/crio-c1b0adb198325fa7a35002b38c5538ef61fce6b06d89827c11e709beceb96077 WatchSource:0}: Error finding container c1b0adb198325fa7a35002b38c5538ef61fce6b06d89827c11e709beceb96077: Status 404 returned error can't find the container with id c1b0adb198325fa7a35002b38c5538ef61fce6b06d89827c11e709beceb96077 Dec 07 16:22:04 crc kubenswrapper[4716]: I1207 16:22:04.418993 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-cbchp" event={"ID":"aef53883-32c3-44d7-9d3e-dd15928778a9","Type":"ContainerStarted","Data":"34dfffe27670489ce32ad670cf9fbdfa109a055adf25d21c955c2e0830783d35"} Dec 07 16:22:04 crc kubenswrapper[4716]: I1207 16:22:04.419417 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-cbchp" event={"ID":"aef53883-32c3-44d7-9d3e-dd15928778a9","Type":"ContainerStarted","Data":"c1b0adb198325fa7a35002b38c5538ef61fce6b06d89827c11e709beceb96077"} Dec 07 16:22:04 crc kubenswrapper[4716]: I1207 16:22:04.445708 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-cbchp" podStartSLOduration=2.445691669 podStartE2EDuration="2.445691669s" podCreationTimestamp="2025-12-07 16:22:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:22:04.435606967 +0000 UTC m=+1187.125891909" watchObservedRunningTime="2025-12-07 16:22:04.445691669 +0000 UTC m=+1187.135976581" Dec 07 16:22:04 crc kubenswrapper[4716]: I1207 16:22:04.777749 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 07 16:22:04 crc kubenswrapper[4716]: I1207 16:22:04.777850 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 07 16:22:05 crc kubenswrapper[4716]: I1207 16:22:05.791222 4716 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.199:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 07 16:22:05 crc kubenswrapper[4716]: I1207 16:22:05.791271 4716 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.199:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 07 16:22:08 crc kubenswrapper[4716]: I1207 16:22:08.465861 4716 generic.go:334] "Generic (PLEG): container finished" podID="aef53883-32c3-44d7-9d3e-dd15928778a9" containerID="34dfffe27670489ce32ad670cf9fbdfa109a055adf25d21c955c2e0830783d35" exitCode=0 Dec 07 16:22:08 crc kubenswrapper[4716]: I1207 16:22:08.465942 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-cbchp" event={"ID":"aef53883-32c3-44d7-9d3e-dd15928778a9","Type":"ContainerDied","Data":"34dfffe27670489ce32ad670cf9fbdfa109a055adf25d21c955c2e0830783d35"} Dec 07 16:22:09 crc kubenswrapper[4716]: I1207 16:22:09.922007 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-cbchp" Dec 07 16:22:10 crc kubenswrapper[4716]: I1207 16:22:10.004837 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aef53883-32c3-44d7-9d3e-dd15928778a9-scripts\") pod \"aef53883-32c3-44d7-9d3e-dd15928778a9\" (UID: \"aef53883-32c3-44d7-9d3e-dd15928778a9\") " Dec 07 16:22:10 crc kubenswrapper[4716]: I1207 16:22:10.005090 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aef53883-32c3-44d7-9d3e-dd15928778a9-config-data\") pod \"aef53883-32c3-44d7-9d3e-dd15928778a9\" (UID: \"aef53883-32c3-44d7-9d3e-dd15928778a9\") " Dec 07 16:22:10 crc kubenswrapper[4716]: I1207 16:22:10.005237 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aef53883-32c3-44d7-9d3e-dd15928778a9-combined-ca-bundle\") pod \"aef53883-32c3-44d7-9d3e-dd15928778a9\" (UID: \"aef53883-32c3-44d7-9d3e-dd15928778a9\") " Dec 07 16:22:10 crc kubenswrapper[4716]: I1207 16:22:10.005317 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hv4wv\" (UniqueName: \"kubernetes.io/projected/aef53883-32c3-44d7-9d3e-dd15928778a9-kube-api-access-hv4wv\") pod \"aef53883-32c3-44d7-9d3e-dd15928778a9\" (UID: \"aef53883-32c3-44d7-9d3e-dd15928778a9\") " Dec 07 16:22:10 crc kubenswrapper[4716]: I1207 16:22:10.010309 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aef53883-32c3-44d7-9d3e-dd15928778a9-scripts" (OuterVolumeSpecName: "scripts") pod "aef53883-32c3-44d7-9d3e-dd15928778a9" (UID: "aef53883-32c3-44d7-9d3e-dd15928778a9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:22:10 crc kubenswrapper[4716]: I1207 16:22:10.010937 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aef53883-32c3-44d7-9d3e-dd15928778a9-kube-api-access-hv4wv" (OuterVolumeSpecName: "kube-api-access-hv4wv") pod "aef53883-32c3-44d7-9d3e-dd15928778a9" (UID: "aef53883-32c3-44d7-9d3e-dd15928778a9"). InnerVolumeSpecName "kube-api-access-hv4wv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:22:10 crc kubenswrapper[4716]: I1207 16:22:10.032825 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aef53883-32c3-44d7-9d3e-dd15928778a9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "aef53883-32c3-44d7-9d3e-dd15928778a9" (UID: "aef53883-32c3-44d7-9d3e-dd15928778a9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:22:10 crc kubenswrapper[4716]: I1207 16:22:10.033408 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aef53883-32c3-44d7-9d3e-dd15928778a9-config-data" (OuterVolumeSpecName: "config-data") pod "aef53883-32c3-44d7-9d3e-dd15928778a9" (UID: "aef53883-32c3-44d7-9d3e-dd15928778a9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:22:10 crc kubenswrapper[4716]: I1207 16:22:10.107971 4716 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aef53883-32c3-44d7-9d3e-dd15928778a9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 16:22:10 crc kubenswrapper[4716]: I1207 16:22:10.108005 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hv4wv\" (UniqueName: \"kubernetes.io/projected/aef53883-32c3-44d7-9d3e-dd15928778a9-kube-api-access-hv4wv\") on node \"crc\" DevicePath \"\"" Dec 07 16:22:10 crc kubenswrapper[4716]: I1207 16:22:10.108014 4716 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aef53883-32c3-44d7-9d3e-dd15928778a9-scripts\") on node \"crc\" DevicePath \"\"" Dec 07 16:22:10 crc kubenswrapper[4716]: I1207 16:22:10.108023 4716 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aef53883-32c3-44d7-9d3e-dd15928778a9-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 16:22:10 crc kubenswrapper[4716]: I1207 16:22:10.498369 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-cbchp" event={"ID":"aef53883-32c3-44d7-9d3e-dd15928778a9","Type":"ContainerDied","Data":"c1b0adb198325fa7a35002b38c5538ef61fce6b06d89827c11e709beceb96077"} Dec 07 16:22:10 crc kubenswrapper[4716]: I1207 16:22:10.498448 4716 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c1b0adb198325fa7a35002b38c5538ef61fce6b06d89827c11e709beceb96077" Dec 07 16:22:10 crc kubenswrapper[4716]: I1207 16:22:10.498548 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-cbchp" Dec 07 16:22:10 crc kubenswrapper[4716]: I1207 16:22:10.697008 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 07 16:22:10 crc kubenswrapper[4716]: I1207 16:22:10.697800 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1" containerName="nova-api-log" containerID="cri-o://f9c628f1f05b2f3810e5993338e5e68c94669cc9af38b9e58a18e886cb3108f1" gracePeriod=30 Dec 07 16:22:10 crc kubenswrapper[4716]: I1207 16:22:10.697878 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1" containerName="nova-api-api" containerID="cri-o://448533af007b033c0997dd7c1efbf8576f5ef9f35642d4b1d15ed2db601ed66d" gracePeriod=30 Dec 07 16:22:10 crc kubenswrapper[4716]: I1207 16:22:10.718114 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 07 16:22:10 crc kubenswrapper[4716]: I1207 16:22:10.718362 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="ed6ed6ec-40bd-4628-91c8-2410b68f7384" containerName="nova-scheduler-scheduler" containerID="cri-o://57fcac8ed81c947a24d209d04c59bfeb69b36b49d18fa729d309902f2e7b5e08" gracePeriod=30 Dec 07 16:22:10 crc kubenswrapper[4716]: I1207 16:22:10.738136 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 07 16:22:10 crc kubenswrapper[4716]: I1207 16:22:10.739308 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="fdcb1fee-b3e5-4ed1-b868-875220d969be" containerName="nova-metadata-metadata" containerID="cri-o://906d4a87c6ec35f764939dacb31e576c109400d4b65951469b0bde9c6098df07" gracePeriod=30 Dec 07 16:22:10 crc kubenswrapper[4716]: I1207 16:22:10.739037 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="fdcb1fee-b3e5-4ed1-b868-875220d969be" containerName="nova-metadata-log" containerID="cri-o://ef9a5cfc9371486feec9d5c023ed78c6fd30243c2a69c617f4dcb9287c1dac96" gracePeriod=30 Dec 07 16:22:11 crc kubenswrapper[4716]: I1207 16:22:11.516361 4716 generic.go:334] "Generic (PLEG): container finished" podID="19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1" containerID="f9c628f1f05b2f3810e5993338e5e68c94669cc9af38b9e58a18e886cb3108f1" exitCode=143 Dec 07 16:22:11 crc kubenswrapper[4716]: I1207 16:22:11.516411 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1","Type":"ContainerDied","Data":"f9c628f1f05b2f3810e5993338e5e68c94669cc9af38b9e58a18e886cb3108f1"} Dec 07 16:22:11 crc kubenswrapper[4716]: I1207 16:22:11.520935 4716 generic.go:334] "Generic (PLEG): container finished" podID="fdcb1fee-b3e5-4ed1-b868-875220d969be" containerID="ef9a5cfc9371486feec9d5c023ed78c6fd30243c2a69c617f4dcb9287c1dac96" exitCode=143 Dec 07 16:22:11 crc kubenswrapper[4716]: I1207 16:22:11.520982 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fdcb1fee-b3e5-4ed1-b868-875220d969be","Type":"ContainerDied","Data":"ef9a5cfc9371486feec9d5c023ed78c6fd30243c2a69c617f4dcb9287c1dac96"} Dec 07 16:22:12 crc kubenswrapper[4716]: E1207 16:22:12.500307 4716 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="57fcac8ed81c947a24d209d04c59bfeb69b36b49d18fa729d309902f2e7b5e08" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 07 16:22:12 crc kubenswrapper[4716]: E1207 16:22:12.503607 4716 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="57fcac8ed81c947a24d209d04c59bfeb69b36b49d18fa729d309902f2e7b5e08" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 07 16:22:12 crc kubenswrapper[4716]: E1207 16:22:12.506029 4716 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="57fcac8ed81c947a24d209d04c59bfeb69b36b49d18fa729d309902f2e7b5e08" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 07 16:22:12 crc kubenswrapper[4716]: E1207 16:22:12.506155 4716 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="ed6ed6ec-40bd-4628-91c8-2410b68f7384" containerName="nova-scheduler-scheduler" Dec 07 16:22:13 crc kubenswrapper[4716]: I1207 16:22:13.882595 4716 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="fdcb1fee-b3e5-4ed1-b868-875220d969be" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.195:8775/\": read tcp 10.217.0.2:47294->10.217.0.195:8775: read: connection reset by peer" Dec 07 16:22:13 crc kubenswrapper[4716]: I1207 16:22:13.883222 4716 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="fdcb1fee-b3e5-4ed1-b868-875220d969be" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.195:8775/\": read tcp 10.217.0.2:47308->10.217.0.195:8775: read: connection reset by peer" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.299204 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.380708 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.400891 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1-config-data\") pod \"19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1\" (UID: \"19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1\") " Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.400949 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1-logs\") pod \"19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1\" (UID: \"19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1\") " Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.401028 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1-public-tls-certs\") pod \"19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1\" (UID: \"19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1\") " Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.401065 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1-internal-tls-certs\") pod \"19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1\" (UID: \"19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1\") " Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.401202 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qmt72\" (UniqueName: \"kubernetes.io/projected/19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1-kube-api-access-qmt72\") pod \"19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1\" (UID: \"19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1\") " Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.401318 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1-combined-ca-bundle\") pod \"19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1\" (UID: \"19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1\") " Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.402094 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1-logs" (OuterVolumeSpecName: "logs") pod "19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1" (UID: "19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.402598 4716 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1-logs\") on node \"crc\" DevicePath \"\"" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.412749 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1-kube-api-access-qmt72" (OuterVolumeSpecName: "kube-api-access-qmt72") pod "19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1" (UID: "19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1"). InnerVolumeSpecName "kube-api-access-qmt72". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.441752 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1" (UID: "19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.463313 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1-config-data" (OuterVolumeSpecName: "config-data") pod "19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1" (UID: "19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.464164 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1" (UID: "19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.478319 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1" (UID: "19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.504135 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/fdcb1fee-b3e5-4ed1-b868-875220d969be-nova-metadata-tls-certs\") pod \"fdcb1fee-b3e5-4ed1-b868-875220d969be\" (UID: \"fdcb1fee-b3e5-4ed1-b868-875220d969be\") " Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.504320 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fdcb1fee-b3e5-4ed1-b868-875220d969be-logs\") pod \"fdcb1fee-b3e5-4ed1-b868-875220d969be\" (UID: \"fdcb1fee-b3e5-4ed1-b868-875220d969be\") " Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.504352 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdcb1fee-b3e5-4ed1-b868-875220d969be-combined-ca-bundle\") pod \"fdcb1fee-b3e5-4ed1-b868-875220d969be\" (UID: \"fdcb1fee-b3e5-4ed1-b868-875220d969be\") " Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.504492 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4jqpm\" (UniqueName: \"kubernetes.io/projected/fdcb1fee-b3e5-4ed1-b868-875220d969be-kube-api-access-4jqpm\") pod \"fdcb1fee-b3e5-4ed1-b868-875220d969be\" (UID: \"fdcb1fee-b3e5-4ed1-b868-875220d969be\") " Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.504538 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fdcb1fee-b3e5-4ed1-b868-875220d969be-config-data\") pod \"fdcb1fee-b3e5-4ed1-b868-875220d969be\" (UID: \"fdcb1fee-b3e5-4ed1-b868-875220d969be\") " Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.505056 4716 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.505096 4716 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.505114 4716 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.505126 4716 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.505139 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qmt72\" (UniqueName: \"kubernetes.io/projected/19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1-kube-api-access-qmt72\") on node \"crc\" DevicePath \"\"" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.505182 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fdcb1fee-b3e5-4ed1-b868-875220d969be-logs" (OuterVolumeSpecName: "logs") pod "fdcb1fee-b3e5-4ed1-b868-875220d969be" (UID: "fdcb1fee-b3e5-4ed1-b868-875220d969be"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.508329 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fdcb1fee-b3e5-4ed1-b868-875220d969be-kube-api-access-4jqpm" (OuterVolumeSpecName: "kube-api-access-4jqpm") pod "fdcb1fee-b3e5-4ed1-b868-875220d969be" (UID: "fdcb1fee-b3e5-4ed1-b868-875220d969be"). InnerVolumeSpecName "kube-api-access-4jqpm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.529547 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fdcb1fee-b3e5-4ed1-b868-875220d969be-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fdcb1fee-b3e5-4ed1-b868-875220d969be" (UID: "fdcb1fee-b3e5-4ed1-b868-875220d969be"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.541317 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fdcb1fee-b3e5-4ed1-b868-875220d969be-config-data" (OuterVolumeSpecName: "config-data") pod "fdcb1fee-b3e5-4ed1-b868-875220d969be" (UID: "fdcb1fee-b3e5-4ed1-b868-875220d969be"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.564869 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fdcb1fee-b3e5-4ed1-b868-875220d969be-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "fdcb1fee-b3e5-4ed1-b868-875220d969be" (UID: "fdcb1fee-b3e5-4ed1-b868-875220d969be"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.569496 4716 generic.go:334] "Generic (PLEG): container finished" podID="fdcb1fee-b3e5-4ed1-b868-875220d969be" containerID="906d4a87c6ec35f764939dacb31e576c109400d4b65951469b0bde9c6098df07" exitCode=0 Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.569573 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.569600 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fdcb1fee-b3e5-4ed1-b868-875220d969be","Type":"ContainerDied","Data":"906d4a87c6ec35f764939dacb31e576c109400d4b65951469b0bde9c6098df07"} Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.569658 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fdcb1fee-b3e5-4ed1-b868-875220d969be","Type":"ContainerDied","Data":"8888e46d76c01eb12230bcdefa7ef3662a8bdf028b1170cfe2cc088ffd8fe61b"} Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.569680 4716 scope.go:117] "RemoveContainer" containerID="906d4a87c6ec35f764939dacb31e576c109400d4b65951469b0bde9c6098df07" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.573822 4716 generic.go:334] "Generic (PLEG): container finished" podID="19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1" containerID="448533af007b033c0997dd7c1efbf8576f5ef9f35642d4b1d15ed2db601ed66d" exitCode=0 Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.573866 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1","Type":"ContainerDied","Data":"448533af007b033c0997dd7c1efbf8576f5ef9f35642d4b1d15ed2db601ed66d"} Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.573886 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.573894 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1","Type":"ContainerDied","Data":"607d5d91dea0f6ca5a576bf2e1c57bdfa42ce778f94f5b70df6c42aba2c00017"} Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.621002 4716 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/fdcb1fee-b3e5-4ed1-b868-875220d969be-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.621053 4716 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fdcb1fee-b3e5-4ed1-b868-875220d969be-logs\") on node \"crc\" DevicePath \"\"" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.621072 4716 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdcb1fee-b3e5-4ed1-b868-875220d969be-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.621268 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4jqpm\" (UniqueName: \"kubernetes.io/projected/fdcb1fee-b3e5-4ed1-b868-875220d969be-kube-api-access-4jqpm\") on node \"crc\" DevicePath \"\"" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.621287 4716 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fdcb1fee-b3e5-4ed1-b868-875220d969be-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.622239 4716 scope.go:117] "RemoveContainer" containerID="ef9a5cfc9371486feec9d5c023ed78c6fd30243c2a69c617f4dcb9287c1dac96" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.626895 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.642389 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.649906 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.659413 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.668467 4716 scope.go:117] "RemoveContainer" containerID="906d4a87c6ec35f764939dacb31e576c109400d4b65951469b0bde9c6098df07" Dec 07 16:22:14 crc kubenswrapper[4716]: E1207 16:22:14.668996 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"906d4a87c6ec35f764939dacb31e576c109400d4b65951469b0bde9c6098df07\": container with ID starting with 906d4a87c6ec35f764939dacb31e576c109400d4b65951469b0bde9c6098df07 not found: ID does not exist" containerID="906d4a87c6ec35f764939dacb31e576c109400d4b65951469b0bde9c6098df07" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.669048 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"906d4a87c6ec35f764939dacb31e576c109400d4b65951469b0bde9c6098df07"} err="failed to get container status \"906d4a87c6ec35f764939dacb31e576c109400d4b65951469b0bde9c6098df07\": rpc error: code = NotFound desc = could not find container \"906d4a87c6ec35f764939dacb31e576c109400d4b65951469b0bde9c6098df07\": container with ID starting with 906d4a87c6ec35f764939dacb31e576c109400d4b65951469b0bde9c6098df07 not found: ID does not exist" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.669103 4716 scope.go:117] "RemoveContainer" containerID="ef9a5cfc9371486feec9d5c023ed78c6fd30243c2a69c617f4dcb9287c1dac96" Dec 07 16:22:14 crc kubenswrapper[4716]: E1207 16:22:14.669664 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef9a5cfc9371486feec9d5c023ed78c6fd30243c2a69c617f4dcb9287c1dac96\": container with ID starting with ef9a5cfc9371486feec9d5c023ed78c6fd30243c2a69c617f4dcb9287c1dac96 not found: ID does not exist" containerID="ef9a5cfc9371486feec9d5c023ed78c6fd30243c2a69c617f4dcb9287c1dac96" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.669695 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef9a5cfc9371486feec9d5c023ed78c6fd30243c2a69c617f4dcb9287c1dac96"} err="failed to get container status \"ef9a5cfc9371486feec9d5c023ed78c6fd30243c2a69c617f4dcb9287c1dac96\": rpc error: code = NotFound desc = could not find container \"ef9a5cfc9371486feec9d5c023ed78c6fd30243c2a69c617f4dcb9287c1dac96\": container with ID starting with ef9a5cfc9371486feec9d5c023ed78c6fd30243c2a69c617f4dcb9287c1dac96 not found: ID does not exist" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.669716 4716 scope.go:117] "RemoveContainer" containerID="448533af007b033c0997dd7c1efbf8576f5ef9f35642d4b1d15ed2db601ed66d" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.678479 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 07 16:22:14 crc kubenswrapper[4716]: E1207 16:22:14.678889 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1" containerName="nova-api-api" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.678905 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1" containerName="nova-api-api" Dec 07 16:22:14 crc kubenswrapper[4716]: E1207 16:22:14.678921 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1" containerName="nova-api-log" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.678930 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1" containerName="nova-api-log" Dec 07 16:22:14 crc kubenswrapper[4716]: E1207 16:22:14.678944 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdcb1fee-b3e5-4ed1-b868-875220d969be" containerName="nova-metadata-log" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.678950 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdcb1fee-b3e5-4ed1-b868-875220d969be" containerName="nova-metadata-log" Dec 07 16:22:14 crc kubenswrapper[4716]: E1207 16:22:14.678964 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdcb1fee-b3e5-4ed1-b868-875220d969be" containerName="nova-metadata-metadata" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.678970 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdcb1fee-b3e5-4ed1-b868-875220d969be" containerName="nova-metadata-metadata" Dec 07 16:22:14 crc kubenswrapper[4716]: E1207 16:22:14.679001 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aef53883-32c3-44d7-9d3e-dd15928778a9" containerName="nova-manage" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.679006 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="aef53883-32c3-44d7-9d3e-dd15928778a9" containerName="nova-manage" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.679210 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="fdcb1fee-b3e5-4ed1-b868-875220d969be" containerName="nova-metadata-log" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.679229 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="fdcb1fee-b3e5-4ed1-b868-875220d969be" containerName="nova-metadata-metadata" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.679240 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1" containerName="nova-api-log" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.679257 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1" containerName="nova-api-api" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.679265 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="aef53883-32c3-44d7-9d3e-dd15928778a9" containerName="nova-manage" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.680365 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.692010 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.692396 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.692823 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.694410 4716 scope.go:117] "RemoveContainer" containerID="f9c628f1f05b2f3810e5993338e5e68c94669cc9af38b9e58a18e886cb3108f1" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.702468 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.713466 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.714881 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.718394 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.718611 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.726331 4716 scope.go:117] "RemoveContainer" containerID="448533af007b033c0997dd7c1efbf8576f5ef9f35642d4b1d15ed2db601ed66d" Dec 07 16:22:14 crc kubenswrapper[4716]: E1207 16:22:14.727351 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"448533af007b033c0997dd7c1efbf8576f5ef9f35642d4b1d15ed2db601ed66d\": container with ID starting with 448533af007b033c0997dd7c1efbf8576f5ef9f35642d4b1d15ed2db601ed66d not found: ID does not exist" containerID="448533af007b033c0997dd7c1efbf8576f5ef9f35642d4b1d15ed2db601ed66d" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.727398 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"448533af007b033c0997dd7c1efbf8576f5ef9f35642d4b1d15ed2db601ed66d"} err="failed to get container status \"448533af007b033c0997dd7c1efbf8576f5ef9f35642d4b1d15ed2db601ed66d\": rpc error: code = NotFound desc = could not find container \"448533af007b033c0997dd7c1efbf8576f5ef9f35642d4b1d15ed2db601ed66d\": container with ID starting with 448533af007b033c0997dd7c1efbf8576f5ef9f35642d4b1d15ed2db601ed66d not found: ID does not exist" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.727425 4716 scope.go:117] "RemoveContainer" containerID="f9c628f1f05b2f3810e5993338e5e68c94669cc9af38b9e58a18e886cb3108f1" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.736696 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 07 16:22:14 crc kubenswrapper[4716]: E1207 16:22:14.736712 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f9c628f1f05b2f3810e5993338e5e68c94669cc9af38b9e58a18e886cb3108f1\": container with ID starting with f9c628f1f05b2f3810e5993338e5e68c94669cc9af38b9e58a18e886cb3108f1 not found: ID does not exist" containerID="f9c628f1f05b2f3810e5993338e5e68c94669cc9af38b9e58a18e886cb3108f1" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.736770 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f9c628f1f05b2f3810e5993338e5e68c94669cc9af38b9e58a18e886cb3108f1"} err="failed to get container status \"f9c628f1f05b2f3810e5993338e5e68c94669cc9af38b9e58a18e886cb3108f1\": rpc error: code = NotFound desc = could not find container \"f9c628f1f05b2f3810e5993338e5e68c94669cc9af38b9e58a18e886cb3108f1\": container with ID starting with f9c628f1f05b2f3810e5993338e5e68c94669cc9af38b9e58a18e886cb3108f1 not found: ID does not exist" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.830029 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3953832e-aa6d-49f2-980d-e2308f0d1ece-config-data\") pod \"nova-api-0\" (UID: \"3953832e-aa6d-49f2-980d-e2308f0d1ece\") " pod="openstack/nova-api-0" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.830104 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3953832e-aa6d-49f2-980d-e2308f0d1ece-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3953832e-aa6d-49f2-980d-e2308f0d1ece\") " pod="openstack/nova-api-0" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.830159 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2ac894fd-e293-4f28-bdb5-e80e01601903-logs\") pod \"nova-metadata-0\" (UID: \"2ac894fd-e293-4f28-bdb5-e80e01601903\") " pod="openstack/nova-metadata-0" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.830211 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-282g2\" (UniqueName: \"kubernetes.io/projected/3953832e-aa6d-49f2-980d-e2308f0d1ece-kube-api-access-282g2\") pod \"nova-api-0\" (UID: \"3953832e-aa6d-49f2-980d-e2308f0d1ece\") " pod="openstack/nova-api-0" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.830237 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ac894fd-e293-4f28-bdb5-e80e01601903-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"2ac894fd-e293-4f28-bdb5-e80e01601903\") " pod="openstack/nova-metadata-0" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.830304 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3953832e-aa6d-49f2-980d-e2308f0d1ece-public-tls-certs\") pod \"nova-api-0\" (UID: \"3953832e-aa6d-49f2-980d-e2308f0d1ece\") " pod="openstack/nova-api-0" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.830358 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79qnv\" (UniqueName: \"kubernetes.io/projected/2ac894fd-e293-4f28-bdb5-e80e01601903-kube-api-access-79qnv\") pod \"nova-metadata-0\" (UID: \"2ac894fd-e293-4f28-bdb5-e80e01601903\") " pod="openstack/nova-metadata-0" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.830397 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3953832e-aa6d-49f2-980d-e2308f0d1ece-logs\") pod \"nova-api-0\" (UID: \"3953832e-aa6d-49f2-980d-e2308f0d1ece\") " pod="openstack/nova-api-0" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.830421 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3953832e-aa6d-49f2-980d-e2308f0d1ece-internal-tls-certs\") pod \"nova-api-0\" (UID: \"3953832e-aa6d-49f2-980d-e2308f0d1ece\") " pod="openstack/nova-api-0" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.830478 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ac894fd-e293-4f28-bdb5-e80e01601903-config-data\") pod \"nova-metadata-0\" (UID: \"2ac894fd-e293-4f28-bdb5-e80e01601903\") " pod="openstack/nova-metadata-0" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.830498 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ac894fd-e293-4f28-bdb5-e80e01601903-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2ac894fd-e293-4f28-bdb5-e80e01601903\") " pod="openstack/nova-metadata-0" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.931675 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3953832e-aa6d-49f2-980d-e2308f0d1ece-config-data\") pod \"nova-api-0\" (UID: \"3953832e-aa6d-49f2-980d-e2308f0d1ece\") " pod="openstack/nova-api-0" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.933204 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3953832e-aa6d-49f2-980d-e2308f0d1ece-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3953832e-aa6d-49f2-980d-e2308f0d1ece\") " pod="openstack/nova-api-0" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.933596 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2ac894fd-e293-4f28-bdb5-e80e01601903-logs\") pod \"nova-metadata-0\" (UID: \"2ac894fd-e293-4f28-bdb5-e80e01601903\") " pod="openstack/nova-metadata-0" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.933754 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-282g2\" (UniqueName: \"kubernetes.io/projected/3953832e-aa6d-49f2-980d-e2308f0d1ece-kube-api-access-282g2\") pod \"nova-api-0\" (UID: \"3953832e-aa6d-49f2-980d-e2308f0d1ece\") " pod="openstack/nova-api-0" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.933971 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ac894fd-e293-4f28-bdb5-e80e01601903-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"2ac894fd-e293-4f28-bdb5-e80e01601903\") " pod="openstack/nova-metadata-0" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.934116 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2ac894fd-e293-4f28-bdb5-e80e01601903-logs\") pod \"nova-metadata-0\" (UID: \"2ac894fd-e293-4f28-bdb5-e80e01601903\") " pod="openstack/nova-metadata-0" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.934106 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3953832e-aa6d-49f2-980d-e2308f0d1ece-public-tls-certs\") pod \"nova-api-0\" (UID: \"3953832e-aa6d-49f2-980d-e2308f0d1ece\") " pod="openstack/nova-api-0" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.936257 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79qnv\" (UniqueName: \"kubernetes.io/projected/2ac894fd-e293-4f28-bdb5-e80e01601903-kube-api-access-79qnv\") pod \"nova-metadata-0\" (UID: \"2ac894fd-e293-4f28-bdb5-e80e01601903\") " pod="openstack/nova-metadata-0" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.936333 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3953832e-aa6d-49f2-980d-e2308f0d1ece-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3953832e-aa6d-49f2-980d-e2308f0d1ece\") " pod="openstack/nova-api-0" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.936485 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3953832e-aa6d-49f2-980d-e2308f0d1ece-logs\") pod \"nova-api-0\" (UID: \"3953832e-aa6d-49f2-980d-e2308f0d1ece\") " pod="openstack/nova-api-0" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.937682 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3953832e-aa6d-49f2-980d-e2308f0d1ece-internal-tls-certs\") pod \"nova-api-0\" (UID: \"3953832e-aa6d-49f2-980d-e2308f0d1ece\") " pod="openstack/nova-api-0" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.936929 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3953832e-aa6d-49f2-980d-e2308f0d1ece-logs\") pod \"nova-api-0\" (UID: \"3953832e-aa6d-49f2-980d-e2308f0d1ece\") " pod="openstack/nova-api-0" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.937974 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ac894fd-e293-4f28-bdb5-e80e01601903-config-data\") pod \"nova-metadata-0\" (UID: \"2ac894fd-e293-4f28-bdb5-e80e01601903\") " pod="openstack/nova-metadata-0" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.938055 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ac894fd-e293-4f28-bdb5-e80e01601903-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2ac894fd-e293-4f28-bdb5-e80e01601903\") " pod="openstack/nova-metadata-0" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.938811 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3953832e-aa6d-49f2-980d-e2308f0d1ece-config-data\") pod \"nova-api-0\" (UID: \"3953832e-aa6d-49f2-980d-e2308f0d1ece\") " pod="openstack/nova-api-0" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.940627 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ac894fd-e293-4f28-bdb5-e80e01601903-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2ac894fd-e293-4f28-bdb5-e80e01601903\") " pod="openstack/nova-metadata-0" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.940848 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3953832e-aa6d-49f2-980d-e2308f0d1ece-public-tls-certs\") pod \"nova-api-0\" (UID: \"3953832e-aa6d-49f2-980d-e2308f0d1ece\") " pod="openstack/nova-api-0" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.942143 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3953832e-aa6d-49f2-980d-e2308f0d1ece-internal-tls-certs\") pod \"nova-api-0\" (UID: \"3953832e-aa6d-49f2-980d-e2308f0d1ece\") " pod="openstack/nova-api-0" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.943403 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ac894fd-e293-4f28-bdb5-e80e01601903-config-data\") pod \"nova-metadata-0\" (UID: \"2ac894fd-e293-4f28-bdb5-e80e01601903\") " pod="openstack/nova-metadata-0" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.949646 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ac894fd-e293-4f28-bdb5-e80e01601903-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"2ac894fd-e293-4f28-bdb5-e80e01601903\") " pod="openstack/nova-metadata-0" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.954674 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-282g2\" (UniqueName: \"kubernetes.io/projected/3953832e-aa6d-49f2-980d-e2308f0d1ece-kube-api-access-282g2\") pod \"nova-api-0\" (UID: \"3953832e-aa6d-49f2-980d-e2308f0d1ece\") " pod="openstack/nova-api-0" Dec 07 16:22:14 crc kubenswrapper[4716]: I1207 16:22:14.955070 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79qnv\" (UniqueName: \"kubernetes.io/projected/2ac894fd-e293-4f28-bdb5-e80e01601903-kube-api-access-79qnv\") pod \"nova-metadata-0\" (UID: \"2ac894fd-e293-4f28-bdb5-e80e01601903\") " pod="openstack/nova-metadata-0" Dec 07 16:22:15 crc kubenswrapper[4716]: I1207 16:22:15.012648 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 07 16:22:15 crc kubenswrapper[4716]: I1207 16:22:15.037057 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 07 16:22:15 crc kubenswrapper[4716]: I1207 16:22:15.483000 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 07 16:22:15 crc kubenswrapper[4716]: W1207 16:22:15.495701 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3953832e_aa6d_49f2_980d_e2308f0d1ece.slice/crio-645e9b1390e16b5012d16bb7fb1ed0e4ed0dadc7b3448f91c8d3f21cb03883b4 WatchSource:0}: Error finding container 645e9b1390e16b5012d16bb7fb1ed0e4ed0dadc7b3448f91c8d3f21cb03883b4: Status 404 returned error can't find the container with id 645e9b1390e16b5012d16bb7fb1ed0e4ed0dadc7b3448f91c8d3f21cb03883b4 Dec 07 16:22:15 crc kubenswrapper[4716]: I1207 16:22:15.530990 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 07 16:22:15 crc kubenswrapper[4716]: W1207 16:22:15.545743 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2ac894fd_e293_4f28_bdb5_e80e01601903.slice/crio-747fdd75543d82b4f61e79c702fa976e5bd97cf31ab18b3e13ccd9a289daac04 WatchSource:0}: Error finding container 747fdd75543d82b4f61e79c702fa976e5bd97cf31ab18b3e13ccd9a289daac04: Status 404 returned error can't find the container with id 747fdd75543d82b4f61e79c702fa976e5bd97cf31ab18b3e13ccd9a289daac04 Dec 07 16:22:15 crc kubenswrapper[4716]: I1207 16:22:15.595410 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3953832e-aa6d-49f2-980d-e2308f0d1ece","Type":"ContainerStarted","Data":"645e9b1390e16b5012d16bb7fb1ed0e4ed0dadc7b3448f91c8d3f21cb03883b4"} Dec 07 16:22:15 crc kubenswrapper[4716]: I1207 16:22:15.599536 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2ac894fd-e293-4f28-bdb5-e80e01601903","Type":"ContainerStarted","Data":"747fdd75543d82b4f61e79c702fa976e5bd97cf31ab18b3e13ccd9a289daac04"} Dec 07 16:22:15 crc kubenswrapper[4716]: I1207 16:22:15.677701 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1" path="/var/lib/kubelet/pods/19cf6b5f-e5d6-44e1-99ac-d19bfb0baee1/volumes" Dec 07 16:22:15 crc kubenswrapper[4716]: I1207 16:22:15.678635 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fdcb1fee-b3e5-4ed1-b868-875220d969be" path="/var/lib/kubelet/pods/fdcb1fee-b3e5-4ed1-b868-875220d969be/volumes" Dec 07 16:22:16 crc kubenswrapper[4716]: I1207 16:22:16.357585 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 07 16:22:16 crc kubenswrapper[4716]: I1207 16:22:16.464768 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed6ed6ec-40bd-4628-91c8-2410b68f7384-config-data\") pod \"ed6ed6ec-40bd-4628-91c8-2410b68f7384\" (UID: \"ed6ed6ec-40bd-4628-91c8-2410b68f7384\") " Dec 07 16:22:16 crc kubenswrapper[4716]: I1207 16:22:16.464919 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed6ed6ec-40bd-4628-91c8-2410b68f7384-combined-ca-bundle\") pod \"ed6ed6ec-40bd-4628-91c8-2410b68f7384\" (UID: \"ed6ed6ec-40bd-4628-91c8-2410b68f7384\") " Dec 07 16:22:16 crc kubenswrapper[4716]: I1207 16:22:16.465092 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mlh98\" (UniqueName: \"kubernetes.io/projected/ed6ed6ec-40bd-4628-91c8-2410b68f7384-kube-api-access-mlh98\") pod \"ed6ed6ec-40bd-4628-91c8-2410b68f7384\" (UID: \"ed6ed6ec-40bd-4628-91c8-2410b68f7384\") " Dec 07 16:22:16 crc kubenswrapper[4716]: I1207 16:22:16.478072 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed6ed6ec-40bd-4628-91c8-2410b68f7384-kube-api-access-mlh98" (OuterVolumeSpecName: "kube-api-access-mlh98") pod "ed6ed6ec-40bd-4628-91c8-2410b68f7384" (UID: "ed6ed6ec-40bd-4628-91c8-2410b68f7384"). InnerVolumeSpecName "kube-api-access-mlh98". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:22:16 crc kubenswrapper[4716]: I1207 16:22:16.492631 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed6ed6ec-40bd-4628-91c8-2410b68f7384-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ed6ed6ec-40bd-4628-91c8-2410b68f7384" (UID: "ed6ed6ec-40bd-4628-91c8-2410b68f7384"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:22:16 crc kubenswrapper[4716]: I1207 16:22:16.495536 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed6ed6ec-40bd-4628-91c8-2410b68f7384-config-data" (OuterVolumeSpecName: "config-data") pod "ed6ed6ec-40bd-4628-91c8-2410b68f7384" (UID: "ed6ed6ec-40bd-4628-91c8-2410b68f7384"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:22:16 crc kubenswrapper[4716]: I1207 16:22:16.567279 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mlh98\" (UniqueName: \"kubernetes.io/projected/ed6ed6ec-40bd-4628-91c8-2410b68f7384-kube-api-access-mlh98\") on node \"crc\" DevicePath \"\"" Dec 07 16:22:16 crc kubenswrapper[4716]: I1207 16:22:16.567309 4716 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed6ed6ec-40bd-4628-91c8-2410b68f7384-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 16:22:16 crc kubenswrapper[4716]: I1207 16:22:16.567319 4716 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed6ed6ec-40bd-4628-91c8-2410b68f7384-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 16:22:16 crc kubenswrapper[4716]: I1207 16:22:16.612619 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3953832e-aa6d-49f2-980d-e2308f0d1ece","Type":"ContainerStarted","Data":"861a52a098fb156247005c14e526b345178402f25a78d0cafc99c923489dc97a"} Dec 07 16:22:16 crc kubenswrapper[4716]: I1207 16:22:16.612669 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3953832e-aa6d-49f2-980d-e2308f0d1ece","Type":"ContainerStarted","Data":"7ed05b519dc88a90e9a8e1e6136cf7851fa70e5f6568edfcf2c2dc3c4e74ead4"} Dec 07 16:22:16 crc kubenswrapper[4716]: I1207 16:22:16.618250 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2ac894fd-e293-4f28-bdb5-e80e01601903","Type":"ContainerStarted","Data":"32d0f518c42c8779c719961283b82b02a561d3f1eea2612f63ac847059cb9b28"} Dec 07 16:22:16 crc kubenswrapper[4716]: I1207 16:22:16.618301 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2ac894fd-e293-4f28-bdb5-e80e01601903","Type":"ContainerStarted","Data":"59ee6ea9d3973548a67ad157975577fb8df1673952546935640524a837154b3f"} Dec 07 16:22:16 crc kubenswrapper[4716]: I1207 16:22:16.619639 4716 generic.go:334] "Generic (PLEG): container finished" podID="ed6ed6ec-40bd-4628-91c8-2410b68f7384" containerID="57fcac8ed81c947a24d209d04c59bfeb69b36b49d18fa729d309902f2e7b5e08" exitCode=0 Dec 07 16:22:16 crc kubenswrapper[4716]: I1207 16:22:16.619698 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ed6ed6ec-40bd-4628-91c8-2410b68f7384","Type":"ContainerDied","Data":"57fcac8ed81c947a24d209d04c59bfeb69b36b49d18fa729d309902f2e7b5e08"} Dec 07 16:22:16 crc kubenswrapper[4716]: I1207 16:22:16.620108 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 07 16:22:16 crc kubenswrapper[4716]: I1207 16:22:16.620190 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ed6ed6ec-40bd-4628-91c8-2410b68f7384","Type":"ContainerDied","Data":"b5767352f18a00cb5ba3b22faaca38a55b64a41cb97249d96fd0f9fcf4ef684b"} Dec 07 16:22:16 crc kubenswrapper[4716]: I1207 16:22:16.620225 4716 scope.go:117] "RemoveContainer" containerID="57fcac8ed81c947a24d209d04c59bfeb69b36b49d18fa729d309902f2e7b5e08" Dec 07 16:22:16 crc kubenswrapper[4716]: I1207 16:22:16.644729 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.64470733 podStartE2EDuration="2.64470733s" podCreationTimestamp="2025-12-07 16:22:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:22:16.642417268 +0000 UTC m=+1199.332702180" watchObservedRunningTime="2025-12-07 16:22:16.64470733 +0000 UTC m=+1199.334992252" Dec 07 16:22:16 crc kubenswrapper[4716]: I1207 16:22:16.661650 4716 scope.go:117] "RemoveContainer" containerID="57fcac8ed81c947a24d209d04c59bfeb69b36b49d18fa729d309902f2e7b5e08" Dec 07 16:22:16 crc kubenswrapper[4716]: E1207 16:22:16.662406 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"57fcac8ed81c947a24d209d04c59bfeb69b36b49d18fa729d309902f2e7b5e08\": container with ID starting with 57fcac8ed81c947a24d209d04c59bfeb69b36b49d18fa729d309902f2e7b5e08 not found: ID does not exist" containerID="57fcac8ed81c947a24d209d04c59bfeb69b36b49d18fa729d309902f2e7b5e08" Dec 07 16:22:16 crc kubenswrapper[4716]: I1207 16:22:16.662448 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"57fcac8ed81c947a24d209d04c59bfeb69b36b49d18fa729d309902f2e7b5e08"} err="failed to get container status \"57fcac8ed81c947a24d209d04c59bfeb69b36b49d18fa729d309902f2e7b5e08\": rpc error: code = NotFound desc = could not find container \"57fcac8ed81c947a24d209d04c59bfeb69b36b49d18fa729d309902f2e7b5e08\": container with ID starting with 57fcac8ed81c947a24d209d04c59bfeb69b36b49d18fa729d309902f2e7b5e08 not found: ID does not exist" Dec 07 16:22:16 crc kubenswrapper[4716]: I1207 16:22:16.678170 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.6781514509999997 podStartE2EDuration="2.678151451s" podCreationTimestamp="2025-12-07 16:22:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:22:16.658763259 +0000 UTC m=+1199.349048181" watchObservedRunningTime="2025-12-07 16:22:16.678151451 +0000 UTC m=+1199.368436363" Dec 07 16:22:16 crc kubenswrapper[4716]: I1207 16:22:16.690500 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 07 16:22:16 crc kubenswrapper[4716]: I1207 16:22:16.701353 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 07 16:22:16 crc kubenswrapper[4716]: I1207 16:22:16.708012 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 07 16:22:16 crc kubenswrapper[4716]: E1207 16:22:16.708404 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed6ed6ec-40bd-4628-91c8-2410b68f7384" containerName="nova-scheduler-scheduler" Dec 07 16:22:16 crc kubenswrapper[4716]: I1207 16:22:16.708422 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed6ed6ec-40bd-4628-91c8-2410b68f7384" containerName="nova-scheduler-scheduler" Dec 07 16:22:16 crc kubenswrapper[4716]: I1207 16:22:16.708634 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed6ed6ec-40bd-4628-91c8-2410b68f7384" containerName="nova-scheduler-scheduler" Dec 07 16:22:16 crc kubenswrapper[4716]: I1207 16:22:16.709205 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 07 16:22:16 crc kubenswrapper[4716]: I1207 16:22:16.717622 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 07 16:22:16 crc kubenswrapper[4716]: I1207 16:22:16.718753 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 07 16:22:16 crc kubenswrapper[4716]: I1207 16:22:16.771537 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/794a9d02-75dc-4ad6-bdb1-51bccaab7c9a-config-data\") pod \"nova-scheduler-0\" (UID: \"794a9d02-75dc-4ad6-bdb1-51bccaab7c9a\") " pod="openstack/nova-scheduler-0" Dec 07 16:22:16 crc kubenswrapper[4716]: I1207 16:22:16.771629 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-24smp\" (UniqueName: \"kubernetes.io/projected/794a9d02-75dc-4ad6-bdb1-51bccaab7c9a-kube-api-access-24smp\") pod \"nova-scheduler-0\" (UID: \"794a9d02-75dc-4ad6-bdb1-51bccaab7c9a\") " pod="openstack/nova-scheduler-0" Dec 07 16:22:16 crc kubenswrapper[4716]: I1207 16:22:16.771745 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/794a9d02-75dc-4ad6-bdb1-51bccaab7c9a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"794a9d02-75dc-4ad6-bdb1-51bccaab7c9a\") " pod="openstack/nova-scheduler-0" Dec 07 16:22:16 crc kubenswrapper[4716]: I1207 16:22:16.873695 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/794a9d02-75dc-4ad6-bdb1-51bccaab7c9a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"794a9d02-75dc-4ad6-bdb1-51bccaab7c9a\") " pod="openstack/nova-scheduler-0" Dec 07 16:22:16 crc kubenswrapper[4716]: I1207 16:22:16.873815 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/794a9d02-75dc-4ad6-bdb1-51bccaab7c9a-config-data\") pod \"nova-scheduler-0\" (UID: \"794a9d02-75dc-4ad6-bdb1-51bccaab7c9a\") " pod="openstack/nova-scheduler-0" Dec 07 16:22:16 crc kubenswrapper[4716]: I1207 16:22:16.873855 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-24smp\" (UniqueName: \"kubernetes.io/projected/794a9d02-75dc-4ad6-bdb1-51bccaab7c9a-kube-api-access-24smp\") pod \"nova-scheduler-0\" (UID: \"794a9d02-75dc-4ad6-bdb1-51bccaab7c9a\") " pod="openstack/nova-scheduler-0" Dec 07 16:22:16 crc kubenswrapper[4716]: I1207 16:22:16.877443 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/794a9d02-75dc-4ad6-bdb1-51bccaab7c9a-config-data\") pod \"nova-scheduler-0\" (UID: \"794a9d02-75dc-4ad6-bdb1-51bccaab7c9a\") " pod="openstack/nova-scheduler-0" Dec 07 16:22:16 crc kubenswrapper[4716]: I1207 16:22:16.877895 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/794a9d02-75dc-4ad6-bdb1-51bccaab7c9a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"794a9d02-75dc-4ad6-bdb1-51bccaab7c9a\") " pod="openstack/nova-scheduler-0" Dec 07 16:22:16 crc kubenswrapper[4716]: I1207 16:22:16.892904 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-24smp\" (UniqueName: \"kubernetes.io/projected/794a9d02-75dc-4ad6-bdb1-51bccaab7c9a-kube-api-access-24smp\") pod \"nova-scheduler-0\" (UID: \"794a9d02-75dc-4ad6-bdb1-51bccaab7c9a\") " pod="openstack/nova-scheduler-0" Dec 07 16:22:17 crc kubenswrapper[4716]: I1207 16:22:17.030748 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 07 16:22:17 crc kubenswrapper[4716]: W1207 16:22:17.461591 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod794a9d02_75dc_4ad6_bdb1_51bccaab7c9a.slice/crio-4dad3cd372472afd4e8d4cffef8ab035293739494f0733fd41badd1a8f17b627 WatchSource:0}: Error finding container 4dad3cd372472afd4e8d4cffef8ab035293739494f0733fd41badd1a8f17b627: Status 404 returned error can't find the container with id 4dad3cd372472afd4e8d4cffef8ab035293739494f0733fd41badd1a8f17b627 Dec 07 16:22:17 crc kubenswrapper[4716]: I1207 16:22:17.462584 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 07 16:22:17 crc kubenswrapper[4716]: I1207 16:22:17.628964 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"794a9d02-75dc-4ad6-bdb1-51bccaab7c9a","Type":"ContainerStarted","Data":"4dad3cd372472afd4e8d4cffef8ab035293739494f0733fd41badd1a8f17b627"} Dec 07 16:22:17 crc kubenswrapper[4716]: I1207 16:22:17.672873 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed6ed6ec-40bd-4628-91c8-2410b68f7384" path="/var/lib/kubelet/pods/ed6ed6ec-40bd-4628-91c8-2410b68f7384/volumes" Dec 07 16:22:18 crc kubenswrapper[4716]: I1207 16:22:18.641654 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"794a9d02-75dc-4ad6-bdb1-51bccaab7c9a","Type":"ContainerStarted","Data":"e46f04fd6f65a9a01993867b773e15d5e897def4c9984cfd92ed477ad4b27f6a"} Dec 07 16:22:18 crc kubenswrapper[4716]: I1207 16:22:18.658791 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.658773451 podStartE2EDuration="2.658773451s" podCreationTimestamp="2025-12-07 16:22:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:22:18.65575321 +0000 UTC m=+1201.346038132" watchObservedRunningTime="2025-12-07 16:22:18.658773451 +0000 UTC m=+1201.349058373" Dec 07 16:22:20 crc kubenswrapper[4716]: I1207 16:22:20.039196 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 07 16:22:20 crc kubenswrapper[4716]: I1207 16:22:20.040777 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 07 16:22:22 crc kubenswrapper[4716]: I1207 16:22:22.031658 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 07 16:22:22 crc kubenswrapper[4716]: I1207 16:22:22.761702 4716 patch_prober.go:28] interesting pod/machine-config-daemon-zcxxp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 16:22:22 crc kubenswrapper[4716]: I1207 16:22:22.761790 4716 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 16:22:23 crc kubenswrapper[4716]: I1207 16:22:23.969564 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 07 16:22:25 crc kubenswrapper[4716]: I1207 16:22:25.013169 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 07 16:22:25 crc kubenswrapper[4716]: I1207 16:22:25.014189 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 07 16:22:25 crc kubenswrapper[4716]: I1207 16:22:25.037556 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 07 16:22:25 crc kubenswrapper[4716]: I1207 16:22:25.037603 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 07 16:22:26 crc kubenswrapper[4716]: I1207 16:22:26.029221 4716 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="3953832e-aa6d-49f2-980d-e2308f0d1ece" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.201:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 07 16:22:26 crc kubenswrapper[4716]: I1207 16:22:26.029696 4716 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="3953832e-aa6d-49f2-980d-e2308f0d1ece" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.201:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 07 16:22:26 crc kubenswrapper[4716]: I1207 16:22:26.047214 4716 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="2ac894fd-e293-4f28-bdb5-e80e01601903" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.202:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 07 16:22:26 crc kubenswrapper[4716]: I1207 16:22:26.047216 4716 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="2ac894fd-e293-4f28-bdb5-e80e01601903" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.202:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 07 16:22:27 crc kubenswrapper[4716]: I1207 16:22:27.032005 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 07 16:22:27 crc kubenswrapper[4716]: I1207 16:22:27.063879 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 07 16:22:27 crc kubenswrapper[4716]: I1207 16:22:27.791971 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 07 16:22:35 crc kubenswrapper[4716]: I1207 16:22:35.023089 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 07 16:22:35 crc kubenswrapper[4716]: I1207 16:22:35.024049 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 07 16:22:35 crc kubenswrapper[4716]: I1207 16:22:35.028341 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 07 16:22:35 crc kubenswrapper[4716]: I1207 16:22:35.030889 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 07 16:22:35 crc kubenswrapper[4716]: I1207 16:22:35.043717 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 07 16:22:35 crc kubenswrapper[4716]: I1207 16:22:35.045429 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 07 16:22:35 crc kubenswrapper[4716]: I1207 16:22:35.058268 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 07 16:22:35 crc kubenswrapper[4716]: I1207 16:22:35.860463 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 07 16:22:35 crc kubenswrapper[4716]: I1207 16:22:35.866829 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 07 16:22:35 crc kubenswrapper[4716]: I1207 16:22:35.869132 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 07 16:22:44 crc kubenswrapper[4716]: I1207 16:22:44.742210 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 07 16:22:45 crc kubenswrapper[4716]: I1207 16:22:45.761658 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 07 16:22:49 crc kubenswrapper[4716]: I1207 16:22:49.302794 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="f653df87-c234-4cb6-8ee8-6b16b4fe0044" containerName="rabbitmq" containerID="cri-o://e24ca70ffe380dd6ae0b3ce56869f1dfe6b1c7a31169cf168b558517d0a1aa65" gracePeriod=604796 Dec 07 16:22:49 crc kubenswrapper[4716]: I1207 16:22:49.861128 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe" containerName="rabbitmq" containerID="cri-o://9c52f289d4d35a7ab538719f9c3c5eb4d453bb8387f9e697d5cf26704c98cd04" gracePeriod=604796 Dec 07 16:22:52 crc kubenswrapper[4716]: I1207 16:22:52.761617 4716 patch_prober.go:28] interesting pod/machine-config-daemon-zcxxp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 16:22:52 crc kubenswrapper[4716]: I1207 16:22:52.761918 4716 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 16:22:55 crc kubenswrapper[4716]: I1207 16:22:55.845967 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 07 16:22:55 crc kubenswrapper[4716]: I1207 16:22:55.952324 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f653df87-c234-4cb6-8ee8-6b16b4fe0044-rabbitmq-confd\") pod \"f653df87-c234-4cb6-8ee8-6b16b4fe0044\" (UID: \"f653df87-c234-4cb6-8ee8-6b16b4fe0044\") " Dec 07 16:22:55 crc kubenswrapper[4716]: I1207 16:22:55.952430 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f653df87-c234-4cb6-8ee8-6b16b4fe0044-config-data\") pod \"f653df87-c234-4cb6-8ee8-6b16b4fe0044\" (UID: \"f653df87-c234-4cb6-8ee8-6b16b4fe0044\") " Dec 07 16:22:55 crc kubenswrapper[4716]: I1207 16:22:55.952461 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v8bpb\" (UniqueName: \"kubernetes.io/projected/f653df87-c234-4cb6-8ee8-6b16b4fe0044-kube-api-access-v8bpb\") pod \"f653df87-c234-4cb6-8ee8-6b16b4fe0044\" (UID: \"f653df87-c234-4cb6-8ee8-6b16b4fe0044\") " Dec 07 16:22:55 crc kubenswrapper[4716]: I1207 16:22:55.952534 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f653df87-c234-4cb6-8ee8-6b16b4fe0044-server-conf\") pod \"f653df87-c234-4cb6-8ee8-6b16b4fe0044\" (UID: \"f653df87-c234-4cb6-8ee8-6b16b4fe0044\") " Dec 07 16:22:55 crc kubenswrapper[4716]: I1207 16:22:55.952578 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f653df87-c234-4cb6-8ee8-6b16b4fe0044-rabbitmq-plugins\") pod \"f653df87-c234-4cb6-8ee8-6b16b4fe0044\" (UID: \"f653df87-c234-4cb6-8ee8-6b16b4fe0044\") " Dec 07 16:22:55 crc kubenswrapper[4716]: I1207 16:22:55.952628 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f653df87-c234-4cb6-8ee8-6b16b4fe0044-plugins-conf\") pod \"f653df87-c234-4cb6-8ee8-6b16b4fe0044\" (UID: \"f653df87-c234-4cb6-8ee8-6b16b4fe0044\") " Dec 07 16:22:55 crc kubenswrapper[4716]: I1207 16:22:55.953589 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f653df87-c234-4cb6-8ee8-6b16b4fe0044-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "f653df87-c234-4cb6-8ee8-6b16b4fe0044" (UID: "f653df87-c234-4cb6-8ee8-6b16b4fe0044"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:22:55 crc kubenswrapper[4716]: I1207 16:22:55.953838 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f653df87-c234-4cb6-8ee8-6b16b4fe0044-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "f653df87-c234-4cb6-8ee8-6b16b4fe0044" (UID: "f653df87-c234-4cb6-8ee8-6b16b4fe0044"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:22:55 crc kubenswrapper[4716]: I1207 16:22:55.955562 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"f653df87-c234-4cb6-8ee8-6b16b4fe0044\" (UID: \"f653df87-c234-4cb6-8ee8-6b16b4fe0044\") " Dec 07 16:22:55 crc kubenswrapper[4716]: I1207 16:22:55.955682 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f653df87-c234-4cb6-8ee8-6b16b4fe0044-rabbitmq-tls\") pod \"f653df87-c234-4cb6-8ee8-6b16b4fe0044\" (UID: \"f653df87-c234-4cb6-8ee8-6b16b4fe0044\") " Dec 07 16:22:55 crc kubenswrapper[4716]: I1207 16:22:55.955744 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f653df87-c234-4cb6-8ee8-6b16b4fe0044-erlang-cookie-secret\") pod \"f653df87-c234-4cb6-8ee8-6b16b4fe0044\" (UID: \"f653df87-c234-4cb6-8ee8-6b16b4fe0044\") " Dec 07 16:22:55 crc kubenswrapper[4716]: I1207 16:22:55.955778 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f653df87-c234-4cb6-8ee8-6b16b4fe0044-pod-info\") pod \"f653df87-c234-4cb6-8ee8-6b16b4fe0044\" (UID: \"f653df87-c234-4cb6-8ee8-6b16b4fe0044\") " Dec 07 16:22:55 crc kubenswrapper[4716]: I1207 16:22:55.955807 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f653df87-c234-4cb6-8ee8-6b16b4fe0044-rabbitmq-erlang-cookie\") pod \"f653df87-c234-4cb6-8ee8-6b16b4fe0044\" (UID: \"f653df87-c234-4cb6-8ee8-6b16b4fe0044\") " Dec 07 16:22:55 crc kubenswrapper[4716]: I1207 16:22:55.956421 4716 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f653df87-c234-4cb6-8ee8-6b16b4fe0044-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 07 16:22:55 crc kubenswrapper[4716]: I1207 16:22:55.956444 4716 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f653df87-c234-4cb6-8ee8-6b16b4fe0044-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 07 16:22:55 crc kubenswrapper[4716]: I1207 16:22:55.956992 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f653df87-c234-4cb6-8ee8-6b16b4fe0044-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "f653df87-c234-4cb6-8ee8-6b16b4fe0044" (UID: "f653df87-c234-4cb6-8ee8-6b16b4fe0044"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:22:55 crc kubenswrapper[4716]: I1207 16:22:55.960499 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f653df87-c234-4cb6-8ee8-6b16b4fe0044-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "f653df87-c234-4cb6-8ee8-6b16b4fe0044" (UID: "f653df87-c234-4cb6-8ee8-6b16b4fe0044"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:22:55 crc kubenswrapper[4716]: I1207 16:22:55.963563 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "persistence") pod "f653df87-c234-4cb6-8ee8-6b16b4fe0044" (UID: "f653df87-c234-4cb6-8ee8-6b16b4fe0044"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 07 16:22:55 crc kubenswrapper[4716]: I1207 16:22:55.972805 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f653df87-c234-4cb6-8ee8-6b16b4fe0044-kube-api-access-v8bpb" (OuterVolumeSpecName: "kube-api-access-v8bpb") pod "f653df87-c234-4cb6-8ee8-6b16b4fe0044" (UID: "f653df87-c234-4cb6-8ee8-6b16b4fe0044"). InnerVolumeSpecName "kube-api-access-v8bpb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:22:55 crc kubenswrapper[4716]: I1207 16:22:55.973181 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f653df87-c234-4cb6-8ee8-6b16b4fe0044-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "f653df87-c234-4cb6-8ee8-6b16b4fe0044" (UID: "f653df87-c234-4cb6-8ee8-6b16b4fe0044"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:22:55 crc kubenswrapper[4716]: I1207 16:22:55.974709 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/f653df87-c234-4cb6-8ee8-6b16b4fe0044-pod-info" (OuterVolumeSpecName: "pod-info") pod "f653df87-c234-4cb6-8ee8-6b16b4fe0044" (UID: "f653df87-c234-4cb6-8ee8-6b16b4fe0044"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.008955 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f653df87-c234-4cb6-8ee8-6b16b4fe0044-config-data" (OuterVolumeSpecName: "config-data") pod "f653df87-c234-4cb6-8ee8-6b16b4fe0044" (UID: "f653df87-c234-4cb6-8ee8-6b16b4fe0044"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.041237 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f653df87-c234-4cb6-8ee8-6b16b4fe0044-server-conf" (OuterVolumeSpecName: "server-conf") pod "f653df87-c234-4cb6-8ee8-6b16b4fe0044" (UID: "f653df87-c234-4cb6-8ee8-6b16b4fe0044"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.058471 4716 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f653df87-c234-4cb6-8ee8-6b16b4fe0044-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.058709 4716 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f653df87-c234-4cb6-8ee8-6b16b4fe0044-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.058801 4716 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f653df87-c234-4cb6-8ee8-6b16b4fe0044-pod-info\") on node \"crc\" DevicePath \"\"" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.058884 4716 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f653df87-c234-4cb6-8ee8-6b16b4fe0044-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.058958 4716 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f653df87-c234-4cb6-8ee8-6b16b4fe0044-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.059024 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v8bpb\" (UniqueName: \"kubernetes.io/projected/f653df87-c234-4cb6-8ee8-6b16b4fe0044-kube-api-access-v8bpb\") on node \"crc\" DevicePath \"\"" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.059182 4716 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f653df87-c234-4cb6-8ee8-6b16b4fe0044-server-conf\") on node \"crc\" DevicePath \"\"" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.059293 4716 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.092580 4716 generic.go:334] "Generic (PLEG): container finished" podID="f653df87-c234-4cb6-8ee8-6b16b4fe0044" containerID="e24ca70ffe380dd6ae0b3ce56869f1dfe6b1c7a31169cf168b558517d0a1aa65" exitCode=0 Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.092647 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f653df87-c234-4cb6-8ee8-6b16b4fe0044","Type":"ContainerDied","Data":"e24ca70ffe380dd6ae0b3ce56869f1dfe6b1c7a31169cf168b558517d0a1aa65"} Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.092674 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f653df87-c234-4cb6-8ee8-6b16b4fe0044","Type":"ContainerDied","Data":"f7fef3c71a02a05d212998dcea6ed0986974ed48e929dc2cca1cd5fd191035db"} Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.092689 4716 scope.go:117] "RemoveContainer" containerID="e24ca70ffe380dd6ae0b3ce56869f1dfe6b1c7a31169cf168b558517d0a1aa65" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.092698 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.101588 4716 generic.go:334] "Generic (PLEG): container finished" podID="9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe" containerID="9c52f289d4d35a7ab538719f9c3c5eb4d453bb8387f9e697d5cf26704c98cd04" exitCode=0 Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.101621 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe","Type":"ContainerDied","Data":"9c52f289d4d35a7ab538719f9c3c5eb4d453bb8387f9e697d5cf26704c98cd04"} Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.110789 4716 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.111880 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f653df87-c234-4cb6-8ee8-6b16b4fe0044-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "f653df87-c234-4cb6-8ee8-6b16b4fe0044" (UID: "f653df87-c234-4cb6-8ee8-6b16b4fe0044"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.143824 4716 scope.go:117] "RemoveContainer" containerID="c4250e9cb4d0c6bd1c2d3e3bc3ca8c43b9a53e5cd186b5bd154837621d4a55f1" Dec 07 16:22:56 crc kubenswrapper[4716]: E1207 16:22:56.156246 4716 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9ca0dfec_a8e3_40d8_9d66_6269a6aa42fe.slice/crio-conmon-9c52f289d4d35a7ab538719f9c3c5eb4d453bb8387f9e697d5cf26704c98cd04.scope\": RecentStats: unable to find data in memory cache]" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.161609 4716 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.161773 4716 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f653df87-c234-4cb6-8ee8-6b16b4fe0044-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.205486 4716 scope.go:117] "RemoveContainer" containerID="e24ca70ffe380dd6ae0b3ce56869f1dfe6b1c7a31169cf168b558517d0a1aa65" Dec 07 16:22:56 crc kubenswrapper[4716]: E1207 16:22:56.206390 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e24ca70ffe380dd6ae0b3ce56869f1dfe6b1c7a31169cf168b558517d0a1aa65\": container with ID starting with e24ca70ffe380dd6ae0b3ce56869f1dfe6b1c7a31169cf168b558517d0a1aa65 not found: ID does not exist" containerID="e24ca70ffe380dd6ae0b3ce56869f1dfe6b1c7a31169cf168b558517d0a1aa65" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.207171 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e24ca70ffe380dd6ae0b3ce56869f1dfe6b1c7a31169cf168b558517d0a1aa65"} err="failed to get container status \"e24ca70ffe380dd6ae0b3ce56869f1dfe6b1c7a31169cf168b558517d0a1aa65\": rpc error: code = NotFound desc = could not find container \"e24ca70ffe380dd6ae0b3ce56869f1dfe6b1c7a31169cf168b558517d0a1aa65\": container with ID starting with e24ca70ffe380dd6ae0b3ce56869f1dfe6b1c7a31169cf168b558517d0a1aa65 not found: ID does not exist" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.207216 4716 scope.go:117] "RemoveContainer" containerID="c4250e9cb4d0c6bd1c2d3e3bc3ca8c43b9a53e5cd186b5bd154837621d4a55f1" Dec 07 16:22:56 crc kubenswrapper[4716]: E1207 16:22:56.208742 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c4250e9cb4d0c6bd1c2d3e3bc3ca8c43b9a53e5cd186b5bd154837621d4a55f1\": container with ID starting with c4250e9cb4d0c6bd1c2d3e3bc3ca8c43b9a53e5cd186b5bd154837621d4a55f1 not found: ID does not exist" containerID="c4250e9cb4d0c6bd1c2d3e3bc3ca8c43b9a53e5cd186b5bd154837621d4a55f1" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.208793 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4250e9cb4d0c6bd1c2d3e3bc3ca8c43b9a53e5cd186b5bd154837621d4a55f1"} err="failed to get container status \"c4250e9cb4d0c6bd1c2d3e3bc3ca8c43b9a53e5cd186b5bd154837621d4a55f1\": rpc error: code = NotFound desc = could not find container \"c4250e9cb4d0c6bd1c2d3e3bc3ca8c43b9a53e5cd186b5bd154837621d4a55f1\": container with ID starting with c4250e9cb4d0c6bd1c2d3e3bc3ca8c43b9a53e5cd186b5bd154837621d4a55f1 not found: ID does not exist" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.401330 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.457084 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.463955 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.466681 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6zsm9\" (UniqueName: \"kubernetes.io/projected/9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe-kube-api-access-6zsm9\") pod \"9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe\" (UID: \"9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe\") " Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.466721 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe\" (UID: \"9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe\") " Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.466754 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe-server-conf\") pod \"9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe\" (UID: \"9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe\") " Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.466778 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe-erlang-cookie-secret\") pod \"9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe\" (UID: \"9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe\") " Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.466822 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe-config-data\") pod \"9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe\" (UID: \"9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe\") " Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.466854 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe-rabbitmq-tls\") pod \"9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe\" (UID: \"9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe\") " Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.466947 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe-rabbitmq-confd\") pod \"9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe\" (UID: \"9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe\") " Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.467032 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe-rabbitmq-erlang-cookie\") pod \"9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe\" (UID: \"9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe\") " Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.467175 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe-rabbitmq-plugins\") pod \"9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe\" (UID: \"9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe\") " Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.467205 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe-plugins-conf\") pod \"9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe\" (UID: \"9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe\") " Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.467243 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe-pod-info\") pod \"9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe\" (UID: \"9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe\") " Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.467864 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe" (UID: "9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.472772 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe" (UID: "9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.477188 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe" (UID: "9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.497242 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe" (UID: "9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.497586 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe-config-data" (OuterVolumeSpecName: "config-data") pod "9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe" (UID: "9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.499997 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "persistence") pod "9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe" (UID: "9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.500151 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe-pod-info" (OuterVolumeSpecName: "pod-info") pod "9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe" (UID: "9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.500151 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe-kube-api-access-6zsm9" (OuterVolumeSpecName: "kube-api-access-6zsm9") pod "9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe" (UID: "9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe"). InnerVolumeSpecName "kube-api-access-6zsm9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.501883 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 07 16:22:56 crc kubenswrapper[4716]: E1207 16:22:56.502279 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f653df87-c234-4cb6-8ee8-6b16b4fe0044" containerName="setup-container" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.502296 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="f653df87-c234-4cb6-8ee8-6b16b4fe0044" containerName="setup-container" Dec 07 16:22:56 crc kubenswrapper[4716]: E1207 16:22:56.502321 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe" containerName="setup-container" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.502328 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe" containerName="setup-container" Dec 07 16:22:56 crc kubenswrapper[4716]: E1207 16:22:56.502360 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f653df87-c234-4cb6-8ee8-6b16b4fe0044" containerName="rabbitmq" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.502367 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="f653df87-c234-4cb6-8ee8-6b16b4fe0044" containerName="rabbitmq" Dec 07 16:22:56 crc kubenswrapper[4716]: E1207 16:22:56.502377 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe" containerName="rabbitmq" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.502384 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe" containerName="rabbitmq" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.502541 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe" containerName="rabbitmq" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.502554 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="f653df87-c234-4cb6-8ee8-6b16b4fe0044" containerName="rabbitmq" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.503520 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.504248 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe" (UID: "9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.507004 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.507175 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.507285 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.507338 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.507436 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.507538 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-5tlk9" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.507653 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.514727 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.547479 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe-server-conf" (OuterVolumeSpecName: "server-conf") pod "9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe" (UID: "9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.569464 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"db1d1fd3-b1ed-472b-b615-b866706b28a9\") " pod="openstack/rabbitmq-server-0" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.569512 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/db1d1fd3-b1ed-472b-b615-b866706b28a9-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"db1d1fd3-b1ed-472b-b615-b866706b28a9\") " pod="openstack/rabbitmq-server-0" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.569554 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/db1d1fd3-b1ed-472b-b615-b866706b28a9-config-data\") pod \"rabbitmq-server-0\" (UID: \"db1d1fd3-b1ed-472b-b615-b866706b28a9\") " pod="openstack/rabbitmq-server-0" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.569572 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/db1d1fd3-b1ed-472b-b615-b866706b28a9-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"db1d1fd3-b1ed-472b-b615-b866706b28a9\") " pod="openstack/rabbitmq-server-0" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.569607 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/db1d1fd3-b1ed-472b-b615-b866706b28a9-pod-info\") pod \"rabbitmq-server-0\" (UID: \"db1d1fd3-b1ed-472b-b615-b866706b28a9\") " pod="openstack/rabbitmq-server-0" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.569657 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/db1d1fd3-b1ed-472b-b615-b866706b28a9-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"db1d1fd3-b1ed-472b-b615-b866706b28a9\") " pod="openstack/rabbitmq-server-0" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.569689 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjf76\" (UniqueName: \"kubernetes.io/projected/db1d1fd3-b1ed-472b-b615-b866706b28a9-kube-api-access-hjf76\") pod \"rabbitmq-server-0\" (UID: \"db1d1fd3-b1ed-472b-b615-b866706b28a9\") " pod="openstack/rabbitmq-server-0" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.569818 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/db1d1fd3-b1ed-472b-b615-b866706b28a9-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"db1d1fd3-b1ed-472b-b615-b866706b28a9\") " pod="openstack/rabbitmq-server-0" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.569856 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/db1d1fd3-b1ed-472b-b615-b866706b28a9-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"db1d1fd3-b1ed-472b-b615-b866706b28a9\") " pod="openstack/rabbitmq-server-0" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.569900 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/db1d1fd3-b1ed-472b-b615-b866706b28a9-server-conf\") pod \"rabbitmq-server-0\" (UID: \"db1d1fd3-b1ed-472b-b615-b866706b28a9\") " pod="openstack/rabbitmq-server-0" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.569923 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/db1d1fd3-b1ed-472b-b615-b866706b28a9-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"db1d1fd3-b1ed-472b-b615-b866706b28a9\") " pod="openstack/rabbitmq-server-0" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.569991 4716 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.570008 4716 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.570020 4716 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.570029 4716 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe-pod-info\") on node \"crc\" DevicePath \"\"" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.570042 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6zsm9\" (UniqueName: \"kubernetes.io/projected/9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe-kube-api-access-6zsm9\") on node \"crc\" DevicePath \"\"" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.570064 4716 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.570098 4716 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe-server-conf\") on node \"crc\" DevicePath \"\"" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.570109 4716 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.570122 4716 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.570131 4716 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.606484 4716 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.642780 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe" (UID: "9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.671286 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/db1d1fd3-b1ed-472b-b615-b866706b28a9-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"db1d1fd3-b1ed-472b-b615-b866706b28a9\") " pod="openstack/rabbitmq-server-0" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.671590 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/db1d1fd3-b1ed-472b-b615-b866706b28a9-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"db1d1fd3-b1ed-472b-b615-b866706b28a9\") " pod="openstack/rabbitmq-server-0" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.671730 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/db1d1fd3-b1ed-472b-b615-b866706b28a9-server-conf\") pod \"rabbitmq-server-0\" (UID: \"db1d1fd3-b1ed-472b-b615-b866706b28a9\") " pod="openstack/rabbitmq-server-0" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.671810 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/db1d1fd3-b1ed-472b-b615-b866706b28a9-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"db1d1fd3-b1ed-472b-b615-b866706b28a9\") " pod="openstack/rabbitmq-server-0" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.671892 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"db1d1fd3-b1ed-472b-b615-b866706b28a9\") " pod="openstack/rabbitmq-server-0" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.671965 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/db1d1fd3-b1ed-472b-b615-b866706b28a9-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"db1d1fd3-b1ed-472b-b615-b866706b28a9\") " pod="openstack/rabbitmq-server-0" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.672088 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/db1d1fd3-b1ed-472b-b615-b866706b28a9-config-data\") pod \"rabbitmq-server-0\" (UID: \"db1d1fd3-b1ed-472b-b615-b866706b28a9\") " pod="openstack/rabbitmq-server-0" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.672197 4716 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"db1d1fd3-b1ed-472b-b615-b866706b28a9\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/rabbitmq-server-0" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.671987 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/db1d1fd3-b1ed-472b-b615-b866706b28a9-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"db1d1fd3-b1ed-472b-b615-b866706b28a9\") " pod="openstack/rabbitmq-server-0" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.672643 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/db1d1fd3-b1ed-472b-b615-b866706b28a9-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"db1d1fd3-b1ed-472b-b615-b866706b28a9\") " pod="openstack/rabbitmq-server-0" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.673361 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/db1d1fd3-b1ed-472b-b615-b866706b28a9-server-conf\") pod \"rabbitmq-server-0\" (UID: \"db1d1fd3-b1ed-472b-b615-b866706b28a9\") " pod="openstack/rabbitmq-server-0" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.672207 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/db1d1fd3-b1ed-472b-b615-b866706b28a9-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"db1d1fd3-b1ed-472b-b615-b866706b28a9\") " pod="openstack/rabbitmq-server-0" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.673502 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/db1d1fd3-b1ed-472b-b615-b866706b28a9-pod-info\") pod \"rabbitmq-server-0\" (UID: \"db1d1fd3-b1ed-472b-b615-b866706b28a9\") " pod="openstack/rabbitmq-server-0" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.673561 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/db1d1fd3-b1ed-472b-b615-b866706b28a9-config-data\") pod \"rabbitmq-server-0\" (UID: \"db1d1fd3-b1ed-472b-b615-b866706b28a9\") " pod="openstack/rabbitmq-server-0" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.673654 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/db1d1fd3-b1ed-472b-b615-b866706b28a9-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"db1d1fd3-b1ed-472b-b615-b866706b28a9\") " pod="openstack/rabbitmq-server-0" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.673743 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjf76\" (UniqueName: \"kubernetes.io/projected/db1d1fd3-b1ed-472b-b615-b866706b28a9-kube-api-access-hjf76\") pod \"rabbitmq-server-0\" (UID: \"db1d1fd3-b1ed-472b-b615-b866706b28a9\") " pod="openstack/rabbitmq-server-0" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.673940 4716 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.673962 4716 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.674139 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/db1d1fd3-b1ed-472b-b615-b866706b28a9-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"db1d1fd3-b1ed-472b-b615-b866706b28a9\") " pod="openstack/rabbitmq-server-0" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.678575 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/db1d1fd3-b1ed-472b-b615-b866706b28a9-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"db1d1fd3-b1ed-472b-b615-b866706b28a9\") " pod="openstack/rabbitmq-server-0" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.679728 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/db1d1fd3-b1ed-472b-b615-b866706b28a9-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"db1d1fd3-b1ed-472b-b615-b866706b28a9\") " pod="openstack/rabbitmq-server-0" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.680350 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/db1d1fd3-b1ed-472b-b615-b866706b28a9-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"db1d1fd3-b1ed-472b-b615-b866706b28a9\") " pod="openstack/rabbitmq-server-0" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.680791 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/db1d1fd3-b1ed-472b-b615-b866706b28a9-pod-info\") pod \"rabbitmq-server-0\" (UID: \"db1d1fd3-b1ed-472b-b615-b866706b28a9\") " pod="openstack/rabbitmq-server-0" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.696050 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjf76\" (UniqueName: \"kubernetes.io/projected/db1d1fd3-b1ed-472b-b615-b866706b28a9-kube-api-access-hjf76\") pod \"rabbitmq-server-0\" (UID: \"db1d1fd3-b1ed-472b-b615-b866706b28a9\") " pod="openstack/rabbitmq-server-0" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.710054 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"db1d1fd3-b1ed-472b-b615-b866706b28a9\") " pod="openstack/rabbitmq-server-0" Dec 07 16:22:56 crc kubenswrapper[4716]: I1207 16:22:56.769629 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 07 16:22:57 crc kubenswrapper[4716]: I1207 16:22:57.112136 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe","Type":"ContainerDied","Data":"0c115d03172b1cc97cbbfbf14c42f1452617bd97a6f36bca2ea6f8acc3f24cf5"} Dec 07 16:22:57 crc kubenswrapper[4716]: I1207 16:22:57.112478 4716 scope.go:117] "RemoveContainer" containerID="9c52f289d4d35a7ab538719f9c3c5eb4d453bb8387f9e697d5cf26704c98cd04" Dec 07 16:22:57 crc kubenswrapper[4716]: I1207 16:22:57.112172 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:22:57 crc kubenswrapper[4716]: I1207 16:22:57.138059 4716 scope.go:117] "RemoveContainer" containerID="dec9f36ef119ff579ca865b84eed4ecbc4e8e9b8f0a6ad017672916770251362" Dec 07 16:22:57 crc kubenswrapper[4716]: I1207 16:22:57.152888 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 07 16:22:57 crc kubenswrapper[4716]: I1207 16:22:57.160506 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 07 16:22:57 crc kubenswrapper[4716]: I1207 16:22:57.173307 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 07 16:22:57 crc kubenswrapper[4716]: I1207 16:22:57.174892 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:22:57 crc kubenswrapper[4716]: I1207 16:22:57.176614 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 07 16:22:57 crc kubenswrapper[4716]: I1207 16:22:57.177448 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 07 16:22:57 crc kubenswrapper[4716]: I1207 16:22:57.177897 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 07 16:22:57 crc kubenswrapper[4716]: I1207 16:22:57.178051 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-b6fbd" Dec 07 16:22:57 crc kubenswrapper[4716]: I1207 16:22:57.178145 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 07 16:22:57 crc kubenswrapper[4716]: I1207 16:22:57.178153 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 07 16:22:57 crc kubenswrapper[4716]: I1207 16:22:57.178440 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 07 16:22:57 crc kubenswrapper[4716]: I1207 16:22:57.190218 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 07 16:22:57 crc kubenswrapper[4716]: I1207 16:22:57.239861 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 07 16:22:57 crc kubenswrapper[4716]: I1207 16:22:57.283743 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/bb3eb7dd-3210-452a-adf4-92bb2eb2447f-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"bb3eb7dd-3210-452a-adf4-92bb2eb2447f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:22:57 crc kubenswrapper[4716]: I1207 16:22:57.284174 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/bb3eb7dd-3210-452a-adf4-92bb2eb2447f-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"bb3eb7dd-3210-452a-adf4-92bb2eb2447f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:22:57 crc kubenswrapper[4716]: I1207 16:22:57.284199 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"bb3eb7dd-3210-452a-adf4-92bb2eb2447f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:22:57 crc kubenswrapper[4716]: I1207 16:22:57.284215 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/bb3eb7dd-3210-452a-adf4-92bb2eb2447f-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"bb3eb7dd-3210-452a-adf4-92bb2eb2447f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:22:57 crc kubenswrapper[4716]: I1207 16:22:57.284231 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/bb3eb7dd-3210-452a-adf4-92bb2eb2447f-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"bb3eb7dd-3210-452a-adf4-92bb2eb2447f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:22:57 crc kubenswrapper[4716]: I1207 16:22:57.284251 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/bb3eb7dd-3210-452a-adf4-92bb2eb2447f-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"bb3eb7dd-3210-452a-adf4-92bb2eb2447f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:22:57 crc kubenswrapper[4716]: I1207 16:22:57.284289 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/bb3eb7dd-3210-452a-adf4-92bb2eb2447f-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"bb3eb7dd-3210-452a-adf4-92bb2eb2447f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:22:57 crc kubenswrapper[4716]: I1207 16:22:57.284307 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bb3eb7dd-3210-452a-adf4-92bb2eb2447f-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"bb3eb7dd-3210-452a-adf4-92bb2eb2447f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:22:57 crc kubenswrapper[4716]: I1207 16:22:57.284533 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55rbn\" (UniqueName: \"kubernetes.io/projected/bb3eb7dd-3210-452a-adf4-92bb2eb2447f-kube-api-access-55rbn\") pod \"rabbitmq-cell1-server-0\" (UID: \"bb3eb7dd-3210-452a-adf4-92bb2eb2447f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:22:57 crc kubenswrapper[4716]: I1207 16:22:57.284640 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/bb3eb7dd-3210-452a-adf4-92bb2eb2447f-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"bb3eb7dd-3210-452a-adf4-92bb2eb2447f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:22:57 crc kubenswrapper[4716]: I1207 16:22:57.284663 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/bb3eb7dd-3210-452a-adf4-92bb2eb2447f-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"bb3eb7dd-3210-452a-adf4-92bb2eb2447f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:22:57 crc kubenswrapper[4716]: I1207 16:22:57.386737 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/bb3eb7dd-3210-452a-adf4-92bb2eb2447f-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"bb3eb7dd-3210-452a-adf4-92bb2eb2447f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:22:57 crc kubenswrapper[4716]: I1207 16:22:57.387022 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bb3eb7dd-3210-452a-adf4-92bb2eb2447f-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"bb3eb7dd-3210-452a-adf4-92bb2eb2447f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:22:57 crc kubenswrapper[4716]: I1207 16:22:57.387152 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55rbn\" (UniqueName: \"kubernetes.io/projected/bb3eb7dd-3210-452a-adf4-92bb2eb2447f-kube-api-access-55rbn\") pod \"rabbitmq-cell1-server-0\" (UID: \"bb3eb7dd-3210-452a-adf4-92bb2eb2447f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:22:57 crc kubenswrapper[4716]: I1207 16:22:57.387234 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/bb3eb7dd-3210-452a-adf4-92bb2eb2447f-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"bb3eb7dd-3210-452a-adf4-92bb2eb2447f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:22:57 crc kubenswrapper[4716]: I1207 16:22:57.387301 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/bb3eb7dd-3210-452a-adf4-92bb2eb2447f-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"bb3eb7dd-3210-452a-adf4-92bb2eb2447f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:22:57 crc kubenswrapper[4716]: I1207 16:22:57.387420 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/bb3eb7dd-3210-452a-adf4-92bb2eb2447f-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"bb3eb7dd-3210-452a-adf4-92bb2eb2447f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:22:57 crc kubenswrapper[4716]: I1207 16:22:57.387319 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/bb3eb7dd-3210-452a-adf4-92bb2eb2447f-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"bb3eb7dd-3210-452a-adf4-92bb2eb2447f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:22:57 crc kubenswrapper[4716]: I1207 16:22:57.387630 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/bb3eb7dd-3210-452a-adf4-92bb2eb2447f-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"bb3eb7dd-3210-452a-adf4-92bb2eb2447f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:22:57 crc kubenswrapper[4716]: I1207 16:22:57.387719 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"bb3eb7dd-3210-452a-adf4-92bb2eb2447f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:22:57 crc kubenswrapper[4716]: I1207 16:22:57.387798 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/bb3eb7dd-3210-452a-adf4-92bb2eb2447f-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"bb3eb7dd-3210-452a-adf4-92bb2eb2447f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:22:57 crc kubenswrapper[4716]: I1207 16:22:57.387879 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/bb3eb7dd-3210-452a-adf4-92bb2eb2447f-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"bb3eb7dd-3210-452a-adf4-92bb2eb2447f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:22:57 crc kubenswrapper[4716]: I1207 16:22:57.387817 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bb3eb7dd-3210-452a-adf4-92bb2eb2447f-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"bb3eb7dd-3210-452a-adf4-92bb2eb2447f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:22:57 crc kubenswrapper[4716]: I1207 16:22:57.388018 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/bb3eb7dd-3210-452a-adf4-92bb2eb2447f-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"bb3eb7dd-3210-452a-adf4-92bb2eb2447f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:22:57 crc kubenswrapper[4716]: I1207 16:22:57.388205 4716 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"bb3eb7dd-3210-452a-adf4-92bb2eb2447f\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:22:57 crc kubenswrapper[4716]: I1207 16:22:57.388514 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/bb3eb7dd-3210-452a-adf4-92bb2eb2447f-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"bb3eb7dd-3210-452a-adf4-92bb2eb2447f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:22:57 crc kubenswrapper[4716]: I1207 16:22:57.388757 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/bb3eb7dd-3210-452a-adf4-92bb2eb2447f-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"bb3eb7dd-3210-452a-adf4-92bb2eb2447f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:22:57 crc kubenswrapper[4716]: I1207 16:22:57.388799 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/bb3eb7dd-3210-452a-adf4-92bb2eb2447f-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"bb3eb7dd-3210-452a-adf4-92bb2eb2447f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:22:57 crc kubenswrapper[4716]: I1207 16:22:57.391545 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/bb3eb7dd-3210-452a-adf4-92bb2eb2447f-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"bb3eb7dd-3210-452a-adf4-92bb2eb2447f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:22:57 crc kubenswrapper[4716]: I1207 16:22:57.391928 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/bb3eb7dd-3210-452a-adf4-92bb2eb2447f-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"bb3eb7dd-3210-452a-adf4-92bb2eb2447f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:22:57 crc kubenswrapper[4716]: I1207 16:22:57.392560 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/bb3eb7dd-3210-452a-adf4-92bb2eb2447f-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"bb3eb7dd-3210-452a-adf4-92bb2eb2447f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:22:57 crc kubenswrapper[4716]: I1207 16:22:57.393592 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/bb3eb7dd-3210-452a-adf4-92bb2eb2447f-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"bb3eb7dd-3210-452a-adf4-92bb2eb2447f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:22:57 crc kubenswrapper[4716]: I1207 16:22:57.414109 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55rbn\" (UniqueName: \"kubernetes.io/projected/bb3eb7dd-3210-452a-adf4-92bb2eb2447f-kube-api-access-55rbn\") pod \"rabbitmq-cell1-server-0\" (UID: \"bb3eb7dd-3210-452a-adf4-92bb2eb2447f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:22:57 crc kubenswrapper[4716]: I1207 16:22:57.420311 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"bb3eb7dd-3210-452a-adf4-92bb2eb2447f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:22:57 crc kubenswrapper[4716]: I1207 16:22:57.494364 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:22:57 crc kubenswrapper[4716]: I1207 16:22:57.677702 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe" path="/var/lib/kubelet/pods/9ca0dfec-a8e3-40d8-9d66-6269a6aa42fe/volumes" Dec 07 16:22:57 crc kubenswrapper[4716]: I1207 16:22:57.678736 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f653df87-c234-4cb6-8ee8-6b16b4fe0044" path="/var/lib/kubelet/pods/f653df87-c234-4cb6-8ee8-6b16b4fe0044/volumes" Dec 07 16:22:57 crc kubenswrapper[4716]: W1207 16:22:57.769724 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbb3eb7dd_3210_452a_adf4_92bb2eb2447f.slice/crio-462127c9f179ca26560b2a579d1b58e2e1f45c94455acd132f447b17e8f55938 WatchSource:0}: Error finding container 462127c9f179ca26560b2a579d1b58e2e1f45c94455acd132f447b17e8f55938: Status 404 returned error can't find the container with id 462127c9f179ca26560b2a579d1b58e2e1f45c94455acd132f447b17e8f55938 Dec 07 16:22:57 crc kubenswrapper[4716]: I1207 16:22:57.772415 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 07 16:22:58 crc kubenswrapper[4716]: I1207 16:22:58.124892 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"db1d1fd3-b1ed-472b-b615-b866706b28a9","Type":"ContainerStarted","Data":"c371f34e2591288ef5624558ee3ac345e5a4e4dca69bf14cd31294edaf221ab9"} Dec 07 16:22:58 crc kubenswrapper[4716]: I1207 16:22:58.126019 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"bb3eb7dd-3210-452a-adf4-92bb2eb2447f","Type":"ContainerStarted","Data":"462127c9f179ca26560b2a579d1b58e2e1f45c94455acd132f447b17e8f55938"} Dec 07 16:22:59 crc kubenswrapper[4716]: I1207 16:22:59.141323 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"db1d1fd3-b1ed-472b-b615-b866706b28a9","Type":"ContainerStarted","Data":"7db886fb8c881a9f3f04298d3519da3ffbc96ea77ea4788570f9ff764e3088ed"} Dec 07 16:22:59 crc kubenswrapper[4716]: I1207 16:22:59.444387 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-qscrm"] Dec 07 16:22:59 crc kubenswrapper[4716]: I1207 16:22:59.446456 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-qscrm" Dec 07 16:22:59 crc kubenswrapper[4716]: I1207 16:22:59.449409 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Dec 07 16:22:59 crc kubenswrapper[4716]: I1207 16:22:59.513859 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-qscrm"] Dec 07 16:22:59 crc kubenswrapper[4716]: I1207 16:22:59.526118 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a774c86e-8027-4a09-be96-b118f72d7c58-ovsdbserver-sb\") pod \"dnsmasq-dns-79bd4cc8c9-qscrm\" (UID: \"a774c86e-8027-4a09-be96-b118f72d7c58\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-qscrm" Dec 07 16:22:59 crc kubenswrapper[4716]: I1207 16:22:59.526180 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a774c86e-8027-4a09-be96-b118f72d7c58-dns-svc\") pod \"dnsmasq-dns-79bd4cc8c9-qscrm\" (UID: \"a774c86e-8027-4a09-be96-b118f72d7c58\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-qscrm" Dec 07 16:22:59 crc kubenswrapper[4716]: I1207 16:22:59.526297 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a774c86e-8027-4a09-be96-b118f72d7c58-dns-swift-storage-0\") pod \"dnsmasq-dns-79bd4cc8c9-qscrm\" (UID: \"a774c86e-8027-4a09-be96-b118f72d7c58\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-qscrm" Dec 07 16:22:59 crc kubenswrapper[4716]: I1207 16:22:59.526333 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/a774c86e-8027-4a09-be96-b118f72d7c58-openstack-edpm-ipam\") pod \"dnsmasq-dns-79bd4cc8c9-qscrm\" (UID: \"a774c86e-8027-4a09-be96-b118f72d7c58\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-qscrm" Dec 07 16:22:59 crc kubenswrapper[4716]: I1207 16:22:59.526396 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a774c86e-8027-4a09-be96-b118f72d7c58-ovsdbserver-nb\") pod \"dnsmasq-dns-79bd4cc8c9-qscrm\" (UID: \"a774c86e-8027-4a09-be96-b118f72d7c58\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-qscrm" Dec 07 16:22:59 crc kubenswrapper[4716]: I1207 16:22:59.526535 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a774c86e-8027-4a09-be96-b118f72d7c58-config\") pod \"dnsmasq-dns-79bd4cc8c9-qscrm\" (UID: \"a774c86e-8027-4a09-be96-b118f72d7c58\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-qscrm" Dec 07 16:22:59 crc kubenswrapper[4716]: I1207 16:22:59.526585 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7td76\" (UniqueName: \"kubernetes.io/projected/a774c86e-8027-4a09-be96-b118f72d7c58-kube-api-access-7td76\") pod \"dnsmasq-dns-79bd4cc8c9-qscrm\" (UID: \"a774c86e-8027-4a09-be96-b118f72d7c58\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-qscrm" Dec 07 16:22:59 crc kubenswrapper[4716]: I1207 16:22:59.628497 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/a774c86e-8027-4a09-be96-b118f72d7c58-openstack-edpm-ipam\") pod \"dnsmasq-dns-79bd4cc8c9-qscrm\" (UID: \"a774c86e-8027-4a09-be96-b118f72d7c58\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-qscrm" Dec 07 16:22:59 crc kubenswrapper[4716]: I1207 16:22:59.628597 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a774c86e-8027-4a09-be96-b118f72d7c58-ovsdbserver-nb\") pod \"dnsmasq-dns-79bd4cc8c9-qscrm\" (UID: \"a774c86e-8027-4a09-be96-b118f72d7c58\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-qscrm" Dec 07 16:22:59 crc kubenswrapper[4716]: I1207 16:22:59.628641 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a774c86e-8027-4a09-be96-b118f72d7c58-config\") pod \"dnsmasq-dns-79bd4cc8c9-qscrm\" (UID: \"a774c86e-8027-4a09-be96-b118f72d7c58\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-qscrm" Dec 07 16:22:59 crc kubenswrapper[4716]: I1207 16:22:59.628661 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7td76\" (UniqueName: \"kubernetes.io/projected/a774c86e-8027-4a09-be96-b118f72d7c58-kube-api-access-7td76\") pod \"dnsmasq-dns-79bd4cc8c9-qscrm\" (UID: \"a774c86e-8027-4a09-be96-b118f72d7c58\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-qscrm" Dec 07 16:22:59 crc kubenswrapper[4716]: I1207 16:22:59.629550 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/a774c86e-8027-4a09-be96-b118f72d7c58-openstack-edpm-ipam\") pod \"dnsmasq-dns-79bd4cc8c9-qscrm\" (UID: \"a774c86e-8027-4a09-be96-b118f72d7c58\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-qscrm" Dec 07 16:22:59 crc kubenswrapper[4716]: I1207 16:22:59.629572 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a774c86e-8027-4a09-be96-b118f72d7c58-config\") pod \"dnsmasq-dns-79bd4cc8c9-qscrm\" (UID: \"a774c86e-8027-4a09-be96-b118f72d7c58\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-qscrm" Dec 07 16:22:59 crc kubenswrapper[4716]: I1207 16:22:59.629590 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a774c86e-8027-4a09-be96-b118f72d7c58-ovsdbserver-nb\") pod \"dnsmasq-dns-79bd4cc8c9-qscrm\" (UID: \"a774c86e-8027-4a09-be96-b118f72d7c58\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-qscrm" Dec 07 16:22:59 crc kubenswrapper[4716]: I1207 16:22:59.629701 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a774c86e-8027-4a09-be96-b118f72d7c58-ovsdbserver-sb\") pod \"dnsmasq-dns-79bd4cc8c9-qscrm\" (UID: \"a774c86e-8027-4a09-be96-b118f72d7c58\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-qscrm" Dec 07 16:22:59 crc kubenswrapper[4716]: I1207 16:22:59.629729 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a774c86e-8027-4a09-be96-b118f72d7c58-dns-svc\") pod \"dnsmasq-dns-79bd4cc8c9-qscrm\" (UID: \"a774c86e-8027-4a09-be96-b118f72d7c58\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-qscrm" Dec 07 16:22:59 crc kubenswrapper[4716]: I1207 16:22:59.629805 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a774c86e-8027-4a09-be96-b118f72d7c58-dns-swift-storage-0\") pod \"dnsmasq-dns-79bd4cc8c9-qscrm\" (UID: \"a774c86e-8027-4a09-be96-b118f72d7c58\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-qscrm" Dec 07 16:22:59 crc kubenswrapper[4716]: I1207 16:22:59.630338 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a774c86e-8027-4a09-be96-b118f72d7c58-ovsdbserver-sb\") pod \"dnsmasq-dns-79bd4cc8c9-qscrm\" (UID: \"a774c86e-8027-4a09-be96-b118f72d7c58\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-qscrm" Dec 07 16:22:59 crc kubenswrapper[4716]: I1207 16:22:59.630413 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a774c86e-8027-4a09-be96-b118f72d7c58-dns-svc\") pod \"dnsmasq-dns-79bd4cc8c9-qscrm\" (UID: \"a774c86e-8027-4a09-be96-b118f72d7c58\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-qscrm" Dec 07 16:22:59 crc kubenswrapper[4716]: I1207 16:22:59.630503 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a774c86e-8027-4a09-be96-b118f72d7c58-dns-swift-storage-0\") pod \"dnsmasq-dns-79bd4cc8c9-qscrm\" (UID: \"a774c86e-8027-4a09-be96-b118f72d7c58\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-qscrm" Dec 07 16:22:59 crc kubenswrapper[4716]: I1207 16:22:59.649913 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7td76\" (UniqueName: \"kubernetes.io/projected/a774c86e-8027-4a09-be96-b118f72d7c58-kube-api-access-7td76\") pod \"dnsmasq-dns-79bd4cc8c9-qscrm\" (UID: \"a774c86e-8027-4a09-be96-b118f72d7c58\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-qscrm" Dec 07 16:22:59 crc kubenswrapper[4716]: I1207 16:22:59.765786 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-qscrm" Dec 07 16:23:00 crc kubenswrapper[4716]: I1207 16:23:00.150070 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"bb3eb7dd-3210-452a-adf4-92bb2eb2447f","Type":"ContainerStarted","Data":"0054148bf3cb70e1f6a222929f83e33b437b026c2dba22c404ef804ac2e9b68f"} Dec 07 16:23:00 crc kubenswrapper[4716]: I1207 16:23:00.249896 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-qscrm"] Dec 07 16:23:00 crc kubenswrapper[4716]: W1207 16:23:00.250336 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda774c86e_8027_4a09_be96_b118f72d7c58.slice/crio-d4ba87c5e92f19c2406c8ff35f6834fa35a7717033e9cfa7b6fb3af04cc60583 WatchSource:0}: Error finding container d4ba87c5e92f19c2406c8ff35f6834fa35a7717033e9cfa7b6fb3af04cc60583: Status 404 returned error can't find the container with id d4ba87c5e92f19c2406c8ff35f6834fa35a7717033e9cfa7b6fb3af04cc60583 Dec 07 16:23:01 crc kubenswrapper[4716]: I1207 16:23:01.165295 4716 generic.go:334] "Generic (PLEG): container finished" podID="a774c86e-8027-4a09-be96-b118f72d7c58" containerID="f564b143103fcd9d838dbb270af2d500d49accf630dae34e1f338bf61a42905b" exitCode=0 Dec 07 16:23:01 crc kubenswrapper[4716]: I1207 16:23:01.165360 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-qscrm" event={"ID":"a774c86e-8027-4a09-be96-b118f72d7c58","Type":"ContainerDied","Data":"f564b143103fcd9d838dbb270af2d500d49accf630dae34e1f338bf61a42905b"} Dec 07 16:23:01 crc kubenswrapper[4716]: I1207 16:23:01.165892 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-qscrm" event={"ID":"a774c86e-8027-4a09-be96-b118f72d7c58","Type":"ContainerStarted","Data":"d4ba87c5e92f19c2406c8ff35f6834fa35a7717033e9cfa7b6fb3af04cc60583"} Dec 07 16:23:02 crc kubenswrapper[4716]: I1207 16:23:02.178519 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-qscrm" event={"ID":"a774c86e-8027-4a09-be96-b118f72d7c58","Type":"ContainerStarted","Data":"a64231f45cb930e66c95c94d2105040c7447ad07a0fb6d9763385adda898d169"} Dec 07 16:23:02 crc kubenswrapper[4716]: I1207 16:23:02.179343 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-79bd4cc8c9-qscrm" Dec 07 16:23:02 crc kubenswrapper[4716]: I1207 16:23:02.210899 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-79bd4cc8c9-qscrm" podStartSLOduration=3.210879313 podStartE2EDuration="3.210879313s" podCreationTimestamp="2025-12-07 16:22:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:23:02.200911384 +0000 UTC m=+1244.891196306" watchObservedRunningTime="2025-12-07 16:23:02.210879313 +0000 UTC m=+1244.901164225" Dec 07 16:23:09 crc kubenswrapper[4716]: I1207 16:23:09.768250 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-79bd4cc8c9-qscrm" Dec 07 16:23:09 crc kubenswrapper[4716]: I1207 16:23:09.833029 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-ld7pc"] Dec 07 16:23:09 crc kubenswrapper[4716]: I1207 16:23:09.833391 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-89c5cd4d5-ld7pc" podUID="bb9f2412-53a4-4e98-9c37-93b57e6df8a9" containerName="dnsmasq-dns" containerID="cri-o://96a44c2634479106efbff992e763b0c05a596d0ebc22f57a5e6740f60c32c468" gracePeriod=10 Dec 07 16:23:10 crc kubenswrapper[4716]: I1207 16:23:10.000840 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55478c4467-sw78r"] Dec 07 16:23:10 crc kubenswrapper[4716]: I1207 16:23:10.002338 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55478c4467-sw78r" Dec 07 16:23:10 crc kubenswrapper[4716]: I1207 16:23:10.026419 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55478c4467-sw78r"] Dec 07 16:23:10 crc kubenswrapper[4716]: I1207 16:23:10.142564 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/27293e5d-5955-4ccb-b78b-63433622f073-dns-svc\") pod \"dnsmasq-dns-55478c4467-sw78r\" (UID: \"27293e5d-5955-4ccb-b78b-63433622f073\") " pod="openstack/dnsmasq-dns-55478c4467-sw78r" Dec 07 16:23:10 crc kubenswrapper[4716]: I1207 16:23:10.142912 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/27293e5d-5955-4ccb-b78b-63433622f073-ovsdbserver-sb\") pod \"dnsmasq-dns-55478c4467-sw78r\" (UID: \"27293e5d-5955-4ccb-b78b-63433622f073\") " pod="openstack/dnsmasq-dns-55478c4467-sw78r" Dec 07 16:23:10 crc kubenswrapper[4716]: I1207 16:23:10.142969 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/27293e5d-5955-4ccb-b78b-63433622f073-openstack-edpm-ipam\") pod \"dnsmasq-dns-55478c4467-sw78r\" (UID: \"27293e5d-5955-4ccb-b78b-63433622f073\") " pod="openstack/dnsmasq-dns-55478c4467-sw78r" Dec 07 16:23:10 crc kubenswrapper[4716]: I1207 16:23:10.143014 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/27293e5d-5955-4ccb-b78b-63433622f073-config\") pod \"dnsmasq-dns-55478c4467-sw78r\" (UID: \"27293e5d-5955-4ccb-b78b-63433622f073\") " pod="openstack/dnsmasq-dns-55478c4467-sw78r" Dec 07 16:23:10 crc kubenswrapper[4716]: I1207 16:23:10.143199 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/27293e5d-5955-4ccb-b78b-63433622f073-dns-swift-storage-0\") pod \"dnsmasq-dns-55478c4467-sw78r\" (UID: \"27293e5d-5955-4ccb-b78b-63433622f073\") " pod="openstack/dnsmasq-dns-55478c4467-sw78r" Dec 07 16:23:10 crc kubenswrapper[4716]: I1207 16:23:10.143226 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/27293e5d-5955-4ccb-b78b-63433622f073-ovsdbserver-nb\") pod \"dnsmasq-dns-55478c4467-sw78r\" (UID: \"27293e5d-5955-4ccb-b78b-63433622f073\") " pod="openstack/dnsmasq-dns-55478c4467-sw78r" Dec 07 16:23:10 crc kubenswrapper[4716]: I1207 16:23:10.143253 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xn5ql\" (UniqueName: \"kubernetes.io/projected/27293e5d-5955-4ccb-b78b-63433622f073-kube-api-access-xn5ql\") pod \"dnsmasq-dns-55478c4467-sw78r\" (UID: \"27293e5d-5955-4ccb-b78b-63433622f073\") " pod="openstack/dnsmasq-dns-55478c4467-sw78r" Dec 07 16:23:10 crc kubenswrapper[4716]: I1207 16:23:10.245016 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/27293e5d-5955-4ccb-b78b-63433622f073-dns-swift-storage-0\") pod \"dnsmasq-dns-55478c4467-sw78r\" (UID: \"27293e5d-5955-4ccb-b78b-63433622f073\") " pod="openstack/dnsmasq-dns-55478c4467-sw78r" Dec 07 16:23:10 crc kubenswrapper[4716]: I1207 16:23:10.245069 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/27293e5d-5955-4ccb-b78b-63433622f073-ovsdbserver-nb\") pod \"dnsmasq-dns-55478c4467-sw78r\" (UID: \"27293e5d-5955-4ccb-b78b-63433622f073\") " pod="openstack/dnsmasq-dns-55478c4467-sw78r" Dec 07 16:23:10 crc kubenswrapper[4716]: I1207 16:23:10.245108 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xn5ql\" (UniqueName: \"kubernetes.io/projected/27293e5d-5955-4ccb-b78b-63433622f073-kube-api-access-xn5ql\") pod \"dnsmasq-dns-55478c4467-sw78r\" (UID: \"27293e5d-5955-4ccb-b78b-63433622f073\") " pod="openstack/dnsmasq-dns-55478c4467-sw78r" Dec 07 16:23:10 crc kubenswrapper[4716]: I1207 16:23:10.245158 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/27293e5d-5955-4ccb-b78b-63433622f073-dns-svc\") pod \"dnsmasq-dns-55478c4467-sw78r\" (UID: \"27293e5d-5955-4ccb-b78b-63433622f073\") " pod="openstack/dnsmasq-dns-55478c4467-sw78r" Dec 07 16:23:10 crc kubenswrapper[4716]: I1207 16:23:10.245197 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/27293e5d-5955-4ccb-b78b-63433622f073-ovsdbserver-sb\") pod \"dnsmasq-dns-55478c4467-sw78r\" (UID: \"27293e5d-5955-4ccb-b78b-63433622f073\") " pod="openstack/dnsmasq-dns-55478c4467-sw78r" Dec 07 16:23:10 crc kubenswrapper[4716]: I1207 16:23:10.245225 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/27293e5d-5955-4ccb-b78b-63433622f073-openstack-edpm-ipam\") pod \"dnsmasq-dns-55478c4467-sw78r\" (UID: \"27293e5d-5955-4ccb-b78b-63433622f073\") " pod="openstack/dnsmasq-dns-55478c4467-sw78r" Dec 07 16:23:10 crc kubenswrapper[4716]: I1207 16:23:10.245257 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/27293e5d-5955-4ccb-b78b-63433622f073-config\") pod \"dnsmasq-dns-55478c4467-sw78r\" (UID: \"27293e5d-5955-4ccb-b78b-63433622f073\") " pod="openstack/dnsmasq-dns-55478c4467-sw78r" Dec 07 16:23:10 crc kubenswrapper[4716]: I1207 16:23:10.246339 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/27293e5d-5955-4ccb-b78b-63433622f073-config\") pod \"dnsmasq-dns-55478c4467-sw78r\" (UID: \"27293e5d-5955-4ccb-b78b-63433622f073\") " pod="openstack/dnsmasq-dns-55478c4467-sw78r" Dec 07 16:23:10 crc kubenswrapper[4716]: I1207 16:23:10.246555 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/27293e5d-5955-4ccb-b78b-63433622f073-dns-swift-storage-0\") pod \"dnsmasq-dns-55478c4467-sw78r\" (UID: \"27293e5d-5955-4ccb-b78b-63433622f073\") " pod="openstack/dnsmasq-dns-55478c4467-sw78r" Dec 07 16:23:10 crc kubenswrapper[4716]: I1207 16:23:10.247018 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/27293e5d-5955-4ccb-b78b-63433622f073-ovsdbserver-nb\") pod \"dnsmasq-dns-55478c4467-sw78r\" (UID: \"27293e5d-5955-4ccb-b78b-63433622f073\") " pod="openstack/dnsmasq-dns-55478c4467-sw78r" Dec 07 16:23:10 crc kubenswrapper[4716]: I1207 16:23:10.247064 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/27293e5d-5955-4ccb-b78b-63433622f073-dns-svc\") pod \"dnsmasq-dns-55478c4467-sw78r\" (UID: \"27293e5d-5955-4ccb-b78b-63433622f073\") " pod="openstack/dnsmasq-dns-55478c4467-sw78r" Dec 07 16:23:10 crc kubenswrapper[4716]: I1207 16:23:10.247300 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/27293e5d-5955-4ccb-b78b-63433622f073-ovsdbserver-sb\") pod \"dnsmasq-dns-55478c4467-sw78r\" (UID: \"27293e5d-5955-4ccb-b78b-63433622f073\") " pod="openstack/dnsmasq-dns-55478c4467-sw78r" Dec 07 16:23:10 crc kubenswrapper[4716]: I1207 16:23:10.247737 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/27293e5d-5955-4ccb-b78b-63433622f073-openstack-edpm-ipam\") pod \"dnsmasq-dns-55478c4467-sw78r\" (UID: \"27293e5d-5955-4ccb-b78b-63433622f073\") " pod="openstack/dnsmasq-dns-55478c4467-sw78r" Dec 07 16:23:10 crc kubenswrapper[4716]: I1207 16:23:10.277344 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xn5ql\" (UniqueName: \"kubernetes.io/projected/27293e5d-5955-4ccb-b78b-63433622f073-kube-api-access-xn5ql\") pod \"dnsmasq-dns-55478c4467-sw78r\" (UID: \"27293e5d-5955-4ccb-b78b-63433622f073\") " pod="openstack/dnsmasq-dns-55478c4467-sw78r" Dec 07 16:23:10 crc kubenswrapper[4716]: I1207 16:23:10.281533 4716 generic.go:334] "Generic (PLEG): container finished" podID="bb9f2412-53a4-4e98-9c37-93b57e6df8a9" containerID="96a44c2634479106efbff992e763b0c05a596d0ebc22f57a5e6740f60c32c468" exitCode=0 Dec 07 16:23:10 crc kubenswrapper[4716]: I1207 16:23:10.281574 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-ld7pc" event={"ID":"bb9f2412-53a4-4e98-9c37-93b57e6df8a9","Type":"ContainerDied","Data":"96a44c2634479106efbff992e763b0c05a596d0ebc22f57a5e6740f60c32c468"} Dec 07 16:23:10 crc kubenswrapper[4716]: I1207 16:23:10.281598 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-ld7pc" event={"ID":"bb9f2412-53a4-4e98-9c37-93b57e6df8a9","Type":"ContainerDied","Data":"37db71dde61341ccec072e9a56d289580dfda9719d299b4f41110e55c929e643"} Dec 07 16:23:10 crc kubenswrapper[4716]: I1207 16:23:10.281610 4716 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="37db71dde61341ccec072e9a56d289580dfda9719d299b4f41110e55c929e643" Dec 07 16:23:10 crc kubenswrapper[4716]: I1207 16:23:10.325405 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55478c4467-sw78r" Dec 07 16:23:10 crc kubenswrapper[4716]: I1207 16:23:10.356541 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-ld7pc" Dec 07 16:23:10 crc kubenswrapper[4716]: I1207 16:23:10.550263 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bb9f2412-53a4-4e98-9c37-93b57e6df8a9-ovsdbserver-sb\") pod \"bb9f2412-53a4-4e98-9c37-93b57e6df8a9\" (UID: \"bb9f2412-53a4-4e98-9c37-93b57e6df8a9\") " Dec 07 16:23:10 crc kubenswrapper[4716]: I1207 16:23:10.550598 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bb9f2412-53a4-4e98-9c37-93b57e6df8a9-ovsdbserver-nb\") pod \"bb9f2412-53a4-4e98-9c37-93b57e6df8a9\" (UID: \"bb9f2412-53a4-4e98-9c37-93b57e6df8a9\") " Dec 07 16:23:10 crc kubenswrapper[4716]: I1207 16:23:10.550633 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bb9f2412-53a4-4e98-9c37-93b57e6df8a9-dns-swift-storage-0\") pod \"bb9f2412-53a4-4e98-9c37-93b57e6df8a9\" (UID: \"bb9f2412-53a4-4e98-9c37-93b57e6df8a9\") " Dec 07 16:23:10 crc kubenswrapper[4716]: I1207 16:23:10.550682 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w6ptf\" (UniqueName: \"kubernetes.io/projected/bb9f2412-53a4-4e98-9c37-93b57e6df8a9-kube-api-access-w6ptf\") pod \"bb9f2412-53a4-4e98-9c37-93b57e6df8a9\" (UID: \"bb9f2412-53a4-4e98-9c37-93b57e6df8a9\") " Dec 07 16:23:10 crc kubenswrapper[4716]: I1207 16:23:10.550717 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bb9f2412-53a4-4e98-9c37-93b57e6df8a9-dns-svc\") pod \"bb9f2412-53a4-4e98-9c37-93b57e6df8a9\" (UID: \"bb9f2412-53a4-4e98-9c37-93b57e6df8a9\") " Dec 07 16:23:10 crc kubenswrapper[4716]: I1207 16:23:10.550743 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb9f2412-53a4-4e98-9c37-93b57e6df8a9-config\") pod \"bb9f2412-53a4-4e98-9c37-93b57e6df8a9\" (UID: \"bb9f2412-53a4-4e98-9c37-93b57e6df8a9\") " Dec 07 16:23:10 crc kubenswrapper[4716]: I1207 16:23:10.557037 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb9f2412-53a4-4e98-9c37-93b57e6df8a9-kube-api-access-w6ptf" (OuterVolumeSpecName: "kube-api-access-w6ptf") pod "bb9f2412-53a4-4e98-9c37-93b57e6df8a9" (UID: "bb9f2412-53a4-4e98-9c37-93b57e6df8a9"). InnerVolumeSpecName "kube-api-access-w6ptf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:23:10 crc kubenswrapper[4716]: I1207 16:23:10.605054 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb9f2412-53a4-4e98-9c37-93b57e6df8a9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "bb9f2412-53a4-4e98-9c37-93b57e6df8a9" (UID: "bb9f2412-53a4-4e98-9c37-93b57e6df8a9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:23:10 crc kubenswrapper[4716]: I1207 16:23:10.610882 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb9f2412-53a4-4e98-9c37-93b57e6df8a9-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "bb9f2412-53a4-4e98-9c37-93b57e6df8a9" (UID: "bb9f2412-53a4-4e98-9c37-93b57e6df8a9"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:23:10 crc kubenswrapper[4716]: I1207 16:23:10.613680 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb9f2412-53a4-4e98-9c37-93b57e6df8a9-config" (OuterVolumeSpecName: "config") pod "bb9f2412-53a4-4e98-9c37-93b57e6df8a9" (UID: "bb9f2412-53a4-4e98-9c37-93b57e6df8a9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:23:10 crc kubenswrapper[4716]: I1207 16:23:10.615266 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb9f2412-53a4-4e98-9c37-93b57e6df8a9-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "bb9f2412-53a4-4e98-9c37-93b57e6df8a9" (UID: "bb9f2412-53a4-4e98-9c37-93b57e6df8a9"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:23:10 crc kubenswrapper[4716]: I1207 16:23:10.617236 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb9f2412-53a4-4e98-9c37-93b57e6df8a9-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "bb9f2412-53a4-4e98-9c37-93b57e6df8a9" (UID: "bb9f2412-53a4-4e98-9c37-93b57e6df8a9"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:23:10 crc kubenswrapper[4716]: I1207 16:23:10.653293 4716 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bb9f2412-53a4-4e98-9c37-93b57e6df8a9-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 07 16:23:10 crc kubenswrapper[4716]: I1207 16:23:10.653319 4716 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bb9f2412-53a4-4e98-9c37-93b57e6df8a9-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 07 16:23:10 crc kubenswrapper[4716]: I1207 16:23:10.653328 4716 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bb9f2412-53a4-4e98-9c37-93b57e6df8a9-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 07 16:23:10 crc kubenswrapper[4716]: I1207 16:23:10.653337 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w6ptf\" (UniqueName: \"kubernetes.io/projected/bb9f2412-53a4-4e98-9c37-93b57e6df8a9-kube-api-access-w6ptf\") on node \"crc\" DevicePath \"\"" Dec 07 16:23:10 crc kubenswrapper[4716]: I1207 16:23:10.653348 4716 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bb9f2412-53a4-4e98-9c37-93b57e6df8a9-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 07 16:23:10 crc kubenswrapper[4716]: I1207 16:23:10.653356 4716 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb9f2412-53a4-4e98-9c37-93b57e6df8a9-config\") on node \"crc\" DevicePath \"\"" Dec 07 16:23:10 crc kubenswrapper[4716]: W1207 16:23:10.832785 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod27293e5d_5955_4ccb_b78b_63433622f073.slice/crio-c18cc45ed8caa67db44f7f2b609940d5aab14416d5c01419356bdc4b926f3417 WatchSource:0}: Error finding container c18cc45ed8caa67db44f7f2b609940d5aab14416d5c01419356bdc4b926f3417: Status 404 returned error can't find the container with id c18cc45ed8caa67db44f7f2b609940d5aab14416d5c01419356bdc4b926f3417 Dec 07 16:23:10 crc kubenswrapper[4716]: I1207 16:23:10.834064 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55478c4467-sw78r"] Dec 07 16:23:11 crc kubenswrapper[4716]: I1207 16:23:11.290900 4716 generic.go:334] "Generic (PLEG): container finished" podID="27293e5d-5955-4ccb-b78b-63433622f073" containerID="04ebb976a572149d7592d3997a8d10229b43a366e0e06cc456fbff3a7be4340d" exitCode=0 Dec 07 16:23:11 crc kubenswrapper[4716]: I1207 16:23:11.291178 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55478c4467-sw78r" event={"ID":"27293e5d-5955-4ccb-b78b-63433622f073","Type":"ContainerDied","Data":"04ebb976a572149d7592d3997a8d10229b43a366e0e06cc456fbff3a7be4340d"} Dec 07 16:23:11 crc kubenswrapper[4716]: I1207 16:23:11.291242 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-ld7pc" Dec 07 16:23:11 crc kubenswrapper[4716]: I1207 16:23:11.291243 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55478c4467-sw78r" event={"ID":"27293e5d-5955-4ccb-b78b-63433622f073","Type":"ContainerStarted","Data":"c18cc45ed8caa67db44f7f2b609940d5aab14416d5c01419356bdc4b926f3417"} Dec 07 16:23:11 crc kubenswrapper[4716]: I1207 16:23:11.485135 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-ld7pc"] Dec 07 16:23:11 crc kubenswrapper[4716]: I1207 16:23:11.493454 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-ld7pc"] Dec 07 16:23:11 crc kubenswrapper[4716]: I1207 16:23:11.669274 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb9f2412-53a4-4e98-9c37-93b57e6df8a9" path="/var/lib/kubelet/pods/bb9f2412-53a4-4e98-9c37-93b57e6df8a9/volumes" Dec 07 16:23:12 crc kubenswrapper[4716]: I1207 16:23:12.304109 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55478c4467-sw78r" event={"ID":"27293e5d-5955-4ccb-b78b-63433622f073","Type":"ContainerStarted","Data":"d559cb529da8f2bc2c523906b17ecaaec60b2c611988e4c5093a0e29151ef9cb"} Dec 07 16:23:12 crc kubenswrapper[4716]: I1207 16:23:12.304294 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-55478c4467-sw78r" Dec 07 16:23:12 crc kubenswrapper[4716]: I1207 16:23:12.329727 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-55478c4467-sw78r" podStartSLOduration=3.329708586 podStartE2EDuration="3.329708586s" podCreationTimestamp="2025-12-07 16:23:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:23:12.324261719 +0000 UTC m=+1255.014546631" watchObservedRunningTime="2025-12-07 16:23:12.329708586 +0000 UTC m=+1255.019993488" Dec 07 16:23:20 crc kubenswrapper[4716]: I1207 16:23:20.327169 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-55478c4467-sw78r" Dec 07 16:23:20 crc kubenswrapper[4716]: I1207 16:23:20.407883 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-qscrm"] Dec 07 16:23:20 crc kubenswrapper[4716]: I1207 16:23:20.408209 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-79bd4cc8c9-qscrm" podUID="a774c86e-8027-4a09-be96-b118f72d7c58" containerName="dnsmasq-dns" containerID="cri-o://a64231f45cb930e66c95c94d2105040c7447ad07a0fb6d9763385adda898d169" gracePeriod=10 Dec 07 16:23:20 crc kubenswrapper[4716]: I1207 16:23:20.909519 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-qscrm" Dec 07 16:23:21 crc kubenswrapper[4716]: I1207 16:23:21.029230 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/a774c86e-8027-4a09-be96-b118f72d7c58-openstack-edpm-ipam\") pod \"a774c86e-8027-4a09-be96-b118f72d7c58\" (UID: \"a774c86e-8027-4a09-be96-b118f72d7c58\") " Dec 07 16:23:21 crc kubenswrapper[4716]: I1207 16:23:21.029307 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a774c86e-8027-4a09-be96-b118f72d7c58-dns-svc\") pod \"a774c86e-8027-4a09-be96-b118f72d7c58\" (UID: \"a774c86e-8027-4a09-be96-b118f72d7c58\") " Dec 07 16:23:21 crc kubenswrapper[4716]: I1207 16:23:21.029362 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7td76\" (UniqueName: \"kubernetes.io/projected/a774c86e-8027-4a09-be96-b118f72d7c58-kube-api-access-7td76\") pod \"a774c86e-8027-4a09-be96-b118f72d7c58\" (UID: \"a774c86e-8027-4a09-be96-b118f72d7c58\") " Dec 07 16:23:21 crc kubenswrapper[4716]: I1207 16:23:21.029419 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a774c86e-8027-4a09-be96-b118f72d7c58-dns-swift-storage-0\") pod \"a774c86e-8027-4a09-be96-b118f72d7c58\" (UID: \"a774c86e-8027-4a09-be96-b118f72d7c58\") " Dec 07 16:23:21 crc kubenswrapper[4716]: I1207 16:23:21.029481 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a774c86e-8027-4a09-be96-b118f72d7c58-ovsdbserver-nb\") pod \"a774c86e-8027-4a09-be96-b118f72d7c58\" (UID: \"a774c86e-8027-4a09-be96-b118f72d7c58\") " Dec 07 16:23:21 crc kubenswrapper[4716]: I1207 16:23:21.029578 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a774c86e-8027-4a09-be96-b118f72d7c58-ovsdbserver-sb\") pod \"a774c86e-8027-4a09-be96-b118f72d7c58\" (UID: \"a774c86e-8027-4a09-be96-b118f72d7c58\") " Dec 07 16:23:21 crc kubenswrapper[4716]: I1207 16:23:21.029613 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a774c86e-8027-4a09-be96-b118f72d7c58-config\") pod \"a774c86e-8027-4a09-be96-b118f72d7c58\" (UID: \"a774c86e-8027-4a09-be96-b118f72d7c58\") " Dec 07 16:23:21 crc kubenswrapper[4716]: I1207 16:23:21.039906 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a774c86e-8027-4a09-be96-b118f72d7c58-kube-api-access-7td76" (OuterVolumeSpecName: "kube-api-access-7td76") pod "a774c86e-8027-4a09-be96-b118f72d7c58" (UID: "a774c86e-8027-4a09-be96-b118f72d7c58"). InnerVolumeSpecName "kube-api-access-7td76". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:23:21 crc kubenswrapper[4716]: I1207 16:23:21.082842 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a774c86e-8027-4a09-be96-b118f72d7c58-config" (OuterVolumeSpecName: "config") pod "a774c86e-8027-4a09-be96-b118f72d7c58" (UID: "a774c86e-8027-4a09-be96-b118f72d7c58"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:23:21 crc kubenswrapper[4716]: I1207 16:23:21.085211 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a774c86e-8027-4a09-be96-b118f72d7c58-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "a774c86e-8027-4a09-be96-b118f72d7c58" (UID: "a774c86e-8027-4a09-be96-b118f72d7c58"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:23:21 crc kubenswrapper[4716]: I1207 16:23:21.088295 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a774c86e-8027-4a09-be96-b118f72d7c58-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a774c86e-8027-4a09-be96-b118f72d7c58" (UID: "a774c86e-8027-4a09-be96-b118f72d7c58"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:23:21 crc kubenswrapper[4716]: I1207 16:23:21.097956 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a774c86e-8027-4a09-be96-b118f72d7c58-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a774c86e-8027-4a09-be96-b118f72d7c58" (UID: "a774c86e-8027-4a09-be96-b118f72d7c58"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:23:21 crc kubenswrapper[4716]: I1207 16:23:21.098226 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a774c86e-8027-4a09-be96-b118f72d7c58-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "a774c86e-8027-4a09-be96-b118f72d7c58" (UID: "a774c86e-8027-4a09-be96-b118f72d7c58"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:23:21 crc kubenswrapper[4716]: I1207 16:23:21.102146 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a774c86e-8027-4a09-be96-b118f72d7c58-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a774c86e-8027-4a09-be96-b118f72d7c58" (UID: "a774c86e-8027-4a09-be96-b118f72d7c58"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:23:21 crc kubenswrapper[4716]: I1207 16:23:21.132560 4716 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a774c86e-8027-4a09-be96-b118f72d7c58-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 07 16:23:21 crc kubenswrapper[4716]: I1207 16:23:21.132640 4716 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a774c86e-8027-4a09-be96-b118f72d7c58-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 07 16:23:21 crc kubenswrapper[4716]: I1207 16:23:21.132650 4716 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a774c86e-8027-4a09-be96-b118f72d7c58-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 07 16:23:21 crc kubenswrapper[4716]: I1207 16:23:21.132659 4716 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a774c86e-8027-4a09-be96-b118f72d7c58-config\") on node \"crc\" DevicePath \"\"" Dec 07 16:23:21 crc kubenswrapper[4716]: I1207 16:23:21.132668 4716 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/a774c86e-8027-4a09-be96-b118f72d7c58-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 07 16:23:21 crc kubenswrapper[4716]: I1207 16:23:21.132677 4716 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a774c86e-8027-4a09-be96-b118f72d7c58-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 07 16:23:21 crc kubenswrapper[4716]: I1207 16:23:21.132687 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7td76\" (UniqueName: \"kubernetes.io/projected/a774c86e-8027-4a09-be96-b118f72d7c58-kube-api-access-7td76\") on node \"crc\" DevicePath \"\"" Dec 07 16:23:21 crc kubenswrapper[4716]: I1207 16:23:21.432431 4716 generic.go:334] "Generic (PLEG): container finished" podID="a774c86e-8027-4a09-be96-b118f72d7c58" containerID="a64231f45cb930e66c95c94d2105040c7447ad07a0fb6d9763385adda898d169" exitCode=0 Dec 07 16:23:21 crc kubenswrapper[4716]: I1207 16:23:21.432481 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-qscrm" event={"ID":"a774c86e-8027-4a09-be96-b118f72d7c58","Type":"ContainerDied","Data":"a64231f45cb930e66c95c94d2105040c7447ad07a0fb6d9763385adda898d169"} Dec 07 16:23:21 crc kubenswrapper[4716]: I1207 16:23:21.432513 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-qscrm" event={"ID":"a774c86e-8027-4a09-be96-b118f72d7c58","Type":"ContainerDied","Data":"d4ba87c5e92f19c2406c8ff35f6834fa35a7717033e9cfa7b6fb3af04cc60583"} Dec 07 16:23:21 crc kubenswrapper[4716]: I1207 16:23:21.432531 4716 scope.go:117] "RemoveContainer" containerID="a64231f45cb930e66c95c94d2105040c7447ad07a0fb6d9763385adda898d169" Dec 07 16:23:21 crc kubenswrapper[4716]: I1207 16:23:21.433235 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-qscrm" Dec 07 16:23:21 crc kubenswrapper[4716]: I1207 16:23:21.460319 4716 scope.go:117] "RemoveContainer" containerID="f564b143103fcd9d838dbb270af2d500d49accf630dae34e1f338bf61a42905b" Dec 07 16:23:21 crc kubenswrapper[4716]: I1207 16:23:21.463321 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-qscrm"] Dec 07 16:23:21 crc kubenswrapper[4716]: I1207 16:23:21.471825 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-qscrm"] Dec 07 16:23:21 crc kubenswrapper[4716]: I1207 16:23:21.500428 4716 scope.go:117] "RemoveContainer" containerID="a64231f45cb930e66c95c94d2105040c7447ad07a0fb6d9763385adda898d169" Dec 07 16:23:21 crc kubenswrapper[4716]: E1207 16:23:21.500967 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a64231f45cb930e66c95c94d2105040c7447ad07a0fb6d9763385adda898d169\": container with ID starting with a64231f45cb930e66c95c94d2105040c7447ad07a0fb6d9763385adda898d169 not found: ID does not exist" containerID="a64231f45cb930e66c95c94d2105040c7447ad07a0fb6d9763385adda898d169" Dec 07 16:23:21 crc kubenswrapper[4716]: I1207 16:23:21.501032 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a64231f45cb930e66c95c94d2105040c7447ad07a0fb6d9763385adda898d169"} err="failed to get container status \"a64231f45cb930e66c95c94d2105040c7447ad07a0fb6d9763385adda898d169\": rpc error: code = NotFound desc = could not find container \"a64231f45cb930e66c95c94d2105040c7447ad07a0fb6d9763385adda898d169\": container with ID starting with a64231f45cb930e66c95c94d2105040c7447ad07a0fb6d9763385adda898d169 not found: ID does not exist" Dec 07 16:23:21 crc kubenswrapper[4716]: I1207 16:23:21.501109 4716 scope.go:117] "RemoveContainer" containerID="f564b143103fcd9d838dbb270af2d500d49accf630dae34e1f338bf61a42905b" Dec 07 16:23:21 crc kubenswrapper[4716]: E1207 16:23:21.501645 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f564b143103fcd9d838dbb270af2d500d49accf630dae34e1f338bf61a42905b\": container with ID starting with f564b143103fcd9d838dbb270af2d500d49accf630dae34e1f338bf61a42905b not found: ID does not exist" containerID="f564b143103fcd9d838dbb270af2d500d49accf630dae34e1f338bf61a42905b" Dec 07 16:23:21 crc kubenswrapper[4716]: I1207 16:23:21.501687 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f564b143103fcd9d838dbb270af2d500d49accf630dae34e1f338bf61a42905b"} err="failed to get container status \"f564b143103fcd9d838dbb270af2d500d49accf630dae34e1f338bf61a42905b\": rpc error: code = NotFound desc = could not find container \"f564b143103fcd9d838dbb270af2d500d49accf630dae34e1f338bf61a42905b\": container with ID starting with f564b143103fcd9d838dbb270af2d500d49accf630dae34e1f338bf61a42905b not found: ID does not exist" Dec 07 16:23:21 crc kubenswrapper[4716]: I1207 16:23:21.669824 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a774c86e-8027-4a09-be96-b118f72d7c58" path="/var/lib/kubelet/pods/a774c86e-8027-4a09-be96-b118f72d7c58/volumes" Dec 07 16:23:22 crc kubenswrapper[4716]: I1207 16:23:22.760868 4716 patch_prober.go:28] interesting pod/machine-config-daemon-zcxxp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 16:23:22 crc kubenswrapper[4716]: I1207 16:23:22.760935 4716 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 16:23:22 crc kubenswrapper[4716]: I1207 16:23:22.760999 4716 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" Dec 07 16:23:22 crc kubenswrapper[4716]: I1207 16:23:22.762035 4716 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"43e1cbd05079c8cdd0bd98789bfa5b92ef2c8c3e87845aeb18f274ab5529de34"} pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 07 16:23:22 crc kubenswrapper[4716]: I1207 16:23:22.762161 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" containerName="machine-config-daemon" containerID="cri-o://43e1cbd05079c8cdd0bd98789bfa5b92ef2c8c3e87845aeb18f274ab5529de34" gracePeriod=600 Dec 07 16:23:23 crc kubenswrapper[4716]: I1207 16:23:23.452523 4716 generic.go:334] "Generic (PLEG): container finished" podID="c15b59eb-565d-4556-a4ce-75afdf159dc8" containerID="43e1cbd05079c8cdd0bd98789bfa5b92ef2c8c3e87845aeb18f274ab5529de34" exitCode=0 Dec 07 16:23:23 crc kubenswrapper[4716]: I1207 16:23:23.452598 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" event={"ID":"c15b59eb-565d-4556-a4ce-75afdf159dc8","Type":"ContainerDied","Data":"43e1cbd05079c8cdd0bd98789bfa5b92ef2c8c3e87845aeb18f274ab5529de34"} Dec 07 16:23:23 crc kubenswrapper[4716]: I1207 16:23:23.453108 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" event={"ID":"c15b59eb-565d-4556-a4ce-75afdf159dc8","Type":"ContainerStarted","Data":"fea4ce7ae8e17d1efab423015c4c8670e9be41f20a55cff5db1bae1736f619e0"} Dec 07 16:23:23 crc kubenswrapper[4716]: I1207 16:23:23.453133 4716 scope.go:117] "RemoveContainer" containerID="73e652154907931afdb244fb07e6dd85483e6b06afe1a76ebcc2b8bf8b2c7310" Dec 07 16:23:31 crc kubenswrapper[4716]: I1207 16:23:31.538727 4716 generic.go:334] "Generic (PLEG): container finished" podID="db1d1fd3-b1ed-472b-b615-b866706b28a9" containerID="7db886fb8c881a9f3f04298d3519da3ffbc96ea77ea4788570f9ff764e3088ed" exitCode=0 Dec 07 16:23:31 crc kubenswrapper[4716]: I1207 16:23:31.538821 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"db1d1fd3-b1ed-472b-b615-b866706b28a9","Type":"ContainerDied","Data":"7db886fb8c881a9f3f04298d3519da3ffbc96ea77ea4788570f9ff764e3088ed"} Dec 07 16:23:31 crc kubenswrapper[4716]: I1207 16:23:31.540585 4716 generic.go:334] "Generic (PLEG): container finished" podID="bb3eb7dd-3210-452a-adf4-92bb2eb2447f" containerID="0054148bf3cb70e1f6a222929f83e33b437b026c2dba22c404ef804ac2e9b68f" exitCode=0 Dec 07 16:23:31 crc kubenswrapper[4716]: I1207 16:23:31.540608 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"bb3eb7dd-3210-452a-adf4-92bb2eb2447f","Type":"ContainerDied","Data":"0054148bf3cb70e1f6a222929f83e33b437b026c2dba22c404ef804ac2e9b68f"} Dec 07 16:23:32 crc kubenswrapper[4716]: I1207 16:23:32.550836 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"bb3eb7dd-3210-452a-adf4-92bb2eb2447f","Type":"ContainerStarted","Data":"8fe0dec494348dc82f03bc9ff2ec0d478a740a26321bb5c6df2bd4e9aec8ac73"} Dec 07 16:23:32 crc kubenswrapper[4716]: I1207 16:23:32.552491 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:23:32 crc kubenswrapper[4716]: I1207 16:23:32.555284 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"db1d1fd3-b1ed-472b-b615-b866706b28a9","Type":"ContainerStarted","Data":"ed9ba8e6d542592f43c9fbc684186207d7054fb78c0b1b54cb49a063cf9e2181"} Dec 07 16:23:32 crc kubenswrapper[4716]: I1207 16:23:32.556225 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 07 16:23:32 crc kubenswrapper[4716]: I1207 16:23:32.594669 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=35.594640045 podStartE2EDuration="35.594640045s" podCreationTimestamp="2025-12-07 16:22:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:23:32.579531698 +0000 UTC m=+1275.269816610" watchObservedRunningTime="2025-12-07 16:23:32.594640045 +0000 UTC m=+1275.284924967" Dec 07 16:23:34 crc kubenswrapper[4716]: I1207 16:23:34.021980 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=38.021961629 podStartE2EDuration="38.021961629s" podCreationTimestamp="2025-12-07 16:22:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 16:23:32.61453345 +0000 UTC m=+1275.304818362" watchObservedRunningTime="2025-12-07 16:23:34.021961629 +0000 UTC m=+1276.712246541" Dec 07 16:23:34 crc kubenswrapper[4716]: I1207 16:23:34.025512 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-dv7q9"] Dec 07 16:23:34 crc kubenswrapper[4716]: E1207 16:23:34.025864 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb9f2412-53a4-4e98-9c37-93b57e6df8a9" containerName="init" Dec 07 16:23:34 crc kubenswrapper[4716]: I1207 16:23:34.025878 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb9f2412-53a4-4e98-9c37-93b57e6df8a9" containerName="init" Dec 07 16:23:34 crc kubenswrapper[4716]: E1207 16:23:34.025895 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb9f2412-53a4-4e98-9c37-93b57e6df8a9" containerName="dnsmasq-dns" Dec 07 16:23:34 crc kubenswrapper[4716]: I1207 16:23:34.025900 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb9f2412-53a4-4e98-9c37-93b57e6df8a9" containerName="dnsmasq-dns" Dec 07 16:23:34 crc kubenswrapper[4716]: E1207 16:23:34.025917 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a774c86e-8027-4a09-be96-b118f72d7c58" containerName="init" Dec 07 16:23:34 crc kubenswrapper[4716]: I1207 16:23:34.025923 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="a774c86e-8027-4a09-be96-b118f72d7c58" containerName="init" Dec 07 16:23:34 crc kubenswrapper[4716]: E1207 16:23:34.025935 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a774c86e-8027-4a09-be96-b118f72d7c58" containerName="dnsmasq-dns" Dec 07 16:23:34 crc kubenswrapper[4716]: I1207 16:23:34.025942 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="a774c86e-8027-4a09-be96-b118f72d7c58" containerName="dnsmasq-dns" Dec 07 16:23:34 crc kubenswrapper[4716]: I1207 16:23:34.026168 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb9f2412-53a4-4e98-9c37-93b57e6df8a9" containerName="dnsmasq-dns" Dec 07 16:23:34 crc kubenswrapper[4716]: I1207 16:23:34.026181 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="a774c86e-8027-4a09-be96-b118f72d7c58" containerName="dnsmasq-dns" Dec 07 16:23:34 crc kubenswrapper[4716]: I1207 16:23:34.026738 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-dv7q9" Dec 07 16:23:34 crc kubenswrapper[4716]: I1207 16:23:34.030838 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-h9pt2" Dec 07 16:23:34 crc kubenswrapper[4716]: I1207 16:23:34.030934 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 07 16:23:34 crc kubenswrapper[4716]: I1207 16:23:34.030959 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 07 16:23:34 crc kubenswrapper[4716]: I1207 16:23:34.031375 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 07 16:23:34 crc kubenswrapper[4716]: I1207 16:23:34.038011 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-dv7q9"] Dec 07 16:23:34 crc kubenswrapper[4716]: I1207 16:23:34.137233 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2c9bedba-2a2e-4e68-839a-646d5747911c-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-dv7q9\" (UID: \"2c9bedba-2a2e-4e68-839a-646d5747911c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-dv7q9" Dec 07 16:23:34 crc kubenswrapper[4716]: I1207 16:23:34.137495 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c9bedba-2a2e-4e68-839a-646d5747911c-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-dv7q9\" (UID: \"2c9bedba-2a2e-4e68-839a-646d5747911c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-dv7q9" Dec 07 16:23:34 crc kubenswrapper[4716]: I1207 16:23:34.137692 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2c9bedba-2a2e-4e68-839a-646d5747911c-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-dv7q9\" (UID: \"2c9bedba-2a2e-4e68-839a-646d5747911c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-dv7q9" Dec 07 16:23:34 crc kubenswrapper[4716]: I1207 16:23:34.137812 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqw5j\" (UniqueName: \"kubernetes.io/projected/2c9bedba-2a2e-4e68-839a-646d5747911c-kube-api-access-kqw5j\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-dv7q9\" (UID: \"2c9bedba-2a2e-4e68-839a-646d5747911c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-dv7q9" Dec 07 16:23:34 crc kubenswrapper[4716]: I1207 16:23:34.239864 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2c9bedba-2a2e-4e68-839a-646d5747911c-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-dv7q9\" (UID: \"2c9bedba-2a2e-4e68-839a-646d5747911c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-dv7q9" Dec 07 16:23:34 crc kubenswrapper[4716]: I1207 16:23:34.240181 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqw5j\" (UniqueName: \"kubernetes.io/projected/2c9bedba-2a2e-4e68-839a-646d5747911c-kube-api-access-kqw5j\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-dv7q9\" (UID: \"2c9bedba-2a2e-4e68-839a-646d5747911c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-dv7q9" Dec 07 16:23:34 crc kubenswrapper[4716]: I1207 16:23:34.240350 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2c9bedba-2a2e-4e68-839a-646d5747911c-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-dv7q9\" (UID: \"2c9bedba-2a2e-4e68-839a-646d5747911c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-dv7q9" Dec 07 16:23:34 crc kubenswrapper[4716]: I1207 16:23:34.240500 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c9bedba-2a2e-4e68-839a-646d5747911c-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-dv7q9\" (UID: \"2c9bedba-2a2e-4e68-839a-646d5747911c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-dv7q9" Dec 07 16:23:34 crc kubenswrapper[4716]: I1207 16:23:34.246246 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2c9bedba-2a2e-4e68-839a-646d5747911c-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-dv7q9\" (UID: \"2c9bedba-2a2e-4e68-839a-646d5747911c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-dv7q9" Dec 07 16:23:34 crc kubenswrapper[4716]: I1207 16:23:34.246253 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2c9bedba-2a2e-4e68-839a-646d5747911c-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-dv7q9\" (UID: \"2c9bedba-2a2e-4e68-839a-646d5747911c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-dv7q9" Dec 07 16:23:34 crc kubenswrapper[4716]: I1207 16:23:34.246846 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c9bedba-2a2e-4e68-839a-646d5747911c-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-dv7q9\" (UID: \"2c9bedba-2a2e-4e68-839a-646d5747911c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-dv7q9" Dec 07 16:23:34 crc kubenswrapper[4716]: I1207 16:23:34.256975 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqw5j\" (UniqueName: \"kubernetes.io/projected/2c9bedba-2a2e-4e68-839a-646d5747911c-kube-api-access-kqw5j\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-dv7q9\" (UID: \"2c9bedba-2a2e-4e68-839a-646d5747911c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-dv7q9" Dec 07 16:23:34 crc kubenswrapper[4716]: I1207 16:23:34.345267 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-dv7q9" Dec 07 16:23:34 crc kubenswrapper[4716]: I1207 16:23:34.799427 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-dv7q9"] Dec 07 16:23:34 crc kubenswrapper[4716]: W1207 16:23:34.800161 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2c9bedba_2a2e_4e68_839a_646d5747911c.slice/crio-147acfc6d3d1fbdff36afcf06a3a6b0f109d40d21166a350b65e0372b01dfb45 WatchSource:0}: Error finding container 147acfc6d3d1fbdff36afcf06a3a6b0f109d40d21166a350b65e0372b01dfb45: Status 404 returned error can't find the container with id 147acfc6d3d1fbdff36afcf06a3a6b0f109d40d21166a350b65e0372b01dfb45 Dec 07 16:23:34 crc kubenswrapper[4716]: I1207 16:23:34.802407 4716 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 07 16:23:35 crc kubenswrapper[4716]: I1207 16:23:35.593323 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-dv7q9" event={"ID":"2c9bedba-2a2e-4e68-839a-646d5747911c","Type":"ContainerStarted","Data":"147acfc6d3d1fbdff36afcf06a3a6b0f109d40d21166a350b65e0372b01dfb45"} Dec 07 16:23:44 crc kubenswrapper[4716]: I1207 16:23:44.714572 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-dv7q9" event={"ID":"2c9bedba-2a2e-4e68-839a-646d5747911c","Type":"ContainerStarted","Data":"455aa7ed1422c1a12f17a5b1229f0d0b96399bbf68f40552810e7724637f4759"} Dec 07 16:23:44 crc kubenswrapper[4716]: I1207 16:23:44.743696 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-dv7q9" podStartSLOduration=1.56442406 podStartE2EDuration="10.743679185s" podCreationTimestamp="2025-12-07 16:23:34 +0000 UTC" firstStartedPulling="2025-12-07 16:23:34.802158381 +0000 UTC m=+1277.492443293" lastFinishedPulling="2025-12-07 16:23:43.981413486 +0000 UTC m=+1286.671698418" observedRunningTime="2025-12-07 16:23:44.733811671 +0000 UTC m=+1287.424096603" watchObservedRunningTime="2025-12-07 16:23:44.743679185 +0000 UTC m=+1287.433964117" Dec 07 16:23:46 crc kubenswrapper[4716]: I1207 16:23:46.775285 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 07 16:23:47 crc kubenswrapper[4716]: I1207 16:23:47.496285 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 07 16:23:55 crc kubenswrapper[4716]: I1207 16:23:55.819332 4716 generic.go:334] "Generic (PLEG): container finished" podID="2c9bedba-2a2e-4e68-839a-646d5747911c" containerID="455aa7ed1422c1a12f17a5b1229f0d0b96399bbf68f40552810e7724637f4759" exitCode=0 Dec 07 16:23:55 crc kubenswrapper[4716]: I1207 16:23:55.819412 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-dv7q9" event={"ID":"2c9bedba-2a2e-4e68-839a-646d5747911c","Type":"ContainerDied","Data":"455aa7ed1422c1a12f17a5b1229f0d0b96399bbf68f40552810e7724637f4759"} Dec 07 16:23:57 crc kubenswrapper[4716]: I1207 16:23:57.277003 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-dv7q9" Dec 07 16:23:57 crc kubenswrapper[4716]: I1207 16:23:57.475536 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kqw5j\" (UniqueName: \"kubernetes.io/projected/2c9bedba-2a2e-4e68-839a-646d5747911c-kube-api-access-kqw5j\") pod \"2c9bedba-2a2e-4e68-839a-646d5747911c\" (UID: \"2c9bedba-2a2e-4e68-839a-646d5747911c\") " Dec 07 16:23:57 crc kubenswrapper[4716]: I1207 16:23:57.475602 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2c9bedba-2a2e-4e68-839a-646d5747911c-ssh-key\") pod \"2c9bedba-2a2e-4e68-839a-646d5747911c\" (UID: \"2c9bedba-2a2e-4e68-839a-646d5747911c\") " Dec 07 16:23:57 crc kubenswrapper[4716]: I1207 16:23:57.475643 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c9bedba-2a2e-4e68-839a-646d5747911c-repo-setup-combined-ca-bundle\") pod \"2c9bedba-2a2e-4e68-839a-646d5747911c\" (UID: \"2c9bedba-2a2e-4e68-839a-646d5747911c\") " Dec 07 16:23:57 crc kubenswrapper[4716]: I1207 16:23:57.475671 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2c9bedba-2a2e-4e68-839a-646d5747911c-inventory\") pod \"2c9bedba-2a2e-4e68-839a-646d5747911c\" (UID: \"2c9bedba-2a2e-4e68-839a-646d5747911c\") " Dec 07 16:23:57 crc kubenswrapper[4716]: I1207 16:23:57.481735 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c9bedba-2a2e-4e68-839a-646d5747911c-kube-api-access-kqw5j" (OuterVolumeSpecName: "kube-api-access-kqw5j") pod "2c9bedba-2a2e-4e68-839a-646d5747911c" (UID: "2c9bedba-2a2e-4e68-839a-646d5747911c"). InnerVolumeSpecName "kube-api-access-kqw5j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:23:57 crc kubenswrapper[4716]: I1207 16:23:57.482447 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c9bedba-2a2e-4e68-839a-646d5747911c-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "2c9bedba-2a2e-4e68-839a-646d5747911c" (UID: "2c9bedba-2a2e-4e68-839a-646d5747911c"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:23:57 crc kubenswrapper[4716]: I1207 16:23:57.511266 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c9bedba-2a2e-4e68-839a-646d5747911c-inventory" (OuterVolumeSpecName: "inventory") pod "2c9bedba-2a2e-4e68-839a-646d5747911c" (UID: "2c9bedba-2a2e-4e68-839a-646d5747911c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:23:57 crc kubenswrapper[4716]: I1207 16:23:57.511660 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c9bedba-2a2e-4e68-839a-646d5747911c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2c9bedba-2a2e-4e68-839a-646d5747911c" (UID: "2c9bedba-2a2e-4e68-839a-646d5747911c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:23:57 crc kubenswrapper[4716]: I1207 16:23:57.577714 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kqw5j\" (UniqueName: \"kubernetes.io/projected/2c9bedba-2a2e-4e68-839a-646d5747911c-kube-api-access-kqw5j\") on node \"crc\" DevicePath \"\"" Dec 07 16:23:57 crc kubenswrapper[4716]: I1207 16:23:57.577754 4716 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2c9bedba-2a2e-4e68-839a-646d5747911c-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 07 16:23:57 crc kubenswrapper[4716]: I1207 16:23:57.577764 4716 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c9bedba-2a2e-4e68-839a-646d5747911c-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 16:23:57 crc kubenswrapper[4716]: I1207 16:23:57.577773 4716 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2c9bedba-2a2e-4e68-839a-646d5747911c-inventory\") on node \"crc\" DevicePath \"\"" Dec 07 16:23:57 crc kubenswrapper[4716]: I1207 16:23:57.843169 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-dv7q9" event={"ID":"2c9bedba-2a2e-4e68-839a-646d5747911c","Type":"ContainerDied","Data":"147acfc6d3d1fbdff36afcf06a3a6b0f109d40d21166a350b65e0372b01dfb45"} Dec 07 16:23:57 crc kubenswrapper[4716]: I1207 16:23:57.843412 4716 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="147acfc6d3d1fbdff36afcf06a3a6b0f109d40d21166a350b65e0372b01dfb45" Dec 07 16:23:57 crc kubenswrapper[4716]: I1207 16:23:57.843222 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-dv7q9" Dec 07 16:23:57 crc kubenswrapper[4716]: I1207 16:23:57.915593 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-l5phd"] Dec 07 16:23:57 crc kubenswrapper[4716]: E1207 16:23:57.916007 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c9bedba-2a2e-4e68-839a-646d5747911c" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 07 16:23:57 crc kubenswrapper[4716]: I1207 16:23:57.916026 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c9bedba-2a2e-4e68-839a-646d5747911c" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 07 16:23:57 crc kubenswrapper[4716]: I1207 16:23:57.916239 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c9bedba-2a2e-4e68-839a-646d5747911c" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 07 16:23:57 crc kubenswrapper[4716]: I1207 16:23:57.918254 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-l5phd" Dec 07 16:23:57 crc kubenswrapper[4716]: I1207 16:23:57.927044 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-l5phd"] Dec 07 16:23:57 crc kubenswrapper[4716]: I1207 16:23:57.962025 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 07 16:23:57 crc kubenswrapper[4716]: I1207 16:23:57.962064 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 07 16:23:57 crc kubenswrapper[4716]: I1207 16:23:57.962267 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-h9pt2" Dec 07 16:23:57 crc kubenswrapper[4716]: I1207 16:23:57.963398 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 07 16:23:58 crc kubenswrapper[4716]: I1207 16:23:58.089022 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j6wb9\" (UniqueName: \"kubernetes.io/projected/2d823bb1-cb32-4c09-af7c-6ca81d0277bf-kube-api-access-j6wb9\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-l5phd\" (UID: \"2d823bb1-cb32-4c09-af7c-6ca81d0277bf\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-l5phd" Dec 07 16:23:58 crc kubenswrapper[4716]: I1207 16:23:58.089093 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2d823bb1-cb32-4c09-af7c-6ca81d0277bf-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-l5phd\" (UID: \"2d823bb1-cb32-4c09-af7c-6ca81d0277bf\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-l5phd" Dec 07 16:23:58 crc kubenswrapper[4716]: I1207 16:23:58.089236 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2d823bb1-cb32-4c09-af7c-6ca81d0277bf-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-l5phd\" (UID: \"2d823bb1-cb32-4c09-af7c-6ca81d0277bf\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-l5phd" Dec 07 16:23:58 crc kubenswrapper[4716]: I1207 16:23:58.191012 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2d823bb1-cb32-4c09-af7c-6ca81d0277bf-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-l5phd\" (UID: \"2d823bb1-cb32-4c09-af7c-6ca81d0277bf\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-l5phd" Dec 07 16:23:58 crc kubenswrapper[4716]: I1207 16:23:58.191205 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j6wb9\" (UniqueName: \"kubernetes.io/projected/2d823bb1-cb32-4c09-af7c-6ca81d0277bf-kube-api-access-j6wb9\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-l5phd\" (UID: \"2d823bb1-cb32-4c09-af7c-6ca81d0277bf\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-l5phd" Dec 07 16:23:58 crc kubenswrapper[4716]: I1207 16:23:58.191242 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2d823bb1-cb32-4c09-af7c-6ca81d0277bf-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-l5phd\" (UID: \"2d823bb1-cb32-4c09-af7c-6ca81d0277bf\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-l5phd" Dec 07 16:23:58 crc kubenswrapper[4716]: I1207 16:23:58.203669 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2d823bb1-cb32-4c09-af7c-6ca81d0277bf-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-l5phd\" (UID: \"2d823bb1-cb32-4c09-af7c-6ca81d0277bf\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-l5phd" Dec 07 16:23:58 crc kubenswrapper[4716]: I1207 16:23:58.205881 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2d823bb1-cb32-4c09-af7c-6ca81d0277bf-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-l5phd\" (UID: \"2d823bb1-cb32-4c09-af7c-6ca81d0277bf\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-l5phd" Dec 07 16:23:58 crc kubenswrapper[4716]: I1207 16:23:58.206260 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j6wb9\" (UniqueName: \"kubernetes.io/projected/2d823bb1-cb32-4c09-af7c-6ca81d0277bf-kube-api-access-j6wb9\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-l5phd\" (UID: \"2d823bb1-cb32-4c09-af7c-6ca81d0277bf\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-l5phd" Dec 07 16:23:58 crc kubenswrapper[4716]: I1207 16:23:58.285841 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-l5phd" Dec 07 16:23:58 crc kubenswrapper[4716]: I1207 16:23:58.763416 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-l5phd"] Dec 07 16:23:58 crc kubenswrapper[4716]: I1207 16:23:58.860070 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-l5phd" event={"ID":"2d823bb1-cb32-4c09-af7c-6ca81d0277bf","Type":"ContainerStarted","Data":"5b68c2d22a8bec661b3701960b149fea4c7563277a21fc5fabc1373ddf0acc20"} Dec 07 16:23:59 crc kubenswrapper[4716]: I1207 16:23:59.874490 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-l5phd" event={"ID":"2d823bb1-cb32-4c09-af7c-6ca81d0277bf","Type":"ContainerStarted","Data":"9d506d409b2556e44a96ac4ccbcf5522fa35ea8d0d0db09d93925bbb0a834bc5"} Dec 07 16:23:59 crc kubenswrapper[4716]: I1207 16:23:59.892679 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-l5phd" podStartSLOduration=2.507895957 podStartE2EDuration="2.892662664s" podCreationTimestamp="2025-12-07 16:23:57 +0000 UTC" firstStartedPulling="2025-12-07 16:23:58.769025866 +0000 UTC m=+1301.459310778" lastFinishedPulling="2025-12-07 16:23:59.153792573 +0000 UTC m=+1301.844077485" observedRunningTime="2025-12-07 16:23:59.891545254 +0000 UTC m=+1302.581830166" watchObservedRunningTime="2025-12-07 16:23:59.892662664 +0000 UTC m=+1302.582947576" Dec 07 16:24:01 crc kubenswrapper[4716]: I1207 16:24:01.901474 4716 generic.go:334] "Generic (PLEG): container finished" podID="2d823bb1-cb32-4c09-af7c-6ca81d0277bf" containerID="9d506d409b2556e44a96ac4ccbcf5522fa35ea8d0d0db09d93925bbb0a834bc5" exitCode=0 Dec 07 16:24:01 crc kubenswrapper[4716]: I1207 16:24:01.901582 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-l5phd" event={"ID":"2d823bb1-cb32-4c09-af7c-6ca81d0277bf","Type":"ContainerDied","Data":"9d506d409b2556e44a96ac4ccbcf5522fa35ea8d0d0db09d93925bbb0a834bc5"} Dec 07 16:24:03 crc kubenswrapper[4716]: I1207 16:24:03.402456 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-l5phd" Dec 07 16:24:03 crc kubenswrapper[4716]: I1207 16:24:03.591007 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2d823bb1-cb32-4c09-af7c-6ca81d0277bf-ssh-key\") pod \"2d823bb1-cb32-4c09-af7c-6ca81d0277bf\" (UID: \"2d823bb1-cb32-4c09-af7c-6ca81d0277bf\") " Dec 07 16:24:03 crc kubenswrapper[4716]: I1207 16:24:03.591254 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2d823bb1-cb32-4c09-af7c-6ca81d0277bf-inventory\") pod \"2d823bb1-cb32-4c09-af7c-6ca81d0277bf\" (UID: \"2d823bb1-cb32-4c09-af7c-6ca81d0277bf\") " Dec 07 16:24:03 crc kubenswrapper[4716]: I1207 16:24:03.591486 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j6wb9\" (UniqueName: \"kubernetes.io/projected/2d823bb1-cb32-4c09-af7c-6ca81d0277bf-kube-api-access-j6wb9\") pod \"2d823bb1-cb32-4c09-af7c-6ca81d0277bf\" (UID: \"2d823bb1-cb32-4c09-af7c-6ca81d0277bf\") " Dec 07 16:24:03 crc kubenswrapper[4716]: I1207 16:24:03.598831 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d823bb1-cb32-4c09-af7c-6ca81d0277bf-kube-api-access-j6wb9" (OuterVolumeSpecName: "kube-api-access-j6wb9") pod "2d823bb1-cb32-4c09-af7c-6ca81d0277bf" (UID: "2d823bb1-cb32-4c09-af7c-6ca81d0277bf"). InnerVolumeSpecName "kube-api-access-j6wb9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:24:03 crc kubenswrapper[4716]: I1207 16:24:03.627252 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d823bb1-cb32-4c09-af7c-6ca81d0277bf-inventory" (OuterVolumeSpecName: "inventory") pod "2d823bb1-cb32-4c09-af7c-6ca81d0277bf" (UID: "2d823bb1-cb32-4c09-af7c-6ca81d0277bf"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:24:03 crc kubenswrapper[4716]: I1207 16:24:03.642806 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d823bb1-cb32-4c09-af7c-6ca81d0277bf-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2d823bb1-cb32-4c09-af7c-6ca81d0277bf" (UID: "2d823bb1-cb32-4c09-af7c-6ca81d0277bf"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:24:03 crc kubenswrapper[4716]: I1207 16:24:03.694198 4716 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2d823bb1-cb32-4c09-af7c-6ca81d0277bf-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 07 16:24:03 crc kubenswrapper[4716]: I1207 16:24:03.694259 4716 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2d823bb1-cb32-4c09-af7c-6ca81d0277bf-inventory\") on node \"crc\" DevicePath \"\"" Dec 07 16:24:03 crc kubenswrapper[4716]: I1207 16:24:03.694287 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j6wb9\" (UniqueName: \"kubernetes.io/projected/2d823bb1-cb32-4c09-af7c-6ca81d0277bf-kube-api-access-j6wb9\") on node \"crc\" DevicePath \"\"" Dec 07 16:24:03 crc kubenswrapper[4716]: I1207 16:24:03.927192 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-l5phd" event={"ID":"2d823bb1-cb32-4c09-af7c-6ca81d0277bf","Type":"ContainerDied","Data":"5b68c2d22a8bec661b3701960b149fea4c7563277a21fc5fabc1373ddf0acc20"} Dec 07 16:24:03 crc kubenswrapper[4716]: I1207 16:24:03.927250 4716 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5b68c2d22a8bec661b3701960b149fea4c7563277a21fc5fabc1373ddf0acc20" Dec 07 16:24:03 crc kubenswrapper[4716]: I1207 16:24:03.927334 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-l5phd" Dec 07 16:24:04 crc kubenswrapper[4716]: I1207 16:24:04.107985 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bvvcz"] Dec 07 16:24:04 crc kubenswrapper[4716]: E1207 16:24:04.108580 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d823bb1-cb32-4c09-af7c-6ca81d0277bf" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 07 16:24:04 crc kubenswrapper[4716]: I1207 16:24:04.108606 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d823bb1-cb32-4c09-af7c-6ca81d0277bf" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 07 16:24:04 crc kubenswrapper[4716]: I1207 16:24:04.108831 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d823bb1-cb32-4c09-af7c-6ca81d0277bf" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 07 16:24:04 crc kubenswrapper[4716]: I1207 16:24:04.109548 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bvvcz" Dec 07 16:24:04 crc kubenswrapper[4716]: I1207 16:24:04.112359 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 07 16:24:04 crc kubenswrapper[4716]: I1207 16:24:04.112566 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 07 16:24:04 crc kubenswrapper[4716]: I1207 16:24:04.113186 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-h9pt2" Dec 07 16:24:04 crc kubenswrapper[4716]: I1207 16:24:04.113416 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 07 16:24:04 crc kubenswrapper[4716]: I1207 16:24:04.116528 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bvvcz"] Dec 07 16:24:04 crc kubenswrapper[4716]: I1207 16:24:04.131151 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-26n7q\" (UniqueName: \"kubernetes.io/projected/1c545d8b-e4c4-411c-b122-68f22c6befa4-kube-api-access-26n7q\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bvvcz\" (UID: \"1c545d8b-e4c4-411c-b122-68f22c6befa4\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bvvcz" Dec 07 16:24:04 crc kubenswrapper[4716]: I1207 16:24:04.131217 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c545d8b-e4c4-411c-b122-68f22c6befa4-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bvvcz\" (UID: \"1c545d8b-e4c4-411c-b122-68f22c6befa4\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bvvcz" Dec 07 16:24:04 crc kubenswrapper[4716]: I1207 16:24:04.131304 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1c545d8b-e4c4-411c-b122-68f22c6befa4-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bvvcz\" (UID: \"1c545d8b-e4c4-411c-b122-68f22c6befa4\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bvvcz" Dec 07 16:24:04 crc kubenswrapper[4716]: I1207 16:24:04.131330 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1c545d8b-e4c4-411c-b122-68f22c6befa4-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bvvcz\" (UID: \"1c545d8b-e4c4-411c-b122-68f22c6befa4\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bvvcz" Dec 07 16:24:04 crc kubenswrapper[4716]: I1207 16:24:04.232758 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-26n7q\" (UniqueName: \"kubernetes.io/projected/1c545d8b-e4c4-411c-b122-68f22c6befa4-kube-api-access-26n7q\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bvvcz\" (UID: \"1c545d8b-e4c4-411c-b122-68f22c6befa4\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bvvcz" Dec 07 16:24:04 crc kubenswrapper[4716]: I1207 16:24:04.232844 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c545d8b-e4c4-411c-b122-68f22c6befa4-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bvvcz\" (UID: \"1c545d8b-e4c4-411c-b122-68f22c6befa4\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bvvcz" Dec 07 16:24:04 crc kubenswrapper[4716]: I1207 16:24:04.232922 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1c545d8b-e4c4-411c-b122-68f22c6befa4-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bvvcz\" (UID: \"1c545d8b-e4c4-411c-b122-68f22c6befa4\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bvvcz" Dec 07 16:24:04 crc kubenswrapper[4716]: I1207 16:24:04.232943 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1c545d8b-e4c4-411c-b122-68f22c6befa4-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bvvcz\" (UID: \"1c545d8b-e4c4-411c-b122-68f22c6befa4\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bvvcz" Dec 07 16:24:04 crc kubenswrapper[4716]: I1207 16:24:04.237232 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1c545d8b-e4c4-411c-b122-68f22c6befa4-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bvvcz\" (UID: \"1c545d8b-e4c4-411c-b122-68f22c6befa4\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bvvcz" Dec 07 16:24:04 crc kubenswrapper[4716]: I1207 16:24:04.237308 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1c545d8b-e4c4-411c-b122-68f22c6befa4-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bvvcz\" (UID: \"1c545d8b-e4c4-411c-b122-68f22c6befa4\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bvvcz" Dec 07 16:24:04 crc kubenswrapper[4716]: I1207 16:24:04.238039 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c545d8b-e4c4-411c-b122-68f22c6befa4-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bvvcz\" (UID: \"1c545d8b-e4c4-411c-b122-68f22c6befa4\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bvvcz" Dec 07 16:24:04 crc kubenswrapper[4716]: I1207 16:24:04.248686 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-26n7q\" (UniqueName: \"kubernetes.io/projected/1c545d8b-e4c4-411c-b122-68f22c6befa4-kube-api-access-26n7q\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bvvcz\" (UID: \"1c545d8b-e4c4-411c-b122-68f22c6befa4\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bvvcz" Dec 07 16:24:04 crc kubenswrapper[4716]: I1207 16:24:04.432381 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bvvcz" Dec 07 16:24:06 crc kubenswrapper[4716]: I1207 16:24:06.864253 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bvvcz"] Dec 07 16:24:06 crc kubenswrapper[4716]: I1207 16:24:06.962574 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bvvcz" event={"ID":"1c545d8b-e4c4-411c-b122-68f22c6befa4","Type":"ContainerStarted","Data":"d2bf2d004c7d7b1563f55c5a4b30f4f358d53b5ccc78ff789c819e927aaae49b"} Dec 07 16:24:07 crc kubenswrapper[4716]: I1207 16:24:07.978033 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bvvcz" event={"ID":"1c545d8b-e4c4-411c-b122-68f22c6befa4","Type":"ContainerStarted","Data":"45f153147cb48f04ffe9893b8a29774f2c9103b0b70c3b5117f9db339750f93b"} Dec 07 16:24:08 crc kubenswrapper[4716]: I1207 16:24:07.995453 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bvvcz" podStartSLOduration=3.511357471 podStartE2EDuration="3.995432539s" podCreationTimestamp="2025-12-07 16:24:04 +0000 UTC" firstStartedPulling="2025-12-07 16:24:06.870744284 +0000 UTC m=+1309.561029206" lastFinishedPulling="2025-12-07 16:24:07.354819362 +0000 UTC m=+1310.045104274" observedRunningTime="2025-12-07 16:24:07.993058876 +0000 UTC m=+1310.683343788" watchObservedRunningTime="2025-12-07 16:24:07.995432539 +0000 UTC m=+1310.685717471" Dec 07 16:25:32 crc kubenswrapper[4716]: I1207 16:25:32.376507 4716 scope.go:117] "RemoveContainer" containerID="f4bac783db6117885371eea1a50ac5ca886c3831c5a98dd5a59c5aaaf4c1c5cd" Dec 07 16:25:32 crc kubenswrapper[4716]: I1207 16:25:32.403796 4716 scope.go:117] "RemoveContainer" containerID="49e2ed79567e1481a18a12c7c4b8d770a72017bbcff0fed08ff63919176cc886" Dec 07 16:25:32 crc kubenswrapper[4716]: I1207 16:25:32.476967 4716 scope.go:117] "RemoveContainer" containerID="a97b565989cc9fd8373a3674aaa3245f870237bc8cf96066c747220dd37285f9" Dec 07 16:25:52 crc kubenswrapper[4716]: I1207 16:25:52.761515 4716 patch_prober.go:28] interesting pod/machine-config-daemon-zcxxp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 16:25:52 crc kubenswrapper[4716]: I1207 16:25:52.762034 4716 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 16:26:22 crc kubenswrapper[4716]: I1207 16:26:22.761936 4716 patch_prober.go:28] interesting pod/machine-config-daemon-zcxxp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 16:26:22 crc kubenswrapper[4716]: I1207 16:26:22.762506 4716 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 16:26:52 crc kubenswrapper[4716]: I1207 16:26:52.761687 4716 patch_prober.go:28] interesting pod/machine-config-daemon-zcxxp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 16:26:52 crc kubenswrapper[4716]: I1207 16:26:52.762338 4716 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 16:26:52 crc kubenswrapper[4716]: I1207 16:26:52.762390 4716 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" Dec 07 16:26:52 crc kubenswrapper[4716]: I1207 16:26:52.763381 4716 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fea4ce7ae8e17d1efab423015c4c8670e9be41f20a55cff5db1bae1736f619e0"} pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 07 16:26:52 crc kubenswrapper[4716]: I1207 16:26:52.763522 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" containerName="machine-config-daemon" containerID="cri-o://fea4ce7ae8e17d1efab423015c4c8670e9be41f20a55cff5db1bae1736f619e0" gracePeriod=600 Dec 07 16:26:53 crc kubenswrapper[4716]: I1207 16:26:53.681665 4716 generic.go:334] "Generic (PLEG): container finished" podID="c15b59eb-565d-4556-a4ce-75afdf159dc8" containerID="fea4ce7ae8e17d1efab423015c4c8670e9be41f20a55cff5db1bae1736f619e0" exitCode=0 Dec 07 16:26:53 crc kubenswrapper[4716]: I1207 16:26:53.681740 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" event={"ID":"c15b59eb-565d-4556-a4ce-75afdf159dc8","Type":"ContainerDied","Data":"fea4ce7ae8e17d1efab423015c4c8670e9be41f20a55cff5db1bae1736f619e0"} Dec 07 16:26:53 crc kubenswrapper[4716]: I1207 16:26:53.682056 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" event={"ID":"c15b59eb-565d-4556-a4ce-75afdf159dc8","Type":"ContainerStarted","Data":"bae1a3e9d80f5b00293cade8b7601799399c2883ae101574459a4c01e523a154"} Dec 07 16:26:53 crc kubenswrapper[4716]: I1207 16:26:53.682103 4716 scope.go:117] "RemoveContainer" containerID="43e1cbd05079c8cdd0bd98789bfa5b92ef2c8c3e87845aeb18f274ab5529de34" Dec 07 16:27:11 crc kubenswrapper[4716]: I1207 16:27:11.888798 4716 generic.go:334] "Generic (PLEG): container finished" podID="1c545d8b-e4c4-411c-b122-68f22c6befa4" containerID="45f153147cb48f04ffe9893b8a29774f2c9103b0b70c3b5117f9db339750f93b" exitCode=0 Dec 07 16:27:11 crc kubenswrapper[4716]: I1207 16:27:11.888869 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bvvcz" event={"ID":"1c545d8b-e4c4-411c-b122-68f22c6befa4","Type":"ContainerDied","Data":"45f153147cb48f04ffe9893b8a29774f2c9103b0b70c3b5117f9db339750f93b"} Dec 07 16:27:13 crc kubenswrapper[4716]: I1207 16:27:13.415438 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bvvcz" Dec 07 16:27:13 crc kubenswrapper[4716]: I1207 16:27:13.511923 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-26n7q\" (UniqueName: \"kubernetes.io/projected/1c545d8b-e4c4-411c-b122-68f22c6befa4-kube-api-access-26n7q\") pod \"1c545d8b-e4c4-411c-b122-68f22c6befa4\" (UID: \"1c545d8b-e4c4-411c-b122-68f22c6befa4\") " Dec 07 16:27:13 crc kubenswrapper[4716]: I1207 16:27:13.511996 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1c545d8b-e4c4-411c-b122-68f22c6befa4-ssh-key\") pod \"1c545d8b-e4c4-411c-b122-68f22c6befa4\" (UID: \"1c545d8b-e4c4-411c-b122-68f22c6befa4\") " Dec 07 16:27:13 crc kubenswrapper[4716]: I1207 16:27:13.512023 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1c545d8b-e4c4-411c-b122-68f22c6befa4-inventory\") pod \"1c545d8b-e4c4-411c-b122-68f22c6befa4\" (UID: \"1c545d8b-e4c4-411c-b122-68f22c6befa4\") " Dec 07 16:27:13 crc kubenswrapper[4716]: I1207 16:27:13.512068 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c545d8b-e4c4-411c-b122-68f22c6befa4-bootstrap-combined-ca-bundle\") pod \"1c545d8b-e4c4-411c-b122-68f22c6befa4\" (UID: \"1c545d8b-e4c4-411c-b122-68f22c6befa4\") " Dec 07 16:27:13 crc kubenswrapper[4716]: I1207 16:27:13.519884 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c545d8b-e4c4-411c-b122-68f22c6befa4-kube-api-access-26n7q" (OuterVolumeSpecName: "kube-api-access-26n7q") pod "1c545d8b-e4c4-411c-b122-68f22c6befa4" (UID: "1c545d8b-e4c4-411c-b122-68f22c6befa4"). InnerVolumeSpecName "kube-api-access-26n7q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:27:13 crc kubenswrapper[4716]: I1207 16:27:13.529305 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1c545d8b-e4c4-411c-b122-68f22c6befa4-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "1c545d8b-e4c4-411c-b122-68f22c6befa4" (UID: "1c545d8b-e4c4-411c-b122-68f22c6befa4"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:27:13 crc kubenswrapper[4716]: I1207 16:27:13.572147 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1c545d8b-e4c4-411c-b122-68f22c6befa4-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1c545d8b-e4c4-411c-b122-68f22c6befa4" (UID: "1c545d8b-e4c4-411c-b122-68f22c6befa4"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:27:13 crc kubenswrapper[4716]: I1207 16:27:13.575161 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1c545d8b-e4c4-411c-b122-68f22c6befa4-inventory" (OuterVolumeSpecName: "inventory") pod "1c545d8b-e4c4-411c-b122-68f22c6befa4" (UID: "1c545d8b-e4c4-411c-b122-68f22c6befa4"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:27:13 crc kubenswrapper[4716]: I1207 16:27:13.614564 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-26n7q\" (UniqueName: \"kubernetes.io/projected/1c545d8b-e4c4-411c-b122-68f22c6befa4-kube-api-access-26n7q\") on node \"crc\" DevicePath \"\"" Dec 07 16:27:13 crc kubenswrapper[4716]: I1207 16:27:13.614606 4716 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1c545d8b-e4c4-411c-b122-68f22c6befa4-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 07 16:27:13 crc kubenswrapper[4716]: I1207 16:27:13.614619 4716 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1c545d8b-e4c4-411c-b122-68f22c6befa4-inventory\") on node \"crc\" DevicePath \"\"" Dec 07 16:27:13 crc kubenswrapper[4716]: I1207 16:27:13.614631 4716 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c545d8b-e4c4-411c-b122-68f22c6befa4-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 16:27:13 crc kubenswrapper[4716]: I1207 16:27:13.915932 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bvvcz" event={"ID":"1c545d8b-e4c4-411c-b122-68f22c6befa4","Type":"ContainerDied","Data":"d2bf2d004c7d7b1563f55c5a4b30f4f358d53b5ccc78ff789c819e927aaae49b"} Dec 07 16:27:13 crc kubenswrapper[4716]: I1207 16:27:13.915981 4716 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d2bf2d004c7d7b1563f55c5a4b30f4f358d53b5ccc78ff789c819e927aaae49b" Dec 07 16:27:13 crc kubenswrapper[4716]: I1207 16:27:13.916218 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bvvcz" Dec 07 16:27:14 crc kubenswrapper[4716]: I1207 16:27:14.025998 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9d8h7"] Dec 07 16:27:14 crc kubenswrapper[4716]: E1207 16:27:14.026649 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c545d8b-e4c4-411c-b122-68f22c6befa4" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 07 16:27:14 crc kubenswrapper[4716]: I1207 16:27:14.026683 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c545d8b-e4c4-411c-b122-68f22c6befa4" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 07 16:27:14 crc kubenswrapper[4716]: I1207 16:27:14.027237 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c545d8b-e4c4-411c-b122-68f22c6befa4" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 07 16:27:14 crc kubenswrapper[4716]: I1207 16:27:14.028251 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9d8h7" Dec 07 16:27:14 crc kubenswrapper[4716]: I1207 16:27:14.032567 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 07 16:27:14 crc kubenswrapper[4716]: I1207 16:27:14.032640 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 07 16:27:14 crc kubenswrapper[4716]: I1207 16:27:14.034220 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-h9pt2" Dec 07 16:27:14 crc kubenswrapper[4716]: I1207 16:27:14.034544 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 07 16:27:14 crc kubenswrapper[4716]: I1207 16:27:14.036252 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9d8h7"] Dec 07 16:27:14 crc kubenswrapper[4716]: I1207 16:27:14.123414 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7495cf42-56aa-43c9-9d24-a022a3e50505-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-9d8h7\" (UID: \"7495cf42-56aa-43c9-9d24-a022a3e50505\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9d8h7" Dec 07 16:27:14 crc kubenswrapper[4716]: I1207 16:27:14.123786 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7495cf42-56aa-43c9-9d24-a022a3e50505-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-9d8h7\" (UID: \"7495cf42-56aa-43c9-9d24-a022a3e50505\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9d8h7" Dec 07 16:27:14 crc kubenswrapper[4716]: I1207 16:27:14.123815 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2w77\" (UniqueName: \"kubernetes.io/projected/7495cf42-56aa-43c9-9d24-a022a3e50505-kube-api-access-t2w77\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-9d8h7\" (UID: \"7495cf42-56aa-43c9-9d24-a022a3e50505\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9d8h7" Dec 07 16:27:14 crc kubenswrapper[4716]: I1207 16:27:14.225918 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7495cf42-56aa-43c9-9d24-a022a3e50505-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-9d8h7\" (UID: \"7495cf42-56aa-43c9-9d24-a022a3e50505\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9d8h7" Dec 07 16:27:14 crc kubenswrapper[4716]: I1207 16:27:14.226038 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2w77\" (UniqueName: \"kubernetes.io/projected/7495cf42-56aa-43c9-9d24-a022a3e50505-kube-api-access-t2w77\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-9d8h7\" (UID: \"7495cf42-56aa-43c9-9d24-a022a3e50505\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9d8h7" Dec 07 16:27:14 crc kubenswrapper[4716]: I1207 16:27:14.226547 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7495cf42-56aa-43c9-9d24-a022a3e50505-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-9d8h7\" (UID: \"7495cf42-56aa-43c9-9d24-a022a3e50505\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9d8h7" Dec 07 16:27:14 crc kubenswrapper[4716]: I1207 16:27:14.229948 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7495cf42-56aa-43c9-9d24-a022a3e50505-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-9d8h7\" (UID: \"7495cf42-56aa-43c9-9d24-a022a3e50505\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9d8h7" Dec 07 16:27:14 crc kubenswrapper[4716]: I1207 16:27:14.230039 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7495cf42-56aa-43c9-9d24-a022a3e50505-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-9d8h7\" (UID: \"7495cf42-56aa-43c9-9d24-a022a3e50505\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9d8h7" Dec 07 16:27:14 crc kubenswrapper[4716]: I1207 16:27:14.248134 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2w77\" (UniqueName: \"kubernetes.io/projected/7495cf42-56aa-43c9-9d24-a022a3e50505-kube-api-access-t2w77\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-9d8h7\" (UID: \"7495cf42-56aa-43c9-9d24-a022a3e50505\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9d8h7" Dec 07 16:27:14 crc kubenswrapper[4716]: I1207 16:27:14.369612 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9d8h7" Dec 07 16:27:14 crc kubenswrapper[4716]: I1207 16:27:14.912388 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9d8h7"] Dec 07 16:27:14 crc kubenswrapper[4716]: I1207 16:27:14.929139 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9d8h7" event={"ID":"7495cf42-56aa-43c9-9d24-a022a3e50505","Type":"ContainerStarted","Data":"0b681ad96fa545e4d53487591674d93e63b58e02153586c274b0cac81ed3d490"} Dec 07 16:27:15 crc kubenswrapper[4716]: I1207 16:27:15.940705 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9d8h7" event={"ID":"7495cf42-56aa-43c9-9d24-a022a3e50505","Type":"ContainerStarted","Data":"522ffd22fe63962dd32ec2edd14e872c82266d33517f7415098481badc213f50"} Dec 07 16:27:15 crc kubenswrapper[4716]: I1207 16:27:15.970055 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9d8h7" podStartSLOduration=2.475399622 podStartE2EDuration="2.970030797s" podCreationTimestamp="2025-12-07 16:27:13 +0000 UTC" firstStartedPulling="2025-12-07 16:27:14.910785001 +0000 UTC m=+1497.601069913" lastFinishedPulling="2025-12-07 16:27:15.405416156 +0000 UTC m=+1498.095701088" observedRunningTime="2025-12-07 16:27:15.960928192 +0000 UTC m=+1498.651213114" watchObservedRunningTime="2025-12-07 16:27:15.970030797 +0000 UTC m=+1498.660315719" Dec 07 16:27:17 crc kubenswrapper[4716]: I1207 16:27:17.882324 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-9r8px"] Dec 07 16:27:17 crc kubenswrapper[4716]: I1207 16:27:17.886831 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9r8px" Dec 07 16:27:17 crc kubenswrapper[4716]: I1207 16:27:17.896442 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9r8px"] Dec 07 16:27:18 crc kubenswrapper[4716]: I1207 16:27:18.004599 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9aec4026-f10a-446e-b68d-e90f10680919-utilities\") pod \"redhat-marketplace-9r8px\" (UID: \"9aec4026-f10a-446e-b68d-e90f10680919\") " pod="openshift-marketplace/redhat-marketplace-9r8px" Dec 07 16:27:18 crc kubenswrapper[4716]: I1207 16:27:18.004669 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95wwr\" (UniqueName: \"kubernetes.io/projected/9aec4026-f10a-446e-b68d-e90f10680919-kube-api-access-95wwr\") pod \"redhat-marketplace-9r8px\" (UID: \"9aec4026-f10a-446e-b68d-e90f10680919\") " pod="openshift-marketplace/redhat-marketplace-9r8px" Dec 07 16:27:18 crc kubenswrapper[4716]: I1207 16:27:18.004728 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9aec4026-f10a-446e-b68d-e90f10680919-catalog-content\") pod \"redhat-marketplace-9r8px\" (UID: \"9aec4026-f10a-446e-b68d-e90f10680919\") " pod="openshift-marketplace/redhat-marketplace-9r8px" Dec 07 16:27:18 crc kubenswrapper[4716]: I1207 16:27:18.106438 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95wwr\" (UniqueName: \"kubernetes.io/projected/9aec4026-f10a-446e-b68d-e90f10680919-kube-api-access-95wwr\") pod \"redhat-marketplace-9r8px\" (UID: \"9aec4026-f10a-446e-b68d-e90f10680919\") " pod="openshift-marketplace/redhat-marketplace-9r8px" Dec 07 16:27:18 crc kubenswrapper[4716]: I1207 16:27:18.106734 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9aec4026-f10a-446e-b68d-e90f10680919-catalog-content\") pod \"redhat-marketplace-9r8px\" (UID: \"9aec4026-f10a-446e-b68d-e90f10680919\") " pod="openshift-marketplace/redhat-marketplace-9r8px" Dec 07 16:27:18 crc kubenswrapper[4716]: I1207 16:27:18.106908 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9aec4026-f10a-446e-b68d-e90f10680919-utilities\") pod \"redhat-marketplace-9r8px\" (UID: \"9aec4026-f10a-446e-b68d-e90f10680919\") " pod="openshift-marketplace/redhat-marketplace-9r8px" Dec 07 16:27:18 crc kubenswrapper[4716]: I1207 16:27:18.107288 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9aec4026-f10a-446e-b68d-e90f10680919-catalog-content\") pod \"redhat-marketplace-9r8px\" (UID: \"9aec4026-f10a-446e-b68d-e90f10680919\") " pod="openshift-marketplace/redhat-marketplace-9r8px" Dec 07 16:27:18 crc kubenswrapper[4716]: I1207 16:27:18.107364 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9aec4026-f10a-446e-b68d-e90f10680919-utilities\") pod \"redhat-marketplace-9r8px\" (UID: \"9aec4026-f10a-446e-b68d-e90f10680919\") " pod="openshift-marketplace/redhat-marketplace-9r8px" Dec 07 16:27:18 crc kubenswrapper[4716]: I1207 16:27:18.131305 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95wwr\" (UniqueName: \"kubernetes.io/projected/9aec4026-f10a-446e-b68d-e90f10680919-kube-api-access-95wwr\") pod \"redhat-marketplace-9r8px\" (UID: \"9aec4026-f10a-446e-b68d-e90f10680919\") " pod="openshift-marketplace/redhat-marketplace-9r8px" Dec 07 16:27:18 crc kubenswrapper[4716]: I1207 16:27:18.210221 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9r8px" Dec 07 16:27:18 crc kubenswrapper[4716]: I1207 16:27:18.699839 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9r8px"] Dec 07 16:27:18 crc kubenswrapper[4716]: W1207 16:27:18.719563 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9aec4026_f10a_446e_b68d_e90f10680919.slice/crio-d7abdee3cede27ea7528e4f921eaf2a1fea9c2d589dbe2ee8fa3628b4ece33e9 WatchSource:0}: Error finding container d7abdee3cede27ea7528e4f921eaf2a1fea9c2d589dbe2ee8fa3628b4ece33e9: Status 404 returned error can't find the container with id d7abdee3cede27ea7528e4f921eaf2a1fea9c2d589dbe2ee8fa3628b4ece33e9 Dec 07 16:27:18 crc kubenswrapper[4716]: I1207 16:27:18.970360 4716 generic.go:334] "Generic (PLEG): container finished" podID="9aec4026-f10a-446e-b68d-e90f10680919" containerID="32667ca36d90ec8e9bb140019d5898293bdc8813dcaea77a0e3f07fbea4bcbba" exitCode=0 Dec 07 16:27:18 crc kubenswrapper[4716]: I1207 16:27:18.970420 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9r8px" event={"ID":"9aec4026-f10a-446e-b68d-e90f10680919","Type":"ContainerDied","Data":"32667ca36d90ec8e9bb140019d5898293bdc8813dcaea77a0e3f07fbea4bcbba"} Dec 07 16:27:18 crc kubenswrapper[4716]: I1207 16:27:18.970777 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9r8px" event={"ID":"9aec4026-f10a-446e-b68d-e90f10680919","Type":"ContainerStarted","Data":"d7abdee3cede27ea7528e4f921eaf2a1fea9c2d589dbe2ee8fa3628b4ece33e9"} Dec 07 16:27:19 crc kubenswrapper[4716]: I1207 16:27:19.984829 4716 generic.go:334] "Generic (PLEG): container finished" podID="9aec4026-f10a-446e-b68d-e90f10680919" containerID="3e06932feb2f78bb5ee70e6e341894bbb0fceac5c4fa707d204ce2a027a692b1" exitCode=0 Dec 07 16:27:19 crc kubenswrapper[4716]: I1207 16:27:19.984927 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9r8px" event={"ID":"9aec4026-f10a-446e-b68d-e90f10680919","Type":"ContainerDied","Data":"3e06932feb2f78bb5ee70e6e341894bbb0fceac5c4fa707d204ce2a027a692b1"} Dec 07 16:27:21 crc kubenswrapper[4716]: I1207 16:27:21.009115 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9r8px" event={"ID":"9aec4026-f10a-446e-b68d-e90f10680919","Type":"ContainerStarted","Data":"b0de677d421cc44a12fdf93b6e170507609c419aff1bbe23e0db2bc93ed77b26"} Dec 07 16:27:21 crc kubenswrapper[4716]: I1207 16:27:21.051659 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-9r8px" podStartSLOduration=2.574400325 podStartE2EDuration="4.051637906s" podCreationTimestamp="2025-12-07 16:27:17 +0000 UTC" firstStartedPulling="2025-12-07 16:27:18.971716866 +0000 UTC m=+1501.662001768" lastFinishedPulling="2025-12-07 16:27:20.448954437 +0000 UTC m=+1503.139239349" observedRunningTime="2025-12-07 16:27:21.038428569 +0000 UTC m=+1503.728713491" watchObservedRunningTime="2025-12-07 16:27:21.051637906 +0000 UTC m=+1503.741922818" Dec 07 16:27:28 crc kubenswrapper[4716]: I1207 16:27:28.210610 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-9r8px" Dec 07 16:27:28 crc kubenswrapper[4716]: I1207 16:27:28.211426 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-9r8px" Dec 07 16:27:28 crc kubenswrapper[4716]: I1207 16:27:28.283819 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-9r8px" Dec 07 16:27:29 crc kubenswrapper[4716]: I1207 16:27:29.157610 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-9r8px" Dec 07 16:27:29 crc kubenswrapper[4716]: I1207 16:27:29.211403 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9r8px"] Dec 07 16:27:31 crc kubenswrapper[4716]: I1207 16:27:31.103730 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-9r8px" podUID="9aec4026-f10a-446e-b68d-e90f10680919" containerName="registry-server" containerID="cri-o://b0de677d421cc44a12fdf93b6e170507609c419aff1bbe23e0db2bc93ed77b26" gracePeriod=2 Dec 07 16:27:31 crc kubenswrapper[4716]: I1207 16:27:31.646226 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9r8px" Dec 07 16:27:31 crc kubenswrapper[4716]: I1207 16:27:31.718896 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9aec4026-f10a-446e-b68d-e90f10680919-utilities\") pod \"9aec4026-f10a-446e-b68d-e90f10680919\" (UID: \"9aec4026-f10a-446e-b68d-e90f10680919\") " Dec 07 16:27:31 crc kubenswrapper[4716]: I1207 16:27:31.718970 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-95wwr\" (UniqueName: \"kubernetes.io/projected/9aec4026-f10a-446e-b68d-e90f10680919-kube-api-access-95wwr\") pod \"9aec4026-f10a-446e-b68d-e90f10680919\" (UID: \"9aec4026-f10a-446e-b68d-e90f10680919\") " Dec 07 16:27:31 crc kubenswrapper[4716]: I1207 16:27:31.719111 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9aec4026-f10a-446e-b68d-e90f10680919-catalog-content\") pod \"9aec4026-f10a-446e-b68d-e90f10680919\" (UID: \"9aec4026-f10a-446e-b68d-e90f10680919\") " Dec 07 16:27:31 crc kubenswrapper[4716]: I1207 16:27:31.720162 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9aec4026-f10a-446e-b68d-e90f10680919-utilities" (OuterVolumeSpecName: "utilities") pod "9aec4026-f10a-446e-b68d-e90f10680919" (UID: "9aec4026-f10a-446e-b68d-e90f10680919"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:27:31 crc kubenswrapper[4716]: I1207 16:27:31.724567 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9aec4026-f10a-446e-b68d-e90f10680919-kube-api-access-95wwr" (OuterVolumeSpecName: "kube-api-access-95wwr") pod "9aec4026-f10a-446e-b68d-e90f10680919" (UID: "9aec4026-f10a-446e-b68d-e90f10680919"). InnerVolumeSpecName "kube-api-access-95wwr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:27:31 crc kubenswrapper[4716]: I1207 16:27:31.743902 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9aec4026-f10a-446e-b68d-e90f10680919-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9aec4026-f10a-446e-b68d-e90f10680919" (UID: "9aec4026-f10a-446e-b68d-e90f10680919"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:27:31 crc kubenswrapper[4716]: I1207 16:27:31.820854 4716 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9aec4026-f10a-446e-b68d-e90f10680919-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 07 16:27:31 crc kubenswrapper[4716]: I1207 16:27:31.820888 4716 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9aec4026-f10a-446e-b68d-e90f10680919-utilities\") on node \"crc\" DevicePath \"\"" Dec 07 16:27:31 crc kubenswrapper[4716]: I1207 16:27:31.820898 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-95wwr\" (UniqueName: \"kubernetes.io/projected/9aec4026-f10a-446e-b68d-e90f10680919-kube-api-access-95wwr\") on node \"crc\" DevicePath \"\"" Dec 07 16:27:32 crc kubenswrapper[4716]: I1207 16:27:32.117997 4716 generic.go:334] "Generic (PLEG): container finished" podID="9aec4026-f10a-446e-b68d-e90f10680919" containerID="b0de677d421cc44a12fdf93b6e170507609c419aff1bbe23e0db2bc93ed77b26" exitCode=0 Dec 07 16:27:32 crc kubenswrapper[4716]: I1207 16:27:32.118335 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9r8px" event={"ID":"9aec4026-f10a-446e-b68d-e90f10680919","Type":"ContainerDied","Data":"b0de677d421cc44a12fdf93b6e170507609c419aff1bbe23e0db2bc93ed77b26"} Dec 07 16:27:32 crc kubenswrapper[4716]: I1207 16:27:32.118364 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9r8px" event={"ID":"9aec4026-f10a-446e-b68d-e90f10680919","Type":"ContainerDied","Data":"d7abdee3cede27ea7528e4f921eaf2a1fea9c2d589dbe2ee8fa3628b4ece33e9"} Dec 07 16:27:32 crc kubenswrapper[4716]: I1207 16:27:32.118380 4716 scope.go:117] "RemoveContainer" containerID="b0de677d421cc44a12fdf93b6e170507609c419aff1bbe23e0db2bc93ed77b26" Dec 07 16:27:32 crc kubenswrapper[4716]: I1207 16:27:32.118499 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9r8px" Dec 07 16:27:32 crc kubenswrapper[4716]: I1207 16:27:32.152393 4716 scope.go:117] "RemoveContainer" containerID="3e06932feb2f78bb5ee70e6e341894bbb0fceac5c4fa707d204ce2a027a692b1" Dec 07 16:27:32 crc kubenswrapper[4716]: I1207 16:27:32.160203 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9r8px"] Dec 07 16:27:32 crc kubenswrapper[4716]: I1207 16:27:32.168344 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-9r8px"] Dec 07 16:27:32 crc kubenswrapper[4716]: I1207 16:27:32.177872 4716 scope.go:117] "RemoveContainer" containerID="32667ca36d90ec8e9bb140019d5898293bdc8813dcaea77a0e3f07fbea4bcbba" Dec 07 16:27:32 crc kubenswrapper[4716]: I1207 16:27:32.218548 4716 scope.go:117] "RemoveContainer" containerID="b0de677d421cc44a12fdf93b6e170507609c419aff1bbe23e0db2bc93ed77b26" Dec 07 16:27:32 crc kubenswrapper[4716]: E1207 16:27:32.218938 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b0de677d421cc44a12fdf93b6e170507609c419aff1bbe23e0db2bc93ed77b26\": container with ID starting with b0de677d421cc44a12fdf93b6e170507609c419aff1bbe23e0db2bc93ed77b26 not found: ID does not exist" containerID="b0de677d421cc44a12fdf93b6e170507609c419aff1bbe23e0db2bc93ed77b26" Dec 07 16:27:32 crc kubenswrapper[4716]: I1207 16:27:32.218988 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b0de677d421cc44a12fdf93b6e170507609c419aff1bbe23e0db2bc93ed77b26"} err="failed to get container status \"b0de677d421cc44a12fdf93b6e170507609c419aff1bbe23e0db2bc93ed77b26\": rpc error: code = NotFound desc = could not find container \"b0de677d421cc44a12fdf93b6e170507609c419aff1bbe23e0db2bc93ed77b26\": container with ID starting with b0de677d421cc44a12fdf93b6e170507609c419aff1bbe23e0db2bc93ed77b26 not found: ID does not exist" Dec 07 16:27:32 crc kubenswrapper[4716]: I1207 16:27:32.219026 4716 scope.go:117] "RemoveContainer" containerID="3e06932feb2f78bb5ee70e6e341894bbb0fceac5c4fa707d204ce2a027a692b1" Dec 07 16:27:32 crc kubenswrapper[4716]: E1207 16:27:32.219554 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e06932feb2f78bb5ee70e6e341894bbb0fceac5c4fa707d204ce2a027a692b1\": container with ID starting with 3e06932feb2f78bb5ee70e6e341894bbb0fceac5c4fa707d204ce2a027a692b1 not found: ID does not exist" containerID="3e06932feb2f78bb5ee70e6e341894bbb0fceac5c4fa707d204ce2a027a692b1" Dec 07 16:27:32 crc kubenswrapper[4716]: I1207 16:27:32.219582 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e06932feb2f78bb5ee70e6e341894bbb0fceac5c4fa707d204ce2a027a692b1"} err="failed to get container status \"3e06932feb2f78bb5ee70e6e341894bbb0fceac5c4fa707d204ce2a027a692b1\": rpc error: code = NotFound desc = could not find container \"3e06932feb2f78bb5ee70e6e341894bbb0fceac5c4fa707d204ce2a027a692b1\": container with ID starting with 3e06932feb2f78bb5ee70e6e341894bbb0fceac5c4fa707d204ce2a027a692b1 not found: ID does not exist" Dec 07 16:27:32 crc kubenswrapper[4716]: I1207 16:27:32.219600 4716 scope.go:117] "RemoveContainer" containerID="32667ca36d90ec8e9bb140019d5898293bdc8813dcaea77a0e3f07fbea4bcbba" Dec 07 16:27:32 crc kubenswrapper[4716]: E1207 16:27:32.219961 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"32667ca36d90ec8e9bb140019d5898293bdc8813dcaea77a0e3f07fbea4bcbba\": container with ID starting with 32667ca36d90ec8e9bb140019d5898293bdc8813dcaea77a0e3f07fbea4bcbba not found: ID does not exist" containerID="32667ca36d90ec8e9bb140019d5898293bdc8813dcaea77a0e3f07fbea4bcbba" Dec 07 16:27:32 crc kubenswrapper[4716]: I1207 16:27:32.220005 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32667ca36d90ec8e9bb140019d5898293bdc8813dcaea77a0e3f07fbea4bcbba"} err="failed to get container status \"32667ca36d90ec8e9bb140019d5898293bdc8813dcaea77a0e3f07fbea4bcbba\": rpc error: code = NotFound desc = could not find container \"32667ca36d90ec8e9bb140019d5898293bdc8813dcaea77a0e3f07fbea4bcbba\": container with ID starting with 32667ca36d90ec8e9bb140019d5898293bdc8813dcaea77a0e3f07fbea4bcbba not found: ID does not exist" Dec 07 16:27:32 crc kubenswrapper[4716]: I1207 16:27:32.594228 4716 scope.go:117] "RemoveContainer" containerID="08d95b491dfc6c45ca2b0f45ca175db851104e371ff2993f44bc54dd1187a719" Dec 07 16:27:33 crc kubenswrapper[4716]: I1207 16:27:33.669158 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9aec4026-f10a-446e-b68d-e90f10680919" path="/var/lib/kubelet/pods/9aec4026-f10a-446e-b68d-e90f10680919/volumes" Dec 07 16:27:34 crc kubenswrapper[4716]: I1207 16:27:34.531240 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-lrqg9"] Dec 07 16:27:34 crc kubenswrapper[4716]: E1207 16:27:34.531681 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9aec4026-f10a-446e-b68d-e90f10680919" containerName="extract-content" Dec 07 16:27:34 crc kubenswrapper[4716]: I1207 16:27:34.531704 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="9aec4026-f10a-446e-b68d-e90f10680919" containerName="extract-content" Dec 07 16:27:34 crc kubenswrapper[4716]: E1207 16:27:34.531729 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9aec4026-f10a-446e-b68d-e90f10680919" containerName="registry-server" Dec 07 16:27:34 crc kubenswrapper[4716]: I1207 16:27:34.531739 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="9aec4026-f10a-446e-b68d-e90f10680919" containerName="registry-server" Dec 07 16:27:34 crc kubenswrapper[4716]: E1207 16:27:34.531754 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9aec4026-f10a-446e-b68d-e90f10680919" containerName="extract-utilities" Dec 07 16:27:34 crc kubenswrapper[4716]: I1207 16:27:34.531762 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="9aec4026-f10a-446e-b68d-e90f10680919" containerName="extract-utilities" Dec 07 16:27:34 crc kubenswrapper[4716]: I1207 16:27:34.532022 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="9aec4026-f10a-446e-b68d-e90f10680919" containerName="registry-server" Dec 07 16:27:34 crc kubenswrapper[4716]: I1207 16:27:34.533832 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lrqg9" Dec 07 16:27:34 crc kubenswrapper[4716]: I1207 16:27:34.556898 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lrqg9"] Dec 07 16:27:34 crc kubenswrapper[4716]: I1207 16:27:34.681564 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ff9d0bf-af4c-46b9-886e-6ddbe9452fa0-utilities\") pod \"redhat-operators-lrqg9\" (UID: \"6ff9d0bf-af4c-46b9-886e-6ddbe9452fa0\") " pod="openshift-marketplace/redhat-operators-lrqg9" Dec 07 16:27:34 crc kubenswrapper[4716]: I1207 16:27:34.681640 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ff9d0bf-af4c-46b9-886e-6ddbe9452fa0-catalog-content\") pod \"redhat-operators-lrqg9\" (UID: \"6ff9d0bf-af4c-46b9-886e-6ddbe9452fa0\") " pod="openshift-marketplace/redhat-operators-lrqg9" Dec 07 16:27:34 crc kubenswrapper[4716]: I1207 16:27:34.681865 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hgwtp\" (UniqueName: \"kubernetes.io/projected/6ff9d0bf-af4c-46b9-886e-6ddbe9452fa0-kube-api-access-hgwtp\") pod \"redhat-operators-lrqg9\" (UID: \"6ff9d0bf-af4c-46b9-886e-6ddbe9452fa0\") " pod="openshift-marketplace/redhat-operators-lrqg9" Dec 07 16:27:34 crc kubenswrapper[4716]: I1207 16:27:34.784260 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ff9d0bf-af4c-46b9-886e-6ddbe9452fa0-utilities\") pod \"redhat-operators-lrqg9\" (UID: \"6ff9d0bf-af4c-46b9-886e-6ddbe9452fa0\") " pod="openshift-marketplace/redhat-operators-lrqg9" Dec 07 16:27:34 crc kubenswrapper[4716]: I1207 16:27:34.784790 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ff9d0bf-af4c-46b9-886e-6ddbe9452fa0-catalog-content\") pod \"redhat-operators-lrqg9\" (UID: \"6ff9d0bf-af4c-46b9-886e-6ddbe9452fa0\") " pod="openshift-marketplace/redhat-operators-lrqg9" Dec 07 16:27:34 crc kubenswrapper[4716]: I1207 16:27:34.784848 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ff9d0bf-af4c-46b9-886e-6ddbe9452fa0-utilities\") pod \"redhat-operators-lrqg9\" (UID: \"6ff9d0bf-af4c-46b9-886e-6ddbe9452fa0\") " pod="openshift-marketplace/redhat-operators-lrqg9" Dec 07 16:27:34 crc kubenswrapper[4716]: I1207 16:27:34.785010 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hgwtp\" (UniqueName: \"kubernetes.io/projected/6ff9d0bf-af4c-46b9-886e-6ddbe9452fa0-kube-api-access-hgwtp\") pod \"redhat-operators-lrqg9\" (UID: \"6ff9d0bf-af4c-46b9-886e-6ddbe9452fa0\") " pod="openshift-marketplace/redhat-operators-lrqg9" Dec 07 16:27:34 crc kubenswrapper[4716]: I1207 16:27:34.785433 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ff9d0bf-af4c-46b9-886e-6ddbe9452fa0-catalog-content\") pod \"redhat-operators-lrqg9\" (UID: \"6ff9d0bf-af4c-46b9-886e-6ddbe9452fa0\") " pod="openshift-marketplace/redhat-operators-lrqg9" Dec 07 16:27:34 crc kubenswrapper[4716]: I1207 16:27:34.819340 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hgwtp\" (UniqueName: \"kubernetes.io/projected/6ff9d0bf-af4c-46b9-886e-6ddbe9452fa0-kube-api-access-hgwtp\") pod \"redhat-operators-lrqg9\" (UID: \"6ff9d0bf-af4c-46b9-886e-6ddbe9452fa0\") " pod="openshift-marketplace/redhat-operators-lrqg9" Dec 07 16:27:34 crc kubenswrapper[4716]: I1207 16:27:34.863094 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lrqg9" Dec 07 16:27:35 crc kubenswrapper[4716]: W1207 16:27:35.332752 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6ff9d0bf_af4c_46b9_886e_6ddbe9452fa0.slice/crio-bdb6c73fef5cdd7fd75878ddf22023da9f36a11489980018ca2b4eedc2867f76 WatchSource:0}: Error finding container bdb6c73fef5cdd7fd75878ddf22023da9f36a11489980018ca2b4eedc2867f76: Status 404 returned error can't find the container with id bdb6c73fef5cdd7fd75878ddf22023da9f36a11489980018ca2b4eedc2867f76 Dec 07 16:27:35 crc kubenswrapper[4716]: I1207 16:27:35.334202 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lrqg9"] Dec 07 16:27:36 crc kubenswrapper[4716]: I1207 16:27:36.157674 4716 generic.go:334] "Generic (PLEG): container finished" podID="6ff9d0bf-af4c-46b9-886e-6ddbe9452fa0" containerID="d02a111cb6af9e131acfc871a3a1ba058bdd1d72076540606e58ee77087c7197" exitCode=0 Dec 07 16:27:36 crc kubenswrapper[4716]: I1207 16:27:36.157730 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lrqg9" event={"ID":"6ff9d0bf-af4c-46b9-886e-6ddbe9452fa0","Type":"ContainerDied","Data":"d02a111cb6af9e131acfc871a3a1ba058bdd1d72076540606e58ee77087c7197"} Dec 07 16:27:36 crc kubenswrapper[4716]: I1207 16:27:36.158090 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lrqg9" event={"ID":"6ff9d0bf-af4c-46b9-886e-6ddbe9452fa0","Type":"ContainerStarted","Data":"bdb6c73fef5cdd7fd75878ddf22023da9f36a11489980018ca2b4eedc2867f76"} Dec 07 16:27:37 crc kubenswrapper[4716]: I1207 16:27:37.181478 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lrqg9" event={"ID":"6ff9d0bf-af4c-46b9-886e-6ddbe9452fa0","Type":"ContainerStarted","Data":"f84682dd1ebb4daafc4d2a4639b8d38b06de2c289c5b972599319a54ed89b637"} Dec 07 16:27:39 crc kubenswrapper[4716]: I1207 16:27:39.203939 4716 generic.go:334] "Generic (PLEG): container finished" podID="6ff9d0bf-af4c-46b9-886e-6ddbe9452fa0" containerID="f84682dd1ebb4daafc4d2a4639b8d38b06de2c289c5b972599319a54ed89b637" exitCode=0 Dec 07 16:27:39 crc kubenswrapper[4716]: I1207 16:27:39.203990 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lrqg9" event={"ID":"6ff9d0bf-af4c-46b9-886e-6ddbe9452fa0","Type":"ContainerDied","Data":"f84682dd1ebb4daafc4d2a4639b8d38b06de2c289c5b972599319a54ed89b637"} Dec 07 16:27:40 crc kubenswrapper[4716]: I1207 16:27:40.214406 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lrqg9" event={"ID":"6ff9d0bf-af4c-46b9-886e-6ddbe9452fa0","Type":"ContainerStarted","Data":"5c22f773ab9c76232fa643f58a136d3f51734154162028d8c295ab5c6986713a"} Dec 07 16:27:40 crc kubenswrapper[4716]: I1207 16:27:40.237716 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-lrqg9" podStartSLOduration=2.73372933 podStartE2EDuration="6.237668307s" podCreationTimestamp="2025-12-07 16:27:34 +0000 UTC" firstStartedPulling="2025-12-07 16:27:36.159751165 +0000 UTC m=+1518.850036077" lastFinishedPulling="2025-12-07 16:27:39.663690142 +0000 UTC m=+1522.353975054" observedRunningTime="2025-12-07 16:27:40.231198802 +0000 UTC m=+1522.921483714" watchObservedRunningTime="2025-12-07 16:27:40.237668307 +0000 UTC m=+1522.927953219" Dec 07 16:27:44 crc kubenswrapper[4716]: I1207 16:27:44.863678 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-lrqg9" Dec 07 16:27:44 crc kubenswrapper[4716]: I1207 16:27:44.864237 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-lrqg9" Dec 07 16:27:45 crc kubenswrapper[4716]: I1207 16:27:45.909228 4716 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-lrqg9" podUID="6ff9d0bf-af4c-46b9-886e-6ddbe9452fa0" containerName="registry-server" probeResult="failure" output=< Dec 07 16:27:45 crc kubenswrapper[4716]: timeout: failed to connect service ":50051" within 1s Dec 07 16:27:45 crc kubenswrapper[4716]: > Dec 07 16:27:54 crc kubenswrapper[4716]: I1207 16:27:54.925781 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-lrqg9" Dec 07 16:27:54 crc kubenswrapper[4716]: I1207 16:27:54.992041 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-lrqg9" Dec 07 16:27:55 crc kubenswrapper[4716]: I1207 16:27:55.165170 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lrqg9"] Dec 07 16:27:56 crc kubenswrapper[4716]: I1207 16:27:56.397325 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-lrqg9" podUID="6ff9d0bf-af4c-46b9-886e-6ddbe9452fa0" containerName="registry-server" containerID="cri-o://5c22f773ab9c76232fa643f58a136d3f51734154162028d8c295ab5c6986713a" gracePeriod=2 Dec 07 16:27:56 crc kubenswrapper[4716]: I1207 16:27:56.809818 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lrqg9" Dec 07 16:27:56 crc kubenswrapper[4716]: I1207 16:27:56.923700 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hgwtp\" (UniqueName: \"kubernetes.io/projected/6ff9d0bf-af4c-46b9-886e-6ddbe9452fa0-kube-api-access-hgwtp\") pod \"6ff9d0bf-af4c-46b9-886e-6ddbe9452fa0\" (UID: \"6ff9d0bf-af4c-46b9-886e-6ddbe9452fa0\") " Dec 07 16:27:56 crc kubenswrapper[4716]: I1207 16:27:56.923867 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ff9d0bf-af4c-46b9-886e-6ddbe9452fa0-utilities\") pod \"6ff9d0bf-af4c-46b9-886e-6ddbe9452fa0\" (UID: \"6ff9d0bf-af4c-46b9-886e-6ddbe9452fa0\") " Dec 07 16:27:56 crc kubenswrapper[4716]: I1207 16:27:56.924124 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ff9d0bf-af4c-46b9-886e-6ddbe9452fa0-catalog-content\") pod \"6ff9d0bf-af4c-46b9-886e-6ddbe9452fa0\" (UID: \"6ff9d0bf-af4c-46b9-886e-6ddbe9452fa0\") " Dec 07 16:27:56 crc kubenswrapper[4716]: I1207 16:27:56.926012 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6ff9d0bf-af4c-46b9-886e-6ddbe9452fa0-utilities" (OuterVolumeSpecName: "utilities") pod "6ff9d0bf-af4c-46b9-886e-6ddbe9452fa0" (UID: "6ff9d0bf-af4c-46b9-886e-6ddbe9452fa0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:27:56 crc kubenswrapper[4716]: I1207 16:27:56.931398 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ff9d0bf-af4c-46b9-886e-6ddbe9452fa0-kube-api-access-hgwtp" (OuterVolumeSpecName: "kube-api-access-hgwtp") pod "6ff9d0bf-af4c-46b9-886e-6ddbe9452fa0" (UID: "6ff9d0bf-af4c-46b9-886e-6ddbe9452fa0"). InnerVolumeSpecName "kube-api-access-hgwtp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:27:57 crc kubenswrapper[4716]: I1207 16:27:57.027866 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hgwtp\" (UniqueName: \"kubernetes.io/projected/6ff9d0bf-af4c-46b9-886e-6ddbe9452fa0-kube-api-access-hgwtp\") on node \"crc\" DevicePath \"\"" Dec 07 16:27:57 crc kubenswrapper[4716]: I1207 16:27:57.027898 4716 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ff9d0bf-af4c-46b9-886e-6ddbe9452fa0-utilities\") on node \"crc\" DevicePath \"\"" Dec 07 16:27:57 crc kubenswrapper[4716]: I1207 16:27:57.042568 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6ff9d0bf-af4c-46b9-886e-6ddbe9452fa0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6ff9d0bf-af4c-46b9-886e-6ddbe9452fa0" (UID: "6ff9d0bf-af4c-46b9-886e-6ddbe9452fa0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:27:57 crc kubenswrapper[4716]: I1207 16:27:57.129795 4716 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ff9d0bf-af4c-46b9-886e-6ddbe9452fa0-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 07 16:27:57 crc kubenswrapper[4716]: I1207 16:27:57.443026 4716 generic.go:334] "Generic (PLEG): container finished" podID="6ff9d0bf-af4c-46b9-886e-6ddbe9452fa0" containerID="5c22f773ab9c76232fa643f58a136d3f51734154162028d8c295ab5c6986713a" exitCode=0 Dec 07 16:27:57 crc kubenswrapper[4716]: I1207 16:27:57.443206 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lrqg9" Dec 07 16:27:57 crc kubenswrapper[4716]: I1207 16:27:57.444046 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lrqg9" event={"ID":"6ff9d0bf-af4c-46b9-886e-6ddbe9452fa0","Type":"ContainerDied","Data":"5c22f773ab9c76232fa643f58a136d3f51734154162028d8c295ab5c6986713a"} Dec 07 16:27:57 crc kubenswrapper[4716]: I1207 16:27:57.444193 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lrqg9" event={"ID":"6ff9d0bf-af4c-46b9-886e-6ddbe9452fa0","Type":"ContainerDied","Data":"bdb6c73fef5cdd7fd75878ddf22023da9f36a11489980018ca2b4eedc2867f76"} Dec 07 16:27:57 crc kubenswrapper[4716]: I1207 16:27:57.444301 4716 scope.go:117] "RemoveContainer" containerID="5c22f773ab9c76232fa643f58a136d3f51734154162028d8c295ab5c6986713a" Dec 07 16:27:57 crc kubenswrapper[4716]: I1207 16:27:57.486482 4716 scope.go:117] "RemoveContainer" containerID="f84682dd1ebb4daafc4d2a4639b8d38b06de2c289c5b972599319a54ed89b637" Dec 07 16:27:57 crc kubenswrapper[4716]: I1207 16:27:57.500279 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lrqg9"] Dec 07 16:27:57 crc kubenswrapper[4716]: I1207 16:27:57.515834 4716 scope.go:117] "RemoveContainer" containerID="d02a111cb6af9e131acfc871a3a1ba058bdd1d72076540606e58ee77087c7197" Dec 07 16:27:57 crc kubenswrapper[4716]: I1207 16:27:57.518031 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-lrqg9"] Dec 07 16:27:57 crc kubenswrapper[4716]: I1207 16:27:57.581238 4716 scope.go:117] "RemoveContainer" containerID="5c22f773ab9c76232fa643f58a136d3f51734154162028d8c295ab5c6986713a" Dec 07 16:27:57 crc kubenswrapper[4716]: E1207 16:27:57.581913 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c22f773ab9c76232fa643f58a136d3f51734154162028d8c295ab5c6986713a\": container with ID starting with 5c22f773ab9c76232fa643f58a136d3f51734154162028d8c295ab5c6986713a not found: ID does not exist" containerID="5c22f773ab9c76232fa643f58a136d3f51734154162028d8c295ab5c6986713a" Dec 07 16:27:57 crc kubenswrapper[4716]: I1207 16:27:57.581976 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c22f773ab9c76232fa643f58a136d3f51734154162028d8c295ab5c6986713a"} err="failed to get container status \"5c22f773ab9c76232fa643f58a136d3f51734154162028d8c295ab5c6986713a\": rpc error: code = NotFound desc = could not find container \"5c22f773ab9c76232fa643f58a136d3f51734154162028d8c295ab5c6986713a\": container with ID starting with 5c22f773ab9c76232fa643f58a136d3f51734154162028d8c295ab5c6986713a not found: ID does not exist" Dec 07 16:27:57 crc kubenswrapper[4716]: I1207 16:27:57.582013 4716 scope.go:117] "RemoveContainer" containerID="f84682dd1ebb4daafc4d2a4639b8d38b06de2c289c5b972599319a54ed89b637" Dec 07 16:27:57 crc kubenswrapper[4716]: E1207 16:27:57.582424 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f84682dd1ebb4daafc4d2a4639b8d38b06de2c289c5b972599319a54ed89b637\": container with ID starting with f84682dd1ebb4daafc4d2a4639b8d38b06de2c289c5b972599319a54ed89b637 not found: ID does not exist" containerID="f84682dd1ebb4daafc4d2a4639b8d38b06de2c289c5b972599319a54ed89b637" Dec 07 16:27:57 crc kubenswrapper[4716]: I1207 16:27:57.582474 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f84682dd1ebb4daafc4d2a4639b8d38b06de2c289c5b972599319a54ed89b637"} err="failed to get container status \"f84682dd1ebb4daafc4d2a4639b8d38b06de2c289c5b972599319a54ed89b637\": rpc error: code = NotFound desc = could not find container \"f84682dd1ebb4daafc4d2a4639b8d38b06de2c289c5b972599319a54ed89b637\": container with ID starting with f84682dd1ebb4daafc4d2a4639b8d38b06de2c289c5b972599319a54ed89b637 not found: ID does not exist" Dec 07 16:27:57 crc kubenswrapper[4716]: I1207 16:27:57.582508 4716 scope.go:117] "RemoveContainer" containerID="d02a111cb6af9e131acfc871a3a1ba058bdd1d72076540606e58ee77087c7197" Dec 07 16:27:57 crc kubenswrapper[4716]: E1207 16:27:57.582825 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d02a111cb6af9e131acfc871a3a1ba058bdd1d72076540606e58ee77087c7197\": container with ID starting with d02a111cb6af9e131acfc871a3a1ba058bdd1d72076540606e58ee77087c7197 not found: ID does not exist" containerID="d02a111cb6af9e131acfc871a3a1ba058bdd1d72076540606e58ee77087c7197" Dec 07 16:27:57 crc kubenswrapper[4716]: I1207 16:27:57.582876 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d02a111cb6af9e131acfc871a3a1ba058bdd1d72076540606e58ee77087c7197"} err="failed to get container status \"d02a111cb6af9e131acfc871a3a1ba058bdd1d72076540606e58ee77087c7197\": rpc error: code = NotFound desc = could not find container \"d02a111cb6af9e131acfc871a3a1ba058bdd1d72076540606e58ee77087c7197\": container with ID starting with d02a111cb6af9e131acfc871a3a1ba058bdd1d72076540606e58ee77087c7197 not found: ID does not exist" Dec 07 16:27:57 crc kubenswrapper[4716]: I1207 16:27:57.670915 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ff9d0bf-af4c-46b9-886e-6ddbe9452fa0" path="/var/lib/kubelet/pods/6ff9d0bf-af4c-46b9-886e-6ddbe9452fa0/volumes" Dec 07 16:28:10 crc kubenswrapper[4716]: I1207 16:28:10.047678 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-zkgtv"] Dec 07 16:28:10 crc kubenswrapper[4716]: I1207 16:28:10.058659 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-zkgtv"] Dec 07 16:28:10 crc kubenswrapper[4716]: I1207 16:28:10.068588 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-92a4-account-create-update-pzqm6"] Dec 07 16:28:10 crc kubenswrapper[4716]: I1207 16:28:10.075715 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-92a4-account-create-update-pzqm6"] Dec 07 16:28:11 crc kubenswrapper[4716]: I1207 16:28:11.674826 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a8a9543-3bd8-4f32-8e70-82eea76874df" path="/var/lib/kubelet/pods/3a8a9543-3bd8-4f32-8e70-82eea76874df/volumes" Dec 07 16:28:11 crc kubenswrapper[4716]: I1207 16:28:11.676807 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="708e4fff-969d-4753-9223-7ecbe84890a9" path="/var/lib/kubelet/pods/708e4fff-969d-4753-9223-7ecbe84890a9/volumes" Dec 07 16:28:20 crc kubenswrapper[4716]: I1207 16:28:20.046160 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-f5b6-account-create-update-b7hp7"] Dec 07 16:28:20 crc kubenswrapper[4716]: I1207 16:28:20.060794 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-c4d5d"] Dec 07 16:28:20 crc kubenswrapper[4716]: I1207 16:28:20.072528 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-c4d5d"] Dec 07 16:28:20 crc kubenswrapper[4716]: I1207 16:28:20.080726 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-f5b6-account-create-update-b7hp7"] Dec 07 16:28:20 crc kubenswrapper[4716]: I1207 16:28:20.088156 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-ce39-account-create-update-5f2ml"] Dec 07 16:28:20 crc kubenswrapper[4716]: I1207 16:28:20.096811 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-7w7t8"] Dec 07 16:28:20 crc kubenswrapper[4716]: I1207 16:28:20.106891 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-ce39-account-create-update-5f2ml"] Dec 07 16:28:20 crc kubenswrapper[4716]: I1207 16:28:20.118364 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-7w7t8"] Dec 07 16:28:21 crc kubenswrapper[4716]: I1207 16:28:21.672941 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c2f879d-d624-4537-9ab4-f4076614d82e" path="/var/lib/kubelet/pods/1c2f879d-d624-4537-9ab4-f4076614d82e/volumes" Dec 07 16:28:21 crc kubenswrapper[4716]: I1207 16:28:21.674617 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ce6b31a-2beb-4fc8-ba79-07cb3f4a7ec8" path="/var/lib/kubelet/pods/1ce6b31a-2beb-4fc8-ba79-07cb3f4a7ec8/volumes" Dec 07 16:28:21 crc kubenswrapper[4716]: I1207 16:28:21.676339 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f2a2677-eb1a-40c1-a947-c3347f6c8c5a" path="/var/lib/kubelet/pods/9f2a2677-eb1a-40c1-a947-c3347f6c8c5a/volumes" Dec 07 16:28:21 crc kubenswrapper[4716]: I1207 16:28:21.677782 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db19870c-0848-4366-8e07-f6028a151c86" path="/var/lib/kubelet/pods/db19870c-0848-4366-8e07-f6028a151c86/volumes" Dec 07 16:28:32 crc kubenswrapper[4716]: I1207 16:28:32.680144 4716 scope.go:117] "RemoveContainer" containerID="30d3d72035887670236040e883c7c2e5b42e0a519c96902c5b92be3276ab017a" Dec 07 16:28:32 crc kubenswrapper[4716]: I1207 16:28:32.706421 4716 scope.go:117] "RemoveContainer" containerID="97495ae9bc006e7a09396ed08e1ff1a2212153c8be2a793c8ea9ff8c6182ddf1" Dec 07 16:28:32 crc kubenswrapper[4716]: I1207 16:28:32.751148 4716 scope.go:117] "RemoveContainer" containerID="13927b13b0cabeb3877b59bc7c647efd771bc28b6871f9c1cc5308712fa4d0e3" Dec 07 16:28:32 crc kubenswrapper[4716]: I1207 16:28:32.801157 4716 scope.go:117] "RemoveContainer" containerID="510e1c2305c4e0f7b4156bbad7d847d8690ba86d7fd7e2cd035b8d2aaa5a1421" Dec 07 16:28:32 crc kubenswrapper[4716]: I1207 16:28:32.841770 4716 scope.go:117] "RemoveContainer" containerID="2dd4f0eccf36088a3dd7414d172b860996761216771cc27440379f05438017da" Dec 07 16:28:32 crc kubenswrapper[4716]: I1207 16:28:32.882890 4716 scope.go:117] "RemoveContainer" containerID="b5f5cd456586aa7da6430b4cbeb4d206921c062c830d0665cccb4657c7bcf4f5" Dec 07 16:28:32 crc kubenswrapper[4716]: I1207 16:28:32.921445 4716 scope.go:117] "RemoveContainer" containerID="7ee46a0bc54b0c8ca6cfa63abfeb8f724ab05bbea382bf6025caf8c59b98af4e" Dec 07 16:28:32 crc kubenswrapper[4716]: I1207 16:28:32.944809 4716 scope.go:117] "RemoveContainer" containerID="96a44c2634479106efbff992e763b0c05a596d0ebc22f57a5e6740f60c32c468" Dec 07 16:28:38 crc kubenswrapper[4716]: I1207 16:28:38.034358 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-gq62d"] Dec 07 16:28:38 crc kubenswrapper[4716]: I1207 16:28:38.051248 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-gq62d"] Dec 07 16:28:39 crc kubenswrapper[4716]: I1207 16:28:39.668929 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a17893a0-50ef-4d05-8aaf-c6a9e3f8d20d" path="/var/lib/kubelet/pods/a17893a0-50ef-4d05-8aaf-c6a9e3f8d20d/volumes" Dec 07 16:28:55 crc kubenswrapper[4716]: I1207 16:28:55.015791 4716 generic.go:334] "Generic (PLEG): container finished" podID="7495cf42-56aa-43c9-9d24-a022a3e50505" containerID="522ffd22fe63962dd32ec2edd14e872c82266d33517f7415098481badc213f50" exitCode=0 Dec 07 16:28:55 crc kubenswrapper[4716]: I1207 16:28:55.015887 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9d8h7" event={"ID":"7495cf42-56aa-43c9-9d24-a022a3e50505","Type":"ContainerDied","Data":"522ffd22fe63962dd32ec2edd14e872c82266d33517f7415098481badc213f50"} Dec 07 16:28:56 crc kubenswrapper[4716]: I1207 16:28:56.071586 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-9210-account-create-update-lvz68"] Dec 07 16:28:56 crc kubenswrapper[4716]: I1207 16:28:56.094808 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-t7n4f"] Dec 07 16:28:56 crc kubenswrapper[4716]: I1207 16:28:56.115471 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-e8e1-account-create-update-frpnc"] Dec 07 16:28:56 crc kubenswrapper[4716]: I1207 16:28:56.125010 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-jct5c"] Dec 07 16:28:56 crc kubenswrapper[4716]: I1207 16:28:56.132915 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-9b37-account-create-update-rz5hs"] Dec 07 16:28:56 crc kubenswrapper[4716]: I1207 16:28:56.142586 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-9210-account-create-update-lvz68"] Dec 07 16:28:56 crc kubenswrapper[4716]: I1207 16:28:56.149961 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-jh7l6"] Dec 07 16:28:56 crc kubenswrapper[4716]: I1207 16:28:56.157480 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-t7n4f"] Dec 07 16:28:56 crc kubenswrapper[4716]: I1207 16:28:56.165428 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-jct5c"] Dec 07 16:28:56 crc kubenswrapper[4716]: I1207 16:28:56.173346 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-e8e1-account-create-update-frpnc"] Dec 07 16:28:56 crc kubenswrapper[4716]: I1207 16:28:56.181675 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-jh7l6"] Dec 07 16:28:56 crc kubenswrapper[4716]: I1207 16:28:56.189564 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-9b37-account-create-update-rz5hs"] Dec 07 16:28:56 crc kubenswrapper[4716]: I1207 16:28:56.443023 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9d8h7" Dec 07 16:28:56 crc kubenswrapper[4716]: I1207 16:28:56.574924 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7495cf42-56aa-43c9-9d24-a022a3e50505-inventory\") pod \"7495cf42-56aa-43c9-9d24-a022a3e50505\" (UID: \"7495cf42-56aa-43c9-9d24-a022a3e50505\") " Dec 07 16:28:56 crc kubenswrapper[4716]: I1207 16:28:56.575164 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7495cf42-56aa-43c9-9d24-a022a3e50505-ssh-key\") pod \"7495cf42-56aa-43c9-9d24-a022a3e50505\" (UID: \"7495cf42-56aa-43c9-9d24-a022a3e50505\") " Dec 07 16:28:56 crc kubenswrapper[4716]: I1207 16:28:56.575346 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t2w77\" (UniqueName: \"kubernetes.io/projected/7495cf42-56aa-43c9-9d24-a022a3e50505-kube-api-access-t2w77\") pod \"7495cf42-56aa-43c9-9d24-a022a3e50505\" (UID: \"7495cf42-56aa-43c9-9d24-a022a3e50505\") " Dec 07 16:28:56 crc kubenswrapper[4716]: I1207 16:28:56.580495 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7495cf42-56aa-43c9-9d24-a022a3e50505-kube-api-access-t2w77" (OuterVolumeSpecName: "kube-api-access-t2w77") pod "7495cf42-56aa-43c9-9d24-a022a3e50505" (UID: "7495cf42-56aa-43c9-9d24-a022a3e50505"). InnerVolumeSpecName "kube-api-access-t2w77". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:28:56 crc kubenswrapper[4716]: I1207 16:28:56.604497 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7495cf42-56aa-43c9-9d24-a022a3e50505-inventory" (OuterVolumeSpecName: "inventory") pod "7495cf42-56aa-43c9-9d24-a022a3e50505" (UID: "7495cf42-56aa-43c9-9d24-a022a3e50505"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:28:56 crc kubenswrapper[4716]: I1207 16:28:56.608700 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7495cf42-56aa-43c9-9d24-a022a3e50505-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7495cf42-56aa-43c9-9d24-a022a3e50505" (UID: "7495cf42-56aa-43c9-9d24-a022a3e50505"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:28:56 crc kubenswrapper[4716]: I1207 16:28:56.677445 4716 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7495cf42-56aa-43c9-9d24-a022a3e50505-inventory\") on node \"crc\" DevicePath \"\"" Dec 07 16:28:56 crc kubenswrapper[4716]: I1207 16:28:56.677478 4716 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7495cf42-56aa-43c9-9d24-a022a3e50505-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 07 16:28:56 crc kubenswrapper[4716]: I1207 16:28:56.677488 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t2w77\" (UniqueName: \"kubernetes.io/projected/7495cf42-56aa-43c9-9d24-a022a3e50505-kube-api-access-t2w77\") on node \"crc\" DevicePath \"\"" Dec 07 16:28:57 crc kubenswrapper[4716]: I1207 16:28:57.052960 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9d8h7" event={"ID":"7495cf42-56aa-43c9-9d24-a022a3e50505","Type":"ContainerDied","Data":"0b681ad96fa545e4d53487591674d93e63b58e02153586c274b0cac81ed3d490"} Dec 07 16:28:57 crc kubenswrapper[4716]: I1207 16:28:57.053363 4716 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0b681ad96fa545e4d53487591674d93e63b58e02153586c274b0cac81ed3d490" Dec 07 16:28:57 crc kubenswrapper[4716]: I1207 16:28:57.054304 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9d8h7" Dec 07 16:28:57 crc kubenswrapper[4716]: I1207 16:28:57.135920 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zsnd4"] Dec 07 16:28:57 crc kubenswrapper[4716]: E1207 16:28:57.136326 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7495cf42-56aa-43c9-9d24-a022a3e50505" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 07 16:28:57 crc kubenswrapper[4716]: I1207 16:28:57.136342 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="7495cf42-56aa-43c9-9d24-a022a3e50505" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 07 16:28:57 crc kubenswrapper[4716]: E1207 16:28:57.136355 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ff9d0bf-af4c-46b9-886e-6ddbe9452fa0" containerName="registry-server" Dec 07 16:28:57 crc kubenswrapper[4716]: I1207 16:28:57.136364 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ff9d0bf-af4c-46b9-886e-6ddbe9452fa0" containerName="registry-server" Dec 07 16:28:57 crc kubenswrapper[4716]: E1207 16:28:57.136395 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ff9d0bf-af4c-46b9-886e-6ddbe9452fa0" containerName="extract-content" Dec 07 16:28:57 crc kubenswrapper[4716]: I1207 16:28:57.136403 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ff9d0bf-af4c-46b9-886e-6ddbe9452fa0" containerName="extract-content" Dec 07 16:28:57 crc kubenswrapper[4716]: E1207 16:28:57.136421 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ff9d0bf-af4c-46b9-886e-6ddbe9452fa0" containerName="extract-utilities" Dec 07 16:28:57 crc kubenswrapper[4716]: I1207 16:28:57.136428 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ff9d0bf-af4c-46b9-886e-6ddbe9452fa0" containerName="extract-utilities" Dec 07 16:28:57 crc kubenswrapper[4716]: I1207 16:28:57.136635 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ff9d0bf-af4c-46b9-886e-6ddbe9452fa0" containerName="registry-server" Dec 07 16:28:57 crc kubenswrapper[4716]: I1207 16:28:57.136670 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="7495cf42-56aa-43c9-9d24-a022a3e50505" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 07 16:28:57 crc kubenswrapper[4716]: I1207 16:28:57.137961 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zsnd4" Dec 07 16:28:57 crc kubenswrapper[4716]: I1207 16:28:57.146963 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zsnd4"] Dec 07 16:28:57 crc kubenswrapper[4716]: I1207 16:28:57.190623 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 07 16:28:57 crc kubenswrapper[4716]: I1207 16:28:57.190623 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 07 16:28:57 crc kubenswrapper[4716]: I1207 16:28:57.190830 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 07 16:28:57 crc kubenswrapper[4716]: I1207 16:28:57.191015 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-h9pt2" Dec 07 16:28:57 crc kubenswrapper[4716]: I1207 16:28:57.294463 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d2fcfd64-6b34-4f9c-9e8b-3715df35f92d-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-zsnd4\" (UID: \"d2fcfd64-6b34-4f9c-9e8b-3715df35f92d\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zsnd4" Dec 07 16:28:57 crc kubenswrapper[4716]: I1207 16:28:57.294519 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d2fcfd64-6b34-4f9c-9e8b-3715df35f92d-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-zsnd4\" (UID: \"d2fcfd64-6b34-4f9c-9e8b-3715df35f92d\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zsnd4" Dec 07 16:28:57 crc kubenswrapper[4716]: I1207 16:28:57.294770 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x66wj\" (UniqueName: \"kubernetes.io/projected/d2fcfd64-6b34-4f9c-9e8b-3715df35f92d-kube-api-access-x66wj\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-zsnd4\" (UID: \"d2fcfd64-6b34-4f9c-9e8b-3715df35f92d\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zsnd4" Dec 07 16:28:57 crc kubenswrapper[4716]: I1207 16:28:57.396332 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d2fcfd64-6b34-4f9c-9e8b-3715df35f92d-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-zsnd4\" (UID: \"d2fcfd64-6b34-4f9c-9e8b-3715df35f92d\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zsnd4" Dec 07 16:28:57 crc kubenswrapper[4716]: I1207 16:28:57.396390 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d2fcfd64-6b34-4f9c-9e8b-3715df35f92d-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-zsnd4\" (UID: \"d2fcfd64-6b34-4f9c-9e8b-3715df35f92d\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zsnd4" Dec 07 16:28:57 crc kubenswrapper[4716]: I1207 16:28:57.396436 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x66wj\" (UniqueName: \"kubernetes.io/projected/d2fcfd64-6b34-4f9c-9e8b-3715df35f92d-kube-api-access-x66wj\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-zsnd4\" (UID: \"d2fcfd64-6b34-4f9c-9e8b-3715df35f92d\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zsnd4" Dec 07 16:28:57 crc kubenswrapper[4716]: I1207 16:28:57.400763 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d2fcfd64-6b34-4f9c-9e8b-3715df35f92d-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-zsnd4\" (UID: \"d2fcfd64-6b34-4f9c-9e8b-3715df35f92d\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zsnd4" Dec 07 16:28:57 crc kubenswrapper[4716]: I1207 16:28:57.408206 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d2fcfd64-6b34-4f9c-9e8b-3715df35f92d-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-zsnd4\" (UID: \"d2fcfd64-6b34-4f9c-9e8b-3715df35f92d\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zsnd4" Dec 07 16:28:57 crc kubenswrapper[4716]: I1207 16:28:57.412592 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x66wj\" (UniqueName: \"kubernetes.io/projected/d2fcfd64-6b34-4f9c-9e8b-3715df35f92d-kube-api-access-x66wj\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-zsnd4\" (UID: \"d2fcfd64-6b34-4f9c-9e8b-3715df35f92d\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zsnd4" Dec 07 16:28:57 crc kubenswrapper[4716]: I1207 16:28:57.521558 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zsnd4" Dec 07 16:28:57 crc kubenswrapper[4716]: I1207 16:28:57.668398 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b362916-6a42-4fae-88be-f364dc650240" path="/var/lib/kubelet/pods/0b362916-6a42-4fae-88be-f364dc650240/volumes" Dec 07 16:28:57 crc kubenswrapper[4716]: I1207 16:28:57.669010 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34a0af5c-2914-4a67-8ede-f178f08d473f" path="/var/lib/kubelet/pods/34a0af5c-2914-4a67-8ede-f178f08d473f/volumes" Dec 07 16:28:57 crc kubenswrapper[4716]: I1207 16:28:57.669625 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="50a30fe7-3fb4-4b9d-b186-139c2b48295b" path="/var/lib/kubelet/pods/50a30fe7-3fb4-4b9d-b186-139c2b48295b/volumes" Dec 07 16:28:57 crc kubenswrapper[4716]: I1207 16:28:57.670189 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="862acceb-bcf0-4ef4-9a53-df61d6084806" path="/var/lib/kubelet/pods/862acceb-bcf0-4ef4-9a53-df61d6084806/volumes" Dec 07 16:28:57 crc kubenswrapper[4716]: I1207 16:28:57.671270 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d1c048e-719c-4406-bfae-1e5a629534ed" path="/var/lib/kubelet/pods/9d1c048e-719c-4406-bfae-1e5a629534ed/volumes" Dec 07 16:28:57 crc kubenswrapper[4716]: I1207 16:28:57.671811 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd43dd05-4a94-4e12-9e2a-788135a9ee76" path="/var/lib/kubelet/pods/bd43dd05-4a94-4e12-9e2a-788135a9ee76/volumes" Dec 07 16:28:57 crc kubenswrapper[4716]: I1207 16:28:57.861353 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zsnd4"] Dec 07 16:28:57 crc kubenswrapper[4716]: W1207 16:28:57.870763 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd2fcfd64_6b34_4f9c_9e8b_3715df35f92d.slice/crio-a6c206dfe32028e091c1346584cccc356f4efa6da998fe31ede501613a7cdaf9 WatchSource:0}: Error finding container a6c206dfe32028e091c1346584cccc356f4efa6da998fe31ede501613a7cdaf9: Status 404 returned error can't find the container with id a6c206dfe32028e091c1346584cccc356f4efa6da998fe31ede501613a7cdaf9 Dec 07 16:28:57 crc kubenswrapper[4716]: I1207 16:28:57.874221 4716 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 07 16:28:58 crc kubenswrapper[4716]: I1207 16:28:58.069809 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zsnd4" event={"ID":"d2fcfd64-6b34-4f9c-9e8b-3715df35f92d","Type":"ContainerStarted","Data":"a6c206dfe32028e091c1346584cccc356f4efa6da998fe31ede501613a7cdaf9"} Dec 07 16:29:00 crc kubenswrapper[4716]: I1207 16:29:00.095643 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zsnd4" event={"ID":"d2fcfd64-6b34-4f9c-9e8b-3715df35f92d","Type":"ContainerStarted","Data":"ff0d720bc814ceb6a875cfecfda943958cd1f3ecef64315ab7853bbef556bb4d"} Dec 07 16:29:00 crc kubenswrapper[4716]: I1207 16:29:00.137997 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zsnd4" podStartSLOduration=1.982857918 podStartE2EDuration="3.13796392s" podCreationTimestamp="2025-12-07 16:28:57 +0000 UTC" firstStartedPulling="2025-12-07 16:28:57.873914411 +0000 UTC m=+1600.564199323" lastFinishedPulling="2025-12-07 16:28:59.029020393 +0000 UTC m=+1601.719305325" observedRunningTime="2025-12-07 16:29:00.121030432 +0000 UTC m=+1602.811315454" watchObservedRunningTime="2025-12-07 16:29:00.13796392 +0000 UTC m=+1602.828248882" Dec 07 16:29:01 crc kubenswrapper[4716]: I1207 16:29:01.037974 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-rxd6q"] Dec 07 16:29:01 crc kubenswrapper[4716]: I1207 16:29:01.056376 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-rxd6q"] Dec 07 16:29:01 crc kubenswrapper[4716]: I1207 16:29:01.671471 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f94a6cb-d8ed-47f6-a90a-fa705985ecaf" path="/var/lib/kubelet/pods/6f94a6cb-d8ed-47f6-a90a-fa705985ecaf/volumes" Dec 07 16:29:22 crc kubenswrapper[4716]: I1207 16:29:22.761560 4716 patch_prober.go:28] interesting pod/machine-config-daemon-zcxxp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 16:29:22 crc kubenswrapper[4716]: I1207 16:29:22.762747 4716 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 16:29:31 crc kubenswrapper[4716]: I1207 16:29:31.036395 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-kl4tr"] Dec 07 16:29:31 crc kubenswrapper[4716]: I1207 16:29:31.047688 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-kl4tr"] Dec 07 16:29:31 crc kubenswrapper[4716]: I1207 16:29:31.669159 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925be76f-fe07-4eb8-982a-02a0a002ea58" path="/var/lib/kubelet/pods/925be76f-fe07-4eb8-982a-02a0a002ea58/volumes" Dec 07 16:29:33 crc kubenswrapper[4716]: I1207 16:29:33.083433 4716 scope.go:117] "RemoveContainer" containerID="44c6371bdef3dc8c2a62f43268f22b3bae06501f812fba3cc8545e388fd6080a" Dec 07 16:29:33 crc kubenswrapper[4716]: I1207 16:29:33.113987 4716 scope.go:117] "RemoveContainer" containerID="75ce43476b8844c678d48c3ccb9c33485489d20a32ba7acddb269a24adfa5af2" Dec 07 16:29:33 crc kubenswrapper[4716]: I1207 16:29:33.182017 4716 scope.go:117] "RemoveContainer" containerID="ac3d925aa12a9311bcb204061c9de1fc84a0224612badc8a11ad34df2ad5a754" Dec 07 16:29:33 crc kubenswrapper[4716]: I1207 16:29:33.246385 4716 scope.go:117] "RemoveContainer" containerID="9451289360f0da6e6a1c2c0b7496a3f85c00d5f1d96494d9b9b614d5cd4bd237" Dec 07 16:29:33 crc kubenswrapper[4716]: I1207 16:29:33.271913 4716 scope.go:117] "RemoveContainer" containerID="b52f6aad7da0f69d63bf00028fce594bb28e1dc0d3c7d743727abbfd62dff3b7" Dec 07 16:29:33 crc kubenswrapper[4716]: I1207 16:29:33.310408 4716 scope.go:117] "RemoveContainer" containerID="f899907d58ebda8ca729b889e0d1a75de6da61c3d90e525bba1d8e11925ba380" Dec 07 16:29:33 crc kubenswrapper[4716]: I1207 16:29:33.371669 4716 scope.go:117] "RemoveContainer" containerID="17f5dc7f54db07da92aa7837da9cf1b03ba4e247d65da80dd553c5fe36814a0f" Dec 07 16:29:33 crc kubenswrapper[4716]: I1207 16:29:33.392016 4716 scope.go:117] "RemoveContainer" containerID="ffa4933c139800dce7b364c70ad2bd22dc6cb74649ec6f2f4c0be145c6b55016" Dec 07 16:29:33 crc kubenswrapper[4716]: I1207 16:29:33.428166 4716 scope.go:117] "RemoveContainer" containerID="2445d34fc45b616bb4cd2f04302cbcf69cc53f93e76a6c8ef579ce0b0b858aca" Dec 07 16:29:42 crc kubenswrapper[4716]: I1207 16:29:42.043284 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-45d84"] Dec 07 16:29:42 crc kubenswrapper[4716]: I1207 16:29:42.052827 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-qqklx"] Dec 07 16:29:42 crc kubenswrapper[4716]: I1207 16:29:42.063267 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-45d84"] Dec 07 16:29:42 crc kubenswrapper[4716]: I1207 16:29:42.073000 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-qqklx"] Dec 07 16:29:43 crc kubenswrapper[4716]: I1207 16:29:43.677998 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6354ab8c-f35e-450f-96f9-8e305f778a54" path="/var/lib/kubelet/pods/6354ab8c-f35e-450f-96f9-8e305f778a54/volumes" Dec 07 16:29:43 crc kubenswrapper[4716]: I1207 16:29:43.679206 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d103f298-33f8-49df-894b-50b813660371" path="/var/lib/kubelet/pods/d103f298-33f8-49df-894b-50b813660371/volumes" Dec 07 16:29:52 crc kubenswrapper[4716]: I1207 16:29:52.761268 4716 patch_prober.go:28] interesting pod/machine-config-daemon-zcxxp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 16:29:52 crc kubenswrapper[4716]: I1207 16:29:52.762232 4716 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 16:29:58 crc kubenswrapper[4716]: I1207 16:29:58.041226 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-gqg6b"] Dec 07 16:29:58 crc kubenswrapper[4716]: I1207 16:29:58.054114 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-tk47q"] Dec 07 16:29:58 crc kubenswrapper[4716]: I1207 16:29:58.062183 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-gqg6b"] Dec 07 16:29:58 crc kubenswrapper[4716]: I1207 16:29:58.068845 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-tk47q"] Dec 07 16:29:59 crc kubenswrapper[4716]: I1207 16:29:59.670108 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="717a326f-b1ac-4fbe-b1c8-b669081dab80" path="/var/lib/kubelet/pods/717a326f-b1ac-4fbe-b1c8-b669081dab80/volumes" Dec 07 16:29:59 crc kubenswrapper[4716]: I1207 16:29:59.670695 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f9413953-056c-4d73-b534-12e6816e6fb4" path="/var/lib/kubelet/pods/f9413953-056c-4d73-b534-12e6816e6fb4/volumes" Dec 07 16:30:00 crc kubenswrapper[4716]: I1207 16:30:00.159316 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29418750-c9svx"] Dec 07 16:30:00 crc kubenswrapper[4716]: I1207 16:30:00.161547 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29418750-c9svx" Dec 07 16:30:00 crc kubenswrapper[4716]: I1207 16:30:00.163809 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 07 16:30:00 crc kubenswrapper[4716]: I1207 16:30:00.166664 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 07 16:30:00 crc kubenswrapper[4716]: I1207 16:30:00.168787 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29418750-c9svx"] Dec 07 16:30:00 crc kubenswrapper[4716]: I1207 16:30:00.360421 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jptrb\" (UniqueName: \"kubernetes.io/projected/15e3e08f-67c0-4d1b-8ecb-29a0f4d8ff44-kube-api-access-jptrb\") pod \"collect-profiles-29418750-c9svx\" (UID: \"15e3e08f-67c0-4d1b-8ecb-29a0f4d8ff44\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418750-c9svx" Dec 07 16:30:00 crc kubenswrapper[4716]: I1207 16:30:00.360469 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/15e3e08f-67c0-4d1b-8ecb-29a0f4d8ff44-config-volume\") pod \"collect-profiles-29418750-c9svx\" (UID: \"15e3e08f-67c0-4d1b-8ecb-29a0f4d8ff44\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418750-c9svx" Dec 07 16:30:00 crc kubenswrapper[4716]: I1207 16:30:00.360631 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/15e3e08f-67c0-4d1b-8ecb-29a0f4d8ff44-secret-volume\") pod \"collect-profiles-29418750-c9svx\" (UID: \"15e3e08f-67c0-4d1b-8ecb-29a0f4d8ff44\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418750-c9svx" Dec 07 16:30:00 crc kubenswrapper[4716]: I1207 16:30:00.463344 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jptrb\" (UniqueName: \"kubernetes.io/projected/15e3e08f-67c0-4d1b-8ecb-29a0f4d8ff44-kube-api-access-jptrb\") pod \"collect-profiles-29418750-c9svx\" (UID: \"15e3e08f-67c0-4d1b-8ecb-29a0f4d8ff44\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418750-c9svx" Dec 07 16:30:00 crc kubenswrapper[4716]: I1207 16:30:00.463654 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/15e3e08f-67c0-4d1b-8ecb-29a0f4d8ff44-config-volume\") pod \"collect-profiles-29418750-c9svx\" (UID: \"15e3e08f-67c0-4d1b-8ecb-29a0f4d8ff44\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418750-c9svx" Dec 07 16:30:00 crc kubenswrapper[4716]: I1207 16:30:00.463732 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/15e3e08f-67c0-4d1b-8ecb-29a0f4d8ff44-secret-volume\") pod \"collect-profiles-29418750-c9svx\" (UID: \"15e3e08f-67c0-4d1b-8ecb-29a0f4d8ff44\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418750-c9svx" Dec 07 16:30:00 crc kubenswrapper[4716]: I1207 16:30:00.464980 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/15e3e08f-67c0-4d1b-8ecb-29a0f4d8ff44-config-volume\") pod \"collect-profiles-29418750-c9svx\" (UID: \"15e3e08f-67c0-4d1b-8ecb-29a0f4d8ff44\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418750-c9svx" Dec 07 16:30:00 crc kubenswrapper[4716]: I1207 16:30:00.474954 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/15e3e08f-67c0-4d1b-8ecb-29a0f4d8ff44-secret-volume\") pod \"collect-profiles-29418750-c9svx\" (UID: \"15e3e08f-67c0-4d1b-8ecb-29a0f4d8ff44\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418750-c9svx" Dec 07 16:30:00 crc kubenswrapper[4716]: I1207 16:30:00.480604 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jptrb\" (UniqueName: \"kubernetes.io/projected/15e3e08f-67c0-4d1b-8ecb-29a0f4d8ff44-kube-api-access-jptrb\") pod \"collect-profiles-29418750-c9svx\" (UID: \"15e3e08f-67c0-4d1b-8ecb-29a0f4d8ff44\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418750-c9svx" Dec 07 16:30:00 crc kubenswrapper[4716]: I1207 16:30:00.490194 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29418750-c9svx" Dec 07 16:30:00 crc kubenswrapper[4716]: I1207 16:30:00.924698 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29418750-c9svx"] Dec 07 16:30:01 crc kubenswrapper[4716]: I1207 16:30:01.718974 4716 generic.go:334] "Generic (PLEG): container finished" podID="15e3e08f-67c0-4d1b-8ecb-29a0f4d8ff44" containerID="daca886a24b8417a78bdbab31ee6a24649798d0954cfcb7350978fb0fb9b062c" exitCode=0 Dec 07 16:30:01 crc kubenswrapper[4716]: I1207 16:30:01.719139 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29418750-c9svx" event={"ID":"15e3e08f-67c0-4d1b-8ecb-29a0f4d8ff44","Type":"ContainerDied","Data":"daca886a24b8417a78bdbab31ee6a24649798d0954cfcb7350978fb0fb9b062c"} Dec 07 16:30:01 crc kubenswrapper[4716]: I1207 16:30:01.719788 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29418750-c9svx" event={"ID":"15e3e08f-67c0-4d1b-8ecb-29a0f4d8ff44","Type":"ContainerStarted","Data":"7766067a1457f393c8426751fc900d65d8ac60c0f338a04a65da8950f97e7499"} Dec 07 16:30:03 crc kubenswrapper[4716]: I1207 16:30:03.106910 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29418750-c9svx" Dec 07 16:30:03 crc kubenswrapper[4716]: I1207 16:30:03.120773 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/15e3e08f-67c0-4d1b-8ecb-29a0f4d8ff44-config-volume\") pod \"15e3e08f-67c0-4d1b-8ecb-29a0f4d8ff44\" (UID: \"15e3e08f-67c0-4d1b-8ecb-29a0f4d8ff44\") " Dec 07 16:30:03 crc kubenswrapper[4716]: I1207 16:30:03.120890 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jptrb\" (UniqueName: \"kubernetes.io/projected/15e3e08f-67c0-4d1b-8ecb-29a0f4d8ff44-kube-api-access-jptrb\") pod \"15e3e08f-67c0-4d1b-8ecb-29a0f4d8ff44\" (UID: \"15e3e08f-67c0-4d1b-8ecb-29a0f4d8ff44\") " Dec 07 16:30:03 crc kubenswrapper[4716]: I1207 16:30:03.120922 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/15e3e08f-67c0-4d1b-8ecb-29a0f4d8ff44-secret-volume\") pod \"15e3e08f-67c0-4d1b-8ecb-29a0f4d8ff44\" (UID: \"15e3e08f-67c0-4d1b-8ecb-29a0f4d8ff44\") " Dec 07 16:30:03 crc kubenswrapper[4716]: I1207 16:30:03.122600 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/15e3e08f-67c0-4d1b-8ecb-29a0f4d8ff44-config-volume" (OuterVolumeSpecName: "config-volume") pod "15e3e08f-67c0-4d1b-8ecb-29a0f4d8ff44" (UID: "15e3e08f-67c0-4d1b-8ecb-29a0f4d8ff44"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:30:03 crc kubenswrapper[4716]: I1207 16:30:03.126697 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15e3e08f-67c0-4d1b-8ecb-29a0f4d8ff44-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "15e3e08f-67c0-4d1b-8ecb-29a0f4d8ff44" (UID: "15e3e08f-67c0-4d1b-8ecb-29a0f4d8ff44"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:30:03 crc kubenswrapper[4716]: I1207 16:30:03.128141 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15e3e08f-67c0-4d1b-8ecb-29a0f4d8ff44-kube-api-access-jptrb" (OuterVolumeSpecName: "kube-api-access-jptrb") pod "15e3e08f-67c0-4d1b-8ecb-29a0f4d8ff44" (UID: "15e3e08f-67c0-4d1b-8ecb-29a0f4d8ff44"). InnerVolumeSpecName "kube-api-access-jptrb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:30:03 crc kubenswrapper[4716]: I1207 16:30:03.222241 4716 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/15e3e08f-67c0-4d1b-8ecb-29a0f4d8ff44-config-volume\") on node \"crc\" DevicePath \"\"" Dec 07 16:30:03 crc kubenswrapper[4716]: I1207 16:30:03.222573 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jptrb\" (UniqueName: \"kubernetes.io/projected/15e3e08f-67c0-4d1b-8ecb-29a0f4d8ff44-kube-api-access-jptrb\") on node \"crc\" DevicePath \"\"" Dec 07 16:30:03 crc kubenswrapper[4716]: I1207 16:30:03.222586 4716 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/15e3e08f-67c0-4d1b-8ecb-29a0f4d8ff44-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 07 16:30:03 crc kubenswrapper[4716]: I1207 16:30:03.744608 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29418750-c9svx" event={"ID":"15e3e08f-67c0-4d1b-8ecb-29a0f4d8ff44","Type":"ContainerDied","Data":"7766067a1457f393c8426751fc900d65d8ac60c0f338a04a65da8950f97e7499"} Dec 07 16:30:03 crc kubenswrapper[4716]: I1207 16:30:03.744650 4716 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7766067a1457f393c8426751fc900d65d8ac60c0f338a04a65da8950f97e7499" Dec 07 16:30:03 crc kubenswrapper[4716]: I1207 16:30:03.744658 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29418750-c9svx" Dec 07 16:30:12 crc kubenswrapper[4716]: I1207 16:30:12.838532 4716 generic.go:334] "Generic (PLEG): container finished" podID="d2fcfd64-6b34-4f9c-9e8b-3715df35f92d" containerID="ff0d720bc814ceb6a875cfecfda943958cd1f3ecef64315ab7853bbef556bb4d" exitCode=0 Dec 07 16:30:12 crc kubenswrapper[4716]: I1207 16:30:12.838586 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zsnd4" event={"ID":"d2fcfd64-6b34-4f9c-9e8b-3715df35f92d","Type":"ContainerDied","Data":"ff0d720bc814ceb6a875cfecfda943958cd1f3ecef64315ab7853bbef556bb4d"} Dec 07 16:30:14 crc kubenswrapper[4716]: I1207 16:30:14.346985 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zsnd4" Dec 07 16:30:14 crc kubenswrapper[4716]: I1207 16:30:14.438690 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d2fcfd64-6b34-4f9c-9e8b-3715df35f92d-ssh-key\") pod \"d2fcfd64-6b34-4f9c-9e8b-3715df35f92d\" (UID: \"d2fcfd64-6b34-4f9c-9e8b-3715df35f92d\") " Dec 07 16:30:14 crc kubenswrapper[4716]: I1207 16:30:14.438779 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x66wj\" (UniqueName: \"kubernetes.io/projected/d2fcfd64-6b34-4f9c-9e8b-3715df35f92d-kube-api-access-x66wj\") pod \"d2fcfd64-6b34-4f9c-9e8b-3715df35f92d\" (UID: \"d2fcfd64-6b34-4f9c-9e8b-3715df35f92d\") " Dec 07 16:30:14 crc kubenswrapper[4716]: I1207 16:30:14.448571 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2fcfd64-6b34-4f9c-9e8b-3715df35f92d-kube-api-access-x66wj" (OuterVolumeSpecName: "kube-api-access-x66wj") pod "d2fcfd64-6b34-4f9c-9e8b-3715df35f92d" (UID: "d2fcfd64-6b34-4f9c-9e8b-3715df35f92d"). InnerVolumeSpecName "kube-api-access-x66wj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:30:14 crc kubenswrapper[4716]: I1207 16:30:14.474431 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2fcfd64-6b34-4f9c-9e8b-3715df35f92d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d2fcfd64-6b34-4f9c-9e8b-3715df35f92d" (UID: "d2fcfd64-6b34-4f9c-9e8b-3715df35f92d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:30:14 crc kubenswrapper[4716]: I1207 16:30:14.540113 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d2fcfd64-6b34-4f9c-9e8b-3715df35f92d-inventory\") pod \"d2fcfd64-6b34-4f9c-9e8b-3715df35f92d\" (UID: \"d2fcfd64-6b34-4f9c-9e8b-3715df35f92d\") " Dec 07 16:30:14 crc kubenswrapper[4716]: I1207 16:30:14.540791 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x66wj\" (UniqueName: \"kubernetes.io/projected/d2fcfd64-6b34-4f9c-9e8b-3715df35f92d-kube-api-access-x66wj\") on node \"crc\" DevicePath \"\"" Dec 07 16:30:14 crc kubenswrapper[4716]: I1207 16:30:14.540809 4716 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d2fcfd64-6b34-4f9c-9e8b-3715df35f92d-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 07 16:30:14 crc kubenswrapper[4716]: I1207 16:30:14.570351 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2fcfd64-6b34-4f9c-9e8b-3715df35f92d-inventory" (OuterVolumeSpecName: "inventory") pod "d2fcfd64-6b34-4f9c-9e8b-3715df35f92d" (UID: "d2fcfd64-6b34-4f9c-9e8b-3715df35f92d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:30:14 crc kubenswrapper[4716]: I1207 16:30:14.642228 4716 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d2fcfd64-6b34-4f9c-9e8b-3715df35f92d-inventory\") on node \"crc\" DevicePath \"\"" Dec 07 16:30:14 crc kubenswrapper[4716]: I1207 16:30:14.862653 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zsnd4" event={"ID":"d2fcfd64-6b34-4f9c-9e8b-3715df35f92d","Type":"ContainerDied","Data":"a6c206dfe32028e091c1346584cccc356f4efa6da998fe31ede501613a7cdaf9"} Dec 07 16:30:14 crc kubenswrapper[4716]: I1207 16:30:14.863098 4716 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a6c206dfe32028e091c1346584cccc356f4efa6da998fe31ede501613a7cdaf9" Dec 07 16:30:14 crc kubenswrapper[4716]: I1207 16:30:14.862694 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-zsnd4" Dec 07 16:30:14 crc kubenswrapper[4716]: I1207 16:30:14.959550 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kwl9v"] Dec 07 16:30:14 crc kubenswrapper[4716]: E1207 16:30:14.959920 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15e3e08f-67c0-4d1b-8ecb-29a0f4d8ff44" containerName="collect-profiles" Dec 07 16:30:14 crc kubenswrapper[4716]: I1207 16:30:14.959934 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="15e3e08f-67c0-4d1b-8ecb-29a0f4d8ff44" containerName="collect-profiles" Dec 07 16:30:14 crc kubenswrapper[4716]: E1207 16:30:14.959963 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2fcfd64-6b34-4f9c-9e8b-3715df35f92d" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 07 16:30:14 crc kubenswrapper[4716]: I1207 16:30:14.959969 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2fcfd64-6b34-4f9c-9e8b-3715df35f92d" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 07 16:30:14 crc kubenswrapper[4716]: I1207 16:30:14.960175 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2fcfd64-6b34-4f9c-9e8b-3715df35f92d" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 07 16:30:14 crc kubenswrapper[4716]: I1207 16:30:14.960194 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="15e3e08f-67c0-4d1b-8ecb-29a0f4d8ff44" containerName="collect-profiles" Dec 07 16:30:14 crc kubenswrapper[4716]: I1207 16:30:14.960758 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kwl9v" Dec 07 16:30:14 crc kubenswrapper[4716]: I1207 16:30:14.963737 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 07 16:30:14 crc kubenswrapper[4716]: I1207 16:30:14.963768 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 07 16:30:14 crc kubenswrapper[4716]: I1207 16:30:14.963849 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 07 16:30:14 crc kubenswrapper[4716]: I1207 16:30:14.964148 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-h9pt2" Dec 07 16:30:14 crc kubenswrapper[4716]: I1207 16:30:14.973844 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kwl9v"] Dec 07 16:30:15 crc kubenswrapper[4716]: I1207 16:30:15.053243 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7bxn5\" (UniqueName: \"kubernetes.io/projected/fa50585d-9042-4f6e-b926-fdff8344e0ad-kube-api-access-7bxn5\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-kwl9v\" (UID: \"fa50585d-9042-4f6e-b926-fdff8344e0ad\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kwl9v" Dec 07 16:30:15 crc kubenswrapper[4716]: I1207 16:30:15.053292 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fa50585d-9042-4f6e-b926-fdff8344e0ad-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-kwl9v\" (UID: \"fa50585d-9042-4f6e-b926-fdff8344e0ad\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kwl9v" Dec 07 16:30:15 crc kubenswrapper[4716]: I1207 16:30:15.053316 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fa50585d-9042-4f6e-b926-fdff8344e0ad-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-kwl9v\" (UID: \"fa50585d-9042-4f6e-b926-fdff8344e0ad\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kwl9v" Dec 07 16:30:15 crc kubenswrapper[4716]: I1207 16:30:15.155327 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7bxn5\" (UniqueName: \"kubernetes.io/projected/fa50585d-9042-4f6e-b926-fdff8344e0ad-kube-api-access-7bxn5\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-kwl9v\" (UID: \"fa50585d-9042-4f6e-b926-fdff8344e0ad\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kwl9v" Dec 07 16:30:15 crc kubenswrapper[4716]: I1207 16:30:15.155388 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fa50585d-9042-4f6e-b926-fdff8344e0ad-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-kwl9v\" (UID: \"fa50585d-9042-4f6e-b926-fdff8344e0ad\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kwl9v" Dec 07 16:30:15 crc kubenswrapper[4716]: I1207 16:30:15.155421 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fa50585d-9042-4f6e-b926-fdff8344e0ad-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-kwl9v\" (UID: \"fa50585d-9042-4f6e-b926-fdff8344e0ad\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kwl9v" Dec 07 16:30:15 crc kubenswrapper[4716]: I1207 16:30:15.159679 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fa50585d-9042-4f6e-b926-fdff8344e0ad-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-kwl9v\" (UID: \"fa50585d-9042-4f6e-b926-fdff8344e0ad\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kwl9v" Dec 07 16:30:15 crc kubenswrapper[4716]: I1207 16:30:15.159918 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fa50585d-9042-4f6e-b926-fdff8344e0ad-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-kwl9v\" (UID: \"fa50585d-9042-4f6e-b926-fdff8344e0ad\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kwl9v" Dec 07 16:30:15 crc kubenswrapper[4716]: I1207 16:30:15.177641 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7bxn5\" (UniqueName: \"kubernetes.io/projected/fa50585d-9042-4f6e-b926-fdff8344e0ad-kube-api-access-7bxn5\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-kwl9v\" (UID: \"fa50585d-9042-4f6e-b926-fdff8344e0ad\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kwl9v" Dec 07 16:30:15 crc kubenswrapper[4716]: I1207 16:30:15.278474 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kwl9v" Dec 07 16:30:15 crc kubenswrapper[4716]: I1207 16:30:15.676002 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kwl9v"] Dec 07 16:30:15 crc kubenswrapper[4716]: I1207 16:30:15.874528 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kwl9v" event={"ID":"fa50585d-9042-4f6e-b926-fdff8344e0ad","Type":"ContainerStarted","Data":"ff4738a71e3c7c80b55d32487500ea6789909dff431bab94b2c58e8a9fafbcd8"} Dec 07 16:30:16 crc kubenswrapper[4716]: I1207 16:30:16.884499 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kwl9v" event={"ID":"fa50585d-9042-4f6e-b926-fdff8344e0ad","Type":"ContainerStarted","Data":"c830643ede7b29b8001f27675527388948d53502ad794c3db1e69e93960f0eda"} Dec 07 16:30:16 crc kubenswrapper[4716]: I1207 16:30:16.907160 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kwl9v" podStartSLOduration=2.32932452 podStartE2EDuration="2.907141408s" podCreationTimestamp="2025-12-07 16:30:14 +0000 UTC" firstStartedPulling="2025-12-07 16:30:15.698683226 +0000 UTC m=+1678.388968138" lastFinishedPulling="2025-12-07 16:30:16.276500124 +0000 UTC m=+1678.966785026" observedRunningTime="2025-12-07 16:30:16.898572875 +0000 UTC m=+1679.588857777" watchObservedRunningTime="2025-12-07 16:30:16.907141408 +0000 UTC m=+1679.597426320" Dec 07 16:30:21 crc kubenswrapper[4716]: I1207 16:30:21.928542 4716 generic.go:334] "Generic (PLEG): container finished" podID="fa50585d-9042-4f6e-b926-fdff8344e0ad" containerID="c830643ede7b29b8001f27675527388948d53502ad794c3db1e69e93960f0eda" exitCode=0 Dec 07 16:30:21 crc kubenswrapper[4716]: I1207 16:30:21.928620 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kwl9v" event={"ID":"fa50585d-9042-4f6e-b926-fdff8344e0ad","Type":"ContainerDied","Data":"c830643ede7b29b8001f27675527388948d53502ad794c3db1e69e93960f0eda"} Dec 07 16:30:22 crc kubenswrapper[4716]: I1207 16:30:22.761684 4716 patch_prober.go:28] interesting pod/machine-config-daemon-zcxxp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 16:30:22 crc kubenswrapper[4716]: I1207 16:30:22.761772 4716 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 16:30:22 crc kubenswrapper[4716]: I1207 16:30:22.761823 4716 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" Dec 07 16:30:22 crc kubenswrapper[4716]: I1207 16:30:22.762717 4716 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bae1a3e9d80f5b00293cade8b7601799399c2883ae101574459a4c01e523a154"} pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 07 16:30:22 crc kubenswrapper[4716]: I1207 16:30:22.762811 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" containerName="machine-config-daemon" containerID="cri-o://bae1a3e9d80f5b00293cade8b7601799399c2883ae101574459a4c01e523a154" gracePeriod=600 Dec 07 16:30:22 crc kubenswrapper[4716]: E1207 16:30:22.892569 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 16:30:22 crc kubenswrapper[4716]: I1207 16:30:22.944162 4716 generic.go:334] "Generic (PLEG): container finished" podID="c15b59eb-565d-4556-a4ce-75afdf159dc8" containerID="bae1a3e9d80f5b00293cade8b7601799399c2883ae101574459a4c01e523a154" exitCode=0 Dec 07 16:30:22 crc kubenswrapper[4716]: I1207 16:30:22.944241 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" event={"ID":"c15b59eb-565d-4556-a4ce-75afdf159dc8","Type":"ContainerDied","Data":"bae1a3e9d80f5b00293cade8b7601799399c2883ae101574459a4c01e523a154"} Dec 07 16:30:22 crc kubenswrapper[4716]: I1207 16:30:22.944304 4716 scope.go:117] "RemoveContainer" containerID="fea4ce7ae8e17d1efab423015c4c8670e9be41f20a55cff5db1bae1736f619e0" Dec 07 16:30:22 crc kubenswrapper[4716]: I1207 16:30:22.945365 4716 scope.go:117] "RemoveContainer" containerID="bae1a3e9d80f5b00293cade8b7601799399c2883ae101574459a4c01e523a154" Dec 07 16:30:22 crc kubenswrapper[4716]: E1207 16:30:22.945868 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 16:30:23 crc kubenswrapper[4716]: I1207 16:30:23.438032 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kwl9v" Dec 07 16:30:23 crc kubenswrapper[4716]: I1207 16:30:23.527780 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fa50585d-9042-4f6e-b926-fdff8344e0ad-inventory\") pod \"fa50585d-9042-4f6e-b926-fdff8344e0ad\" (UID: \"fa50585d-9042-4f6e-b926-fdff8344e0ad\") " Dec 07 16:30:23 crc kubenswrapper[4716]: I1207 16:30:23.527903 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fa50585d-9042-4f6e-b926-fdff8344e0ad-ssh-key\") pod \"fa50585d-9042-4f6e-b926-fdff8344e0ad\" (UID: \"fa50585d-9042-4f6e-b926-fdff8344e0ad\") " Dec 07 16:30:23 crc kubenswrapper[4716]: I1207 16:30:23.527964 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7bxn5\" (UniqueName: \"kubernetes.io/projected/fa50585d-9042-4f6e-b926-fdff8344e0ad-kube-api-access-7bxn5\") pod \"fa50585d-9042-4f6e-b926-fdff8344e0ad\" (UID: \"fa50585d-9042-4f6e-b926-fdff8344e0ad\") " Dec 07 16:30:23 crc kubenswrapper[4716]: I1207 16:30:23.533225 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa50585d-9042-4f6e-b926-fdff8344e0ad-kube-api-access-7bxn5" (OuterVolumeSpecName: "kube-api-access-7bxn5") pod "fa50585d-9042-4f6e-b926-fdff8344e0ad" (UID: "fa50585d-9042-4f6e-b926-fdff8344e0ad"). InnerVolumeSpecName "kube-api-access-7bxn5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:30:23 crc kubenswrapper[4716]: I1207 16:30:23.557955 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa50585d-9042-4f6e-b926-fdff8344e0ad-inventory" (OuterVolumeSpecName: "inventory") pod "fa50585d-9042-4f6e-b926-fdff8344e0ad" (UID: "fa50585d-9042-4f6e-b926-fdff8344e0ad"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:30:23 crc kubenswrapper[4716]: I1207 16:30:23.560710 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa50585d-9042-4f6e-b926-fdff8344e0ad-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "fa50585d-9042-4f6e-b926-fdff8344e0ad" (UID: "fa50585d-9042-4f6e-b926-fdff8344e0ad"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:30:23 crc kubenswrapper[4716]: I1207 16:30:23.630525 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7bxn5\" (UniqueName: \"kubernetes.io/projected/fa50585d-9042-4f6e-b926-fdff8344e0ad-kube-api-access-7bxn5\") on node \"crc\" DevicePath \"\"" Dec 07 16:30:23 crc kubenswrapper[4716]: I1207 16:30:23.630565 4716 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fa50585d-9042-4f6e-b926-fdff8344e0ad-inventory\") on node \"crc\" DevicePath \"\"" Dec 07 16:30:23 crc kubenswrapper[4716]: I1207 16:30:23.630575 4716 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fa50585d-9042-4f6e-b926-fdff8344e0ad-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 07 16:30:23 crc kubenswrapper[4716]: I1207 16:30:23.954575 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kwl9v" event={"ID":"fa50585d-9042-4f6e-b926-fdff8344e0ad","Type":"ContainerDied","Data":"ff4738a71e3c7c80b55d32487500ea6789909dff431bab94b2c58e8a9fafbcd8"} Dec 07 16:30:23 crc kubenswrapper[4716]: I1207 16:30:23.954609 4716 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ff4738a71e3c7c80b55d32487500ea6789909dff431bab94b2c58e8a9fafbcd8" Dec 07 16:30:23 crc kubenswrapper[4716]: I1207 16:30:23.954624 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kwl9v" Dec 07 16:30:24 crc kubenswrapper[4716]: I1207 16:30:24.038674 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-bd27b"] Dec 07 16:30:24 crc kubenswrapper[4716]: E1207 16:30:24.039675 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa50585d-9042-4f6e-b926-fdff8344e0ad" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 07 16:30:24 crc kubenswrapper[4716]: I1207 16:30:24.039705 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa50585d-9042-4f6e-b926-fdff8344e0ad" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 07 16:30:24 crc kubenswrapper[4716]: I1207 16:30:24.040345 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa50585d-9042-4f6e-b926-fdff8344e0ad" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 07 16:30:24 crc kubenswrapper[4716]: I1207 16:30:24.041957 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-bd27b" Dec 07 16:30:24 crc kubenswrapper[4716]: I1207 16:30:24.048661 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 07 16:30:24 crc kubenswrapper[4716]: I1207 16:30:24.048804 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 07 16:30:24 crc kubenswrapper[4716]: I1207 16:30:24.049245 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-h9pt2" Dec 07 16:30:24 crc kubenswrapper[4716]: I1207 16:30:24.049371 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 07 16:30:24 crc kubenswrapper[4716]: I1207 16:30:24.064882 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-bd27b"] Dec 07 16:30:24 crc kubenswrapper[4716]: I1207 16:30:24.139495 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9e126137-6235-4a75-8d46-05bb3cf60acc-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-bd27b\" (UID: \"9e126137-6235-4a75-8d46-05bb3cf60acc\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-bd27b" Dec 07 16:30:24 crc kubenswrapper[4716]: I1207 16:30:24.139632 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9e126137-6235-4a75-8d46-05bb3cf60acc-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-bd27b\" (UID: \"9e126137-6235-4a75-8d46-05bb3cf60acc\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-bd27b" Dec 07 16:30:24 crc kubenswrapper[4716]: I1207 16:30:24.139725 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g269x\" (UniqueName: \"kubernetes.io/projected/9e126137-6235-4a75-8d46-05bb3cf60acc-kube-api-access-g269x\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-bd27b\" (UID: \"9e126137-6235-4a75-8d46-05bb3cf60acc\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-bd27b" Dec 07 16:30:24 crc kubenswrapper[4716]: I1207 16:30:24.241575 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g269x\" (UniqueName: \"kubernetes.io/projected/9e126137-6235-4a75-8d46-05bb3cf60acc-kube-api-access-g269x\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-bd27b\" (UID: \"9e126137-6235-4a75-8d46-05bb3cf60acc\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-bd27b" Dec 07 16:30:24 crc kubenswrapper[4716]: I1207 16:30:24.241723 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9e126137-6235-4a75-8d46-05bb3cf60acc-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-bd27b\" (UID: \"9e126137-6235-4a75-8d46-05bb3cf60acc\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-bd27b" Dec 07 16:30:24 crc kubenswrapper[4716]: I1207 16:30:24.241774 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9e126137-6235-4a75-8d46-05bb3cf60acc-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-bd27b\" (UID: \"9e126137-6235-4a75-8d46-05bb3cf60acc\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-bd27b" Dec 07 16:30:24 crc kubenswrapper[4716]: I1207 16:30:24.245634 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9e126137-6235-4a75-8d46-05bb3cf60acc-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-bd27b\" (UID: \"9e126137-6235-4a75-8d46-05bb3cf60acc\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-bd27b" Dec 07 16:30:24 crc kubenswrapper[4716]: I1207 16:30:24.254432 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9e126137-6235-4a75-8d46-05bb3cf60acc-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-bd27b\" (UID: \"9e126137-6235-4a75-8d46-05bb3cf60acc\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-bd27b" Dec 07 16:30:24 crc kubenswrapper[4716]: I1207 16:30:24.256735 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g269x\" (UniqueName: \"kubernetes.io/projected/9e126137-6235-4a75-8d46-05bb3cf60acc-kube-api-access-g269x\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-bd27b\" (UID: \"9e126137-6235-4a75-8d46-05bb3cf60acc\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-bd27b" Dec 07 16:30:24 crc kubenswrapper[4716]: I1207 16:30:24.372441 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-bd27b" Dec 07 16:30:24 crc kubenswrapper[4716]: I1207 16:30:24.701204 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-bd27b"] Dec 07 16:30:24 crc kubenswrapper[4716]: I1207 16:30:24.966395 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-bd27b" event={"ID":"9e126137-6235-4a75-8d46-05bb3cf60acc","Type":"ContainerStarted","Data":"8d95b31af049dcc7268ddd914e376c5c6c4aff7fcc06732dbfa4704d5af53159"} Dec 07 16:30:25 crc kubenswrapper[4716]: I1207 16:30:25.975945 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-bd27b" event={"ID":"9e126137-6235-4a75-8d46-05bb3cf60acc","Type":"ContainerStarted","Data":"ffa9e09b3eeefdfdb01be62732efd32385cbe6779cc0cef160512016c2b241d4"} Dec 07 16:30:26 crc kubenswrapper[4716]: I1207 16:30:26.002478 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-bd27b" podStartSLOduration=1.568437296 podStartE2EDuration="2.002459925s" podCreationTimestamp="2025-12-07 16:30:24 +0000 UTC" firstStartedPulling="2025-12-07 16:30:24.701218702 +0000 UTC m=+1687.391503614" lastFinishedPulling="2025-12-07 16:30:25.135241331 +0000 UTC m=+1687.825526243" observedRunningTime="2025-12-07 16:30:25.99339016 +0000 UTC m=+1688.683675092" watchObservedRunningTime="2025-12-07 16:30:26.002459925 +0000 UTC m=+1688.692744847" Dec 07 16:30:33 crc kubenswrapper[4716]: I1207 16:30:33.052815 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-xbk7g"] Dec 07 16:30:33 crc kubenswrapper[4716]: I1207 16:30:33.060947 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-xbk7g"] Dec 07 16:30:33 crc kubenswrapper[4716]: I1207 16:30:33.601887 4716 scope.go:117] "RemoveContainer" containerID="ad8e4e125c3aa6c78a78cf693c30cd14ae8bb48186cfb3c839fde875ef8d4b8f" Dec 07 16:30:33 crc kubenswrapper[4716]: I1207 16:30:33.662374 4716 scope.go:117] "RemoveContainer" containerID="ee99ded3785807ca8c912daa4b281fb89f2e8b61332ef14490a3f3d6d8909f5f" Dec 07 16:30:33 crc kubenswrapper[4716]: I1207 16:30:33.668733 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ead59b4-dcff-45b8-8b1b-d5e449278acb" path="/var/lib/kubelet/pods/7ead59b4-dcff-45b8-8b1b-d5e449278acb/volumes" Dec 07 16:30:33 crc kubenswrapper[4716]: I1207 16:30:33.695060 4716 scope.go:117] "RemoveContainer" containerID="30ab4927e99fb057a7b46218e65f98171201ab712c6f11193b31bc9789eb773a" Dec 07 16:30:33 crc kubenswrapper[4716]: I1207 16:30:33.748984 4716 scope.go:117] "RemoveContainer" containerID="a93e9710e35e18568b8b3e95c22a35043085b7c042fb81b1aa8b550b36ab9a3f" Dec 07 16:30:33 crc kubenswrapper[4716]: I1207 16:30:33.778796 4716 scope.go:117] "RemoveContainer" containerID="a4bb3f5fae952b8c1002cb654179ae8ea80bb4dbe8649a05733636adb9e748f2" Dec 07 16:30:34 crc kubenswrapper[4716]: I1207 16:30:34.037824 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-33d4-account-create-update-2tg2q"] Dec 07 16:30:34 crc kubenswrapper[4716]: I1207 16:30:34.044227 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-5b89-account-create-update-w7tgb"] Dec 07 16:30:34 crc kubenswrapper[4716]: I1207 16:30:34.056561 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-c495-account-create-update-cpqvg"] Dec 07 16:30:34 crc kubenswrapper[4716]: I1207 16:30:34.068691 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-pz462"] Dec 07 16:30:34 crc kubenswrapper[4716]: I1207 16:30:34.077928 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-d9hjx"] Dec 07 16:30:34 crc kubenswrapper[4716]: I1207 16:30:34.085402 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-c495-account-create-update-cpqvg"] Dec 07 16:30:34 crc kubenswrapper[4716]: I1207 16:30:34.092148 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-d9hjx"] Dec 07 16:30:34 crc kubenswrapper[4716]: I1207 16:30:34.100048 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-5b89-account-create-update-w7tgb"] Dec 07 16:30:34 crc kubenswrapper[4716]: I1207 16:30:34.106990 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-33d4-account-create-update-2tg2q"] Dec 07 16:30:34 crc kubenswrapper[4716]: I1207 16:30:34.114045 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-pz462"] Dec 07 16:30:34 crc kubenswrapper[4716]: I1207 16:30:34.658311 4716 scope.go:117] "RemoveContainer" containerID="bae1a3e9d80f5b00293cade8b7601799399c2883ae101574459a4c01e523a154" Dec 07 16:30:34 crc kubenswrapper[4716]: E1207 16:30:34.658748 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 16:30:35 crc kubenswrapper[4716]: I1207 16:30:35.670748 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3052dae6-ee76-44ac-bc5c-6dcb6d9995cf" path="/var/lib/kubelet/pods/3052dae6-ee76-44ac-bc5c-6dcb6d9995cf/volumes" Dec 07 16:30:35 crc kubenswrapper[4716]: I1207 16:30:35.671589 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4647e69c-31d0-495e-a294-f9970aae2c32" path="/var/lib/kubelet/pods/4647e69c-31d0-495e-a294-f9970aae2c32/volumes" Dec 07 16:30:35 crc kubenswrapper[4716]: I1207 16:30:35.672123 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73795e56-2afe-427c-a122-7d604c6ed134" path="/var/lib/kubelet/pods/73795e56-2afe-427c-a122-7d604c6ed134/volumes" Dec 07 16:30:35 crc kubenswrapper[4716]: I1207 16:30:35.672682 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a14d4b2b-4e79-4258-9d9f-46198eca8a41" path="/var/lib/kubelet/pods/a14d4b2b-4e79-4258-9d9f-46198eca8a41/volumes" Dec 07 16:30:35 crc kubenswrapper[4716]: I1207 16:30:35.673755 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a8a5773f-5b23-400b-8911-1d2d40b9bfd0" path="/var/lib/kubelet/pods/a8a5773f-5b23-400b-8911-1d2d40b9bfd0/volumes" Dec 07 16:30:48 crc kubenswrapper[4716]: I1207 16:30:48.657647 4716 scope.go:117] "RemoveContainer" containerID="bae1a3e9d80f5b00293cade8b7601799399c2883ae101574459a4c01e523a154" Dec 07 16:30:48 crc kubenswrapper[4716]: E1207 16:30:48.658417 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 16:31:03 crc kubenswrapper[4716]: I1207 16:31:03.052780 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-6kcdj"] Dec 07 16:31:03 crc kubenswrapper[4716]: I1207 16:31:03.063153 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-6kcdj"] Dec 07 16:31:03 crc kubenswrapper[4716]: I1207 16:31:03.658458 4716 scope.go:117] "RemoveContainer" containerID="bae1a3e9d80f5b00293cade8b7601799399c2883ae101574459a4c01e523a154" Dec 07 16:31:03 crc kubenswrapper[4716]: E1207 16:31:03.658983 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 16:31:03 crc kubenswrapper[4716]: I1207 16:31:03.679367 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a587b1d-a86e-4ee2-907e-6b014c61b09b" path="/var/lib/kubelet/pods/7a587b1d-a86e-4ee2-907e-6b014c61b09b/volumes" Dec 07 16:31:04 crc kubenswrapper[4716]: I1207 16:31:04.348496 4716 generic.go:334] "Generic (PLEG): container finished" podID="9e126137-6235-4a75-8d46-05bb3cf60acc" containerID="ffa9e09b3eeefdfdb01be62732efd32385cbe6779cc0cef160512016c2b241d4" exitCode=0 Dec 07 16:31:04 crc kubenswrapper[4716]: I1207 16:31:04.348553 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-bd27b" event={"ID":"9e126137-6235-4a75-8d46-05bb3cf60acc","Type":"ContainerDied","Data":"ffa9e09b3eeefdfdb01be62732efd32385cbe6779cc0cef160512016c2b241d4"} Dec 07 16:31:05 crc kubenswrapper[4716]: I1207 16:31:05.882489 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-bd27b" Dec 07 16:31:06 crc kubenswrapper[4716]: I1207 16:31:06.082772 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g269x\" (UniqueName: \"kubernetes.io/projected/9e126137-6235-4a75-8d46-05bb3cf60acc-kube-api-access-g269x\") pod \"9e126137-6235-4a75-8d46-05bb3cf60acc\" (UID: \"9e126137-6235-4a75-8d46-05bb3cf60acc\") " Dec 07 16:31:06 crc kubenswrapper[4716]: I1207 16:31:06.083108 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9e126137-6235-4a75-8d46-05bb3cf60acc-inventory\") pod \"9e126137-6235-4a75-8d46-05bb3cf60acc\" (UID: \"9e126137-6235-4a75-8d46-05bb3cf60acc\") " Dec 07 16:31:06 crc kubenswrapper[4716]: I1207 16:31:06.083184 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9e126137-6235-4a75-8d46-05bb3cf60acc-ssh-key\") pod \"9e126137-6235-4a75-8d46-05bb3cf60acc\" (UID: \"9e126137-6235-4a75-8d46-05bb3cf60acc\") " Dec 07 16:31:06 crc kubenswrapper[4716]: I1207 16:31:06.091213 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e126137-6235-4a75-8d46-05bb3cf60acc-kube-api-access-g269x" (OuterVolumeSpecName: "kube-api-access-g269x") pod "9e126137-6235-4a75-8d46-05bb3cf60acc" (UID: "9e126137-6235-4a75-8d46-05bb3cf60acc"). InnerVolumeSpecName "kube-api-access-g269x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:31:06 crc kubenswrapper[4716]: I1207 16:31:06.113165 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e126137-6235-4a75-8d46-05bb3cf60acc-inventory" (OuterVolumeSpecName: "inventory") pod "9e126137-6235-4a75-8d46-05bb3cf60acc" (UID: "9e126137-6235-4a75-8d46-05bb3cf60acc"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:31:06 crc kubenswrapper[4716]: I1207 16:31:06.130089 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e126137-6235-4a75-8d46-05bb3cf60acc-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9e126137-6235-4a75-8d46-05bb3cf60acc" (UID: "9e126137-6235-4a75-8d46-05bb3cf60acc"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:31:06 crc kubenswrapper[4716]: I1207 16:31:06.185547 4716 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9e126137-6235-4a75-8d46-05bb3cf60acc-inventory\") on node \"crc\" DevicePath \"\"" Dec 07 16:31:06 crc kubenswrapper[4716]: I1207 16:31:06.185589 4716 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9e126137-6235-4a75-8d46-05bb3cf60acc-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 07 16:31:06 crc kubenswrapper[4716]: I1207 16:31:06.185598 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g269x\" (UniqueName: \"kubernetes.io/projected/9e126137-6235-4a75-8d46-05bb3cf60acc-kube-api-access-g269x\") on node \"crc\" DevicePath \"\"" Dec 07 16:31:06 crc kubenswrapper[4716]: I1207 16:31:06.371396 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-bd27b" event={"ID":"9e126137-6235-4a75-8d46-05bb3cf60acc","Type":"ContainerDied","Data":"8d95b31af049dcc7268ddd914e376c5c6c4aff7fcc06732dbfa4704d5af53159"} Dec 07 16:31:06 crc kubenswrapper[4716]: I1207 16:31:06.371434 4716 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8d95b31af049dcc7268ddd914e376c5c6c4aff7fcc06732dbfa4704d5af53159" Dec 07 16:31:06 crc kubenswrapper[4716]: I1207 16:31:06.371511 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-bd27b" Dec 07 16:31:06 crc kubenswrapper[4716]: I1207 16:31:06.462056 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kcj2p"] Dec 07 16:31:06 crc kubenswrapper[4716]: E1207 16:31:06.462667 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e126137-6235-4a75-8d46-05bb3cf60acc" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 07 16:31:06 crc kubenswrapper[4716]: I1207 16:31:06.462695 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e126137-6235-4a75-8d46-05bb3cf60acc" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 07 16:31:06 crc kubenswrapper[4716]: I1207 16:31:06.462986 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e126137-6235-4a75-8d46-05bb3cf60acc" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 07 16:31:06 crc kubenswrapper[4716]: I1207 16:31:06.463981 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kcj2p" Dec 07 16:31:06 crc kubenswrapper[4716]: I1207 16:31:06.466459 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 07 16:31:06 crc kubenswrapper[4716]: I1207 16:31:06.466525 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 07 16:31:06 crc kubenswrapper[4716]: I1207 16:31:06.466863 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-h9pt2" Dec 07 16:31:06 crc kubenswrapper[4716]: I1207 16:31:06.467044 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 07 16:31:06 crc kubenswrapper[4716]: I1207 16:31:06.486018 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kcj2p"] Dec 07 16:31:06 crc kubenswrapper[4716]: I1207 16:31:06.593504 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9d00d0ca-8b9a-455b-a7a5-f3d17a6e363a-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-kcj2p\" (UID: \"9d00d0ca-8b9a-455b-a7a5-f3d17a6e363a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kcj2p" Dec 07 16:31:06 crc kubenswrapper[4716]: I1207 16:31:06.593575 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqv4x\" (UniqueName: \"kubernetes.io/projected/9d00d0ca-8b9a-455b-a7a5-f3d17a6e363a-kube-api-access-nqv4x\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-kcj2p\" (UID: \"9d00d0ca-8b9a-455b-a7a5-f3d17a6e363a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kcj2p" Dec 07 16:31:06 crc kubenswrapper[4716]: I1207 16:31:06.593757 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9d00d0ca-8b9a-455b-a7a5-f3d17a6e363a-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-kcj2p\" (UID: \"9d00d0ca-8b9a-455b-a7a5-f3d17a6e363a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kcj2p" Dec 07 16:31:06 crc kubenswrapper[4716]: I1207 16:31:06.695585 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9d00d0ca-8b9a-455b-a7a5-f3d17a6e363a-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-kcj2p\" (UID: \"9d00d0ca-8b9a-455b-a7a5-f3d17a6e363a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kcj2p" Dec 07 16:31:06 crc kubenswrapper[4716]: I1207 16:31:06.695700 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqv4x\" (UniqueName: \"kubernetes.io/projected/9d00d0ca-8b9a-455b-a7a5-f3d17a6e363a-kube-api-access-nqv4x\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-kcj2p\" (UID: \"9d00d0ca-8b9a-455b-a7a5-f3d17a6e363a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kcj2p" Dec 07 16:31:06 crc kubenswrapper[4716]: I1207 16:31:06.695792 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9d00d0ca-8b9a-455b-a7a5-f3d17a6e363a-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-kcj2p\" (UID: \"9d00d0ca-8b9a-455b-a7a5-f3d17a6e363a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kcj2p" Dec 07 16:31:06 crc kubenswrapper[4716]: I1207 16:31:06.699331 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9d00d0ca-8b9a-455b-a7a5-f3d17a6e363a-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-kcj2p\" (UID: \"9d00d0ca-8b9a-455b-a7a5-f3d17a6e363a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kcj2p" Dec 07 16:31:06 crc kubenswrapper[4716]: I1207 16:31:06.699508 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9d00d0ca-8b9a-455b-a7a5-f3d17a6e363a-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-kcj2p\" (UID: \"9d00d0ca-8b9a-455b-a7a5-f3d17a6e363a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kcj2p" Dec 07 16:31:06 crc kubenswrapper[4716]: I1207 16:31:06.714231 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqv4x\" (UniqueName: \"kubernetes.io/projected/9d00d0ca-8b9a-455b-a7a5-f3d17a6e363a-kube-api-access-nqv4x\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-kcj2p\" (UID: \"9d00d0ca-8b9a-455b-a7a5-f3d17a6e363a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kcj2p" Dec 07 16:31:06 crc kubenswrapper[4716]: I1207 16:31:06.784510 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kcj2p" Dec 07 16:31:07 crc kubenswrapper[4716]: I1207 16:31:07.319094 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kcj2p"] Dec 07 16:31:07 crc kubenswrapper[4716]: I1207 16:31:07.380775 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kcj2p" event={"ID":"9d00d0ca-8b9a-455b-a7a5-f3d17a6e363a","Type":"ContainerStarted","Data":"502ba4292d4a7c6de60f640530fb397a74e95f8f826b9370d75c3bd1f1885cbd"} Dec 07 16:31:08 crc kubenswrapper[4716]: I1207 16:31:08.389638 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kcj2p" event={"ID":"9d00d0ca-8b9a-455b-a7a5-f3d17a6e363a","Type":"ContainerStarted","Data":"4ba7e2106827aacfd72bbe1e788d025d6c49fc73ec9b921997e5c962206331d5"} Dec 07 16:31:08 crc kubenswrapper[4716]: I1207 16:31:08.409324 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kcj2p" podStartSLOduration=1.977617009 podStartE2EDuration="2.409304865s" podCreationTimestamp="2025-12-07 16:31:06 +0000 UTC" firstStartedPulling="2025-12-07 16:31:07.340148812 +0000 UTC m=+1730.030433744" lastFinishedPulling="2025-12-07 16:31:07.771836688 +0000 UTC m=+1730.462121600" observedRunningTime="2025-12-07 16:31:08.407831346 +0000 UTC m=+1731.098116258" watchObservedRunningTime="2025-12-07 16:31:08.409304865 +0000 UTC m=+1731.099589767" Dec 07 16:31:14 crc kubenswrapper[4716]: I1207 16:31:14.659205 4716 scope.go:117] "RemoveContainer" containerID="bae1a3e9d80f5b00293cade8b7601799399c2883ae101574459a4c01e523a154" Dec 07 16:31:14 crc kubenswrapper[4716]: E1207 16:31:14.660854 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 16:31:24 crc kubenswrapper[4716]: I1207 16:31:24.061998 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-q9fx7"] Dec 07 16:31:24 crc kubenswrapper[4716]: I1207 16:31:24.073691 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-q9fx7"] Dec 07 16:31:25 crc kubenswrapper[4716]: I1207 16:31:25.033798 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-x9skx"] Dec 07 16:31:25 crc kubenswrapper[4716]: I1207 16:31:25.042239 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-x9skx"] Dec 07 16:31:25 crc kubenswrapper[4716]: I1207 16:31:25.668928 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="367f54c1-6d21-41fc-8c08-d6a7e9e637cf" path="/var/lib/kubelet/pods/367f54c1-6d21-41fc-8c08-d6a7e9e637cf/volumes" Dec 07 16:31:25 crc kubenswrapper[4716]: I1207 16:31:25.669780 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a950343-f5b9-4a42-857b-7de8ef106453" path="/var/lib/kubelet/pods/9a950343-f5b9-4a42-857b-7de8ef106453/volumes" Dec 07 16:31:27 crc kubenswrapper[4716]: I1207 16:31:27.664316 4716 scope.go:117] "RemoveContainer" containerID="bae1a3e9d80f5b00293cade8b7601799399c2883ae101574459a4c01e523a154" Dec 07 16:31:27 crc kubenswrapper[4716]: E1207 16:31:27.664765 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 16:31:33 crc kubenswrapper[4716]: I1207 16:31:33.901761 4716 scope.go:117] "RemoveContainer" containerID="3f6d2d6aa158ff9d5af36e3ebe2ea76619cd872f551bf4eba62dd2343c17d952" Dec 07 16:31:33 crc kubenswrapper[4716]: I1207 16:31:33.965991 4716 scope.go:117] "RemoveContainer" containerID="f918233e5ff893608dff230031ceae9e1dfe9739e05fd2ab94b483e2a33df54f" Dec 07 16:31:34 crc kubenswrapper[4716]: I1207 16:31:34.024914 4716 scope.go:117] "RemoveContainer" containerID="290aaea26096362bace3d7ab97853c9057fe0afaa3cf08806cd429ea3ae13f7a" Dec 07 16:31:34 crc kubenswrapper[4716]: I1207 16:31:34.051851 4716 scope.go:117] "RemoveContainer" containerID="6368cc2ab6c45969b8edd9d2e09cdeb711278ef76446cd516bd195c0d856f570" Dec 07 16:31:34 crc kubenswrapper[4716]: I1207 16:31:34.130306 4716 scope.go:117] "RemoveContainer" containerID="25a2fe086275ad4a71a835f4d1d54853927f11009001e16c4bde04a32db4975d" Dec 07 16:31:34 crc kubenswrapper[4716]: I1207 16:31:34.153217 4716 scope.go:117] "RemoveContainer" containerID="2df3471b336e2c8ff319c52b304c9fbe472180f8eb3d143bcee10f51a7b68902" Dec 07 16:31:34 crc kubenswrapper[4716]: I1207 16:31:34.200792 4716 scope.go:117] "RemoveContainer" containerID="7ad63c67cc7344b576e5c104851f19bd0e23971ef38e833eba4570bbe4d63af5" Dec 07 16:31:34 crc kubenswrapper[4716]: I1207 16:31:34.225311 4716 scope.go:117] "RemoveContainer" containerID="0e6d7d18623b527169bf2eaaab3b349ad13833aa0638e2e22b0ab5fcd0b0f715" Dec 07 16:31:40 crc kubenswrapper[4716]: I1207 16:31:40.657985 4716 scope.go:117] "RemoveContainer" containerID="bae1a3e9d80f5b00293cade8b7601799399c2883ae101574459a4c01e523a154" Dec 07 16:31:40 crc kubenswrapper[4716]: E1207 16:31:40.658768 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 16:31:53 crc kubenswrapper[4716]: I1207 16:31:53.658262 4716 scope.go:117] "RemoveContainer" containerID="bae1a3e9d80f5b00293cade8b7601799399c2883ae101574459a4c01e523a154" Dec 07 16:31:53 crc kubenswrapper[4716]: E1207 16:31:53.659036 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 16:32:05 crc kubenswrapper[4716]: I1207 16:32:05.658251 4716 scope.go:117] "RemoveContainer" containerID="bae1a3e9d80f5b00293cade8b7601799399c2883ae101574459a4c01e523a154" Dec 07 16:32:05 crc kubenswrapper[4716]: E1207 16:32:05.659379 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 16:32:10 crc kubenswrapper[4716]: I1207 16:32:10.057813 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-cbchp"] Dec 07 16:32:10 crc kubenswrapper[4716]: I1207 16:32:10.064610 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-cbchp"] Dec 07 16:32:11 crc kubenswrapper[4716]: I1207 16:32:11.678320 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aef53883-32c3-44d7-9d3e-dd15928778a9" path="/var/lib/kubelet/pods/aef53883-32c3-44d7-9d3e-dd15928778a9/volumes" Dec 07 16:32:12 crc kubenswrapper[4716]: I1207 16:32:12.018364 4716 generic.go:334] "Generic (PLEG): container finished" podID="9d00d0ca-8b9a-455b-a7a5-f3d17a6e363a" containerID="4ba7e2106827aacfd72bbe1e788d025d6c49fc73ec9b921997e5c962206331d5" exitCode=0 Dec 07 16:32:12 crc kubenswrapper[4716]: I1207 16:32:12.018428 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kcj2p" event={"ID":"9d00d0ca-8b9a-455b-a7a5-f3d17a6e363a","Type":"ContainerDied","Data":"4ba7e2106827aacfd72bbe1e788d025d6c49fc73ec9b921997e5c962206331d5"} Dec 07 16:32:13 crc kubenswrapper[4716]: I1207 16:32:13.450581 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kcj2p" Dec 07 16:32:13 crc kubenswrapper[4716]: I1207 16:32:13.507816 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9d00d0ca-8b9a-455b-a7a5-f3d17a6e363a-ssh-key\") pod \"9d00d0ca-8b9a-455b-a7a5-f3d17a6e363a\" (UID: \"9d00d0ca-8b9a-455b-a7a5-f3d17a6e363a\") " Dec 07 16:32:13 crc kubenswrapper[4716]: I1207 16:32:13.508131 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nqv4x\" (UniqueName: \"kubernetes.io/projected/9d00d0ca-8b9a-455b-a7a5-f3d17a6e363a-kube-api-access-nqv4x\") pod \"9d00d0ca-8b9a-455b-a7a5-f3d17a6e363a\" (UID: \"9d00d0ca-8b9a-455b-a7a5-f3d17a6e363a\") " Dec 07 16:32:13 crc kubenswrapper[4716]: I1207 16:32:13.508164 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9d00d0ca-8b9a-455b-a7a5-f3d17a6e363a-inventory\") pod \"9d00d0ca-8b9a-455b-a7a5-f3d17a6e363a\" (UID: \"9d00d0ca-8b9a-455b-a7a5-f3d17a6e363a\") " Dec 07 16:32:13 crc kubenswrapper[4716]: I1207 16:32:13.527931 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d00d0ca-8b9a-455b-a7a5-f3d17a6e363a-kube-api-access-nqv4x" (OuterVolumeSpecName: "kube-api-access-nqv4x") pod "9d00d0ca-8b9a-455b-a7a5-f3d17a6e363a" (UID: "9d00d0ca-8b9a-455b-a7a5-f3d17a6e363a"). InnerVolumeSpecName "kube-api-access-nqv4x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:32:13 crc kubenswrapper[4716]: I1207 16:32:13.535939 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d00d0ca-8b9a-455b-a7a5-f3d17a6e363a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9d00d0ca-8b9a-455b-a7a5-f3d17a6e363a" (UID: "9d00d0ca-8b9a-455b-a7a5-f3d17a6e363a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:32:13 crc kubenswrapper[4716]: I1207 16:32:13.570877 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d00d0ca-8b9a-455b-a7a5-f3d17a6e363a-inventory" (OuterVolumeSpecName: "inventory") pod "9d00d0ca-8b9a-455b-a7a5-f3d17a6e363a" (UID: "9d00d0ca-8b9a-455b-a7a5-f3d17a6e363a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:32:13 crc kubenswrapper[4716]: I1207 16:32:13.609273 4716 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9d00d0ca-8b9a-455b-a7a5-f3d17a6e363a-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 07 16:32:13 crc kubenswrapper[4716]: I1207 16:32:13.609307 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nqv4x\" (UniqueName: \"kubernetes.io/projected/9d00d0ca-8b9a-455b-a7a5-f3d17a6e363a-kube-api-access-nqv4x\") on node \"crc\" DevicePath \"\"" Dec 07 16:32:13 crc kubenswrapper[4716]: I1207 16:32:13.609322 4716 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9d00d0ca-8b9a-455b-a7a5-f3d17a6e363a-inventory\") on node \"crc\" DevicePath \"\"" Dec 07 16:32:14 crc kubenswrapper[4716]: I1207 16:32:14.042594 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kcj2p" event={"ID":"9d00d0ca-8b9a-455b-a7a5-f3d17a6e363a","Type":"ContainerDied","Data":"502ba4292d4a7c6de60f640530fb397a74e95f8f826b9370d75c3bd1f1885cbd"} Dec 07 16:32:14 crc kubenswrapper[4716]: I1207 16:32:14.042633 4716 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="502ba4292d4a7c6de60f640530fb397a74e95f8f826b9370d75c3bd1f1885cbd" Dec 07 16:32:14 crc kubenswrapper[4716]: I1207 16:32:14.042710 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kcj2p" Dec 07 16:32:14 crc kubenswrapper[4716]: I1207 16:32:14.153255 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-vpm5k"] Dec 07 16:32:14 crc kubenswrapper[4716]: E1207 16:32:14.153722 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d00d0ca-8b9a-455b-a7a5-f3d17a6e363a" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 07 16:32:14 crc kubenswrapper[4716]: I1207 16:32:14.153747 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d00d0ca-8b9a-455b-a7a5-f3d17a6e363a" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 07 16:32:14 crc kubenswrapper[4716]: I1207 16:32:14.154033 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d00d0ca-8b9a-455b-a7a5-f3d17a6e363a" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 07 16:32:14 crc kubenswrapper[4716]: I1207 16:32:14.154780 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-vpm5k" Dec 07 16:32:14 crc kubenswrapper[4716]: I1207 16:32:14.156739 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 07 16:32:14 crc kubenswrapper[4716]: I1207 16:32:14.157605 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 07 16:32:14 crc kubenswrapper[4716]: I1207 16:32:14.157849 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 07 16:32:14 crc kubenswrapper[4716]: I1207 16:32:14.157898 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-h9pt2" Dec 07 16:32:14 crc kubenswrapper[4716]: I1207 16:32:14.167420 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-vpm5k"] Dec 07 16:32:14 crc kubenswrapper[4716]: I1207 16:32:14.222753 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/92420ce0-b95e-4857-b6f3-f6d4519c3d86-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-vpm5k\" (UID: \"92420ce0-b95e-4857-b6f3-f6d4519c3d86\") " pod="openstack/ssh-known-hosts-edpm-deployment-vpm5k" Dec 07 16:32:14 crc kubenswrapper[4716]: I1207 16:32:14.222842 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/92420ce0-b95e-4857-b6f3-f6d4519c3d86-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-vpm5k\" (UID: \"92420ce0-b95e-4857-b6f3-f6d4519c3d86\") " pod="openstack/ssh-known-hosts-edpm-deployment-vpm5k" Dec 07 16:32:14 crc kubenswrapper[4716]: I1207 16:32:14.222939 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzmxt\" (UniqueName: \"kubernetes.io/projected/92420ce0-b95e-4857-b6f3-f6d4519c3d86-kube-api-access-fzmxt\") pod \"ssh-known-hosts-edpm-deployment-vpm5k\" (UID: \"92420ce0-b95e-4857-b6f3-f6d4519c3d86\") " pod="openstack/ssh-known-hosts-edpm-deployment-vpm5k" Dec 07 16:32:14 crc kubenswrapper[4716]: I1207 16:32:14.325019 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/92420ce0-b95e-4857-b6f3-f6d4519c3d86-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-vpm5k\" (UID: \"92420ce0-b95e-4857-b6f3-f6d4519c3d86\") " pod="openstack/ssh-known-hosts-edpm-deployment-vpm5k" Dec 07 16:32:14 crc kubenswrapper[4716]: I1207 16:32:14.325130 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzmxt\" (UniqueName: \"kubernetes.io/projected/92420ce0-b95e-4857-b6f3-f6d4519c3d86-kube-api-access-fzmxt\") pod \"ssh-known-hosts-edpm-deployment-vpm5k\" (UID: \"92420ce0-b95e-4857-b6f3-f6d4519c3d86\") " pod="openstack/ssh-known-hosts-edpm-deployment-vpm5k" Dec 07 16:32:14 crc kubenswrapper[4716]: I1207 16:32:14.325428 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/92420ce0-b95e-4857-b6f3-f6d4519c3d86-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-vpm5k\" (UID: \"92420ce0-b95e-4857-b6f3-f6d4519c3d86\") " pod="openstack/ssh-known-hosts-edpm-deployment-vpm5k" Dec 07 16:32:14 crc kubenswrapper[4716]: I1207 16:32:14.329535 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/92420ce0-b95e-4857-b6f3-f6d4519c3d86-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-vpm5k\" (UID: \"92420ce0-b95e-4857-b6f3-f6d4519c3d86\") " pod="openstack/ssh-known-hosts-edpm-deployment-vpm5k" Dec 07 16:32:14 crc kubenswrapper[4716]: I1207 16:32:14.329612 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/92420ce0-b95e-4857-b6f3-f6d4519c3d86-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-vpm5k\" (UID: \"92420ce0-b95e-4857-b6f3-f6d4519c3d86\") " pod="openstack/ssh-known-hosts-edpm-deployment-vpm5k" Dec 07 16:32:14 crc kubenswrapper[4716]: I1207 16:32:14.341641 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzmxt\" (UniqueName: \"kubernetes.io/projected/92420ce0-b95e-4857-b6f3-f6d4519c3d86-kube-api-access-fzmxt\") pod \"ssh-known-hosts-edpm-deployment-vpm5k\" (UID: \"92420ce0-b95e-4857-b6f3-f6d4519c3d86\") " pod="openstack/ssh-known-hosts-edpm-deployment-vpm5k" Dec 07 16:32:14 crc kubenswrapper[4716]: I1207 16:32:14.481873 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-vpm5k" Dec 07 16:32:14 crc kubenswrapper[4716]: I1207 16:32:14.813377 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-vpm5k"] Dec 07 16:32:15 crc kubenswrapper[4716]: I1207 16:32:15.053965 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-vpm5k" event={"ID":"92420ce0-b95e-4857-b6f3-f6d4519c3d86","Type":"ContainerStarted","Data":"d00c2a13c8cf8a8119dc51c1988ebb7872aaa7e16dd2c3e793b59cf8679ac273"} Dec 07 16:32:19 crc kubenswrapper[4716]: I1207 16:32:19.134852 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-vpm5k" event={"ID":"92420ce0-b95e-4857-b6f3-f6d4519c3d86","Type":"ContainerStarted","Data":"5324470dcdc4fd104a4a806e6c42ba436cca5f23e0295c5ca58921b157170312"} Dec 07 16:32:19 crc kubenswrapper[4716]: I1207 16:32:19.659195 4716 scope.go:117] "RemoveContainer" containerID="bae1a3e9d80f5b00293cade8b7601799399c2883ae101574459a4c01e523a154" Dec 07 16:32:19 crc kubenswrapper[4716]: E1207 16:32:19.659498 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 16:32:21 crc kubenswrapper[4716]: I1207 16:32:21.171631 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-vpm5k" podStartSLOduration=4.429418291 podStartE2EDuration="7.171609314s" podCreationTimestamp="2025-12-07 16:32:14 +0000 UTC" firstStartedPulling="2025-12-07 16:32:14.816254717 +0000 UTC m=+1797.506539629" lastFinishedPulling="2025-12-07 16:32:17.55844573 +0000 UTC m=+1800.248730652" observedRunningTime="2025-12-07 16:32:21.168552822 +0000 UTC m=+1803.858837744" watchObservedRunningTime="2025-12-07 16:32:21.171609314 +0000 UTC m=+1803.861894236" Dec 07 16:32:29 crc kubenswrapper[4716]: I1207 16:32:29.224261 4716 generic.go:334] "Generic (PLEG): container finished" podID="92420ce0-b95e-4857-b6f3-f6d4519c3d86" containerID="5324470dcdc4fd104a4a806e6c42ba436cca5f23e0295c5ca58921b157170312" exitCode=0 Dec 07 16:32:29 crc kubenswrapper[4716]: I1207 16:32:29.224343 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-vpm5k" event={"ID":"92420ce0-b95e-4857-b6f3-f6d4519c3d86","Type":"ContainerDied","Data":"5324470dcdc4fd104a4a806e6c42ba436cca5f23e0295c5ca58921b157170312"} Dec 07 16:32:30 crc kubenswrapper[4716]: I1207 16:32:30.671196 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-vpm5k" Dec 07 16:32:30 crc kubenswrapper[4716]: I1207 16:32:30.785134 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/92420ce0-b95e-4857-b6f3-f6d4519c3d86-ssh-key-openstack-edpm-ipam\") pod \"92420ce0-b95e-4857-b6f3-f6d4519c3d86\" (UID: \"92420ce0-b95e-4857-b6f3-f6d4519c3d86\") " Dec 07 16:32:30 crc kubenswrapper[4716]: I1207 16:32:30.785265 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fzmxt\" (UniqueName: \"kubernetes.io/projected/92420ce0-b95e-4857-b6f3-f6d4519c3d86-kube-api-access-fzmxt\") pod \"92420ce0-b95e-4857-b6f3-f6d4519c3d86\" (UID: \"92420ce0-b95e-4857-b6f3-f6d4519c3d86\") " Dec 07 16:32:30 crc kubenswrapper[4716]: I1207 16:32:30.785322 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/92420ce0-b95e-4857-b6f3-f6d4519c3d86-inventory-0\") pod \"92420ce0-b95e-4857-b6f3-f6d4519c3d86\" (UID: \"92420ce0-b95e-4857-b6f3-f6d4519c3d86\") " Dec 07 16:32:30 crc kubenswrapper[4716]: I1207 16:32:30.796406 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92420ce0-b95e-4857-b6f3-f6d4519c3d86-kube-api-access-fzmxt" (OuterVolumeSpecName: "kube-api-access-fzmxt") pod "92420ce0-b95e-4857-b6f3-f6d4519c3d86" (UID: "92420ce0-b95e-4857-b6f3-f6d4519c3d86"). InnerVolumeSpecName "kube-api-access-fzmxt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:32:30 crc kubenswrapper[4716]: I1207 16:32:30.814930 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92420ce0-b95e-4857-b6f3-f6d4519c3d86-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "92420ce0-b95e-4857-b6f3-f6d4519c3d86" (UID: "92420ce0-b95e-4857-b6f3-f6d4519c3d86"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:32:30 crc kubenswrapper[4716]: I1207 16:32:30.820374 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92420ce0-b95e-4857-b6f3-f6d4519c3d86-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "92420ce0-b95e-4857-b6f3-f6d4519c3d86" (UID: "92420ce0-b95e-4857-b6f3-f6d4519c3d86"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:32:30 crc kubenswrapper[4716]: I1207 16:32:30.888579 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fzmxt\" (UniqueName: \"kubernetes.io/projected/92420ce0-b95e-4857-b6f3-f6d4519c3d86-kube-api-access-fzmxt\") on node \"crc\" DevicePath \"\"" Dec 07 16:32:30 crc kubenswrapper[4716]: I1207 16:32:30.888957 4716 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/92420ce0-b95e-4857-b6f3-f6d4519c3d86-inventory-0\") on node \"crc\" DevicePath \"\"" Dec 07 16:32:30 crc kubenswrapper[4716]: I1207 16:32:30.888970 4716 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/92420ce0-b95e-4857-b6f3-f6d4519c3d86-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 07 16:32:31 crc kubenswrapper[4716]: I1207 16:32:31.250971 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-vpm5k" event={"ID":"92420ce0-b95e-4857-b6f3-f6d4519c3d86","Type":"ContainerDied","Data":"d00c2a13c8cf8a8119dc51c1988ebb7872aaa7e16dd2c3e793b59cf8679ac273"} Dec 07 16:32:31 crc kubenswrapper[4716]: I1207 16:32:31.251023 4716 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d00c2a13c8cf8a8119dc51c1988ebb7872aaa7e16dd2c3e793b59cf8679ac273" Dec 07 16:32:31 crc kubenswrapper[4716]: I1207 16:32:31.251113 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-vpm5k" Dec 07 16:32:31 crc kubenswrapper[4716]: I1207 16:32:31.335826 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-9c27f"] Dec 07 16:32:31 crc kubenswrapper[4716]: E1207 16:32:31.336325 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92420ce0-b95e-4857-b6f3-f6d4519c3d86" containerName="ssh-known-hosts-edpm-deployment" Dec 07 16:32:31 crc kubenswrapper[4716]: I1207 16:32:31.336337 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="92420ce0-b95e-4857-b6f3-f6d4519c3d86" containerName="ssh-known-hosts-edpm-deployment" Dec 07 16:32:31 crc kubenswrapper[4716]: I1207 16:32:31.336529 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="92420ce0-b95e-4857-b6f3-f6d4519c3d86" containerName="ssh-known-hosts-edpm-deployment" Dec 07 16:32:31 crc kubenswrapper[4716]: I1207 16:32:31.337212 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9c27f" Dec 07 16:32:31 crc kubenswrapper[4716]: I1207 16:32:31.339313 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-h9pt2" Dec 07 16:32:31 crc kubenswrapper[4716]: I1207 16:32:31.339759 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 07 16:32:31 crc kubenswrapper[4716]: I1207 16:32:31.341293 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 07 16:32:31 crc kubenswrapper[4716]: I1207 16:32:31.344177 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 07 16:32:31 crc kubenswrapper[4716]: I1207 16:32:31.349146 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-9c27f"] Dec 07 16:32:31 crc kubenswrapper[4716]: I1207 16:32:31.502491 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0d4b7b56-033c-41be-b46f-e4090775ab5c-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-9c27f\" (UID: \"0d4b7b56-033c-41be-b46f-e4090775ab5c\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9c27f" Dec 07 16:32:31 crc kubenswrapper[4716]: I1207 16:32:31.502549 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6jl7\" (UniqueName: \"kubernetes.io/projected/0d4b7b56-033c-41be-b46f-e4090775ab5c-kube-api-access-r6jl7\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-9c27f\" (UID: \"0d4b7b56-033c-41be-b46f-e4090775ab5c\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9c27f" Dec 07 16:32:31 crc kubenswrapper[4716]: I1207 16:32:31.502697 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0d4b7b56-033c-41be-b46f-e4090775ab5c-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-9c27f\" (UID: \"0d4b7b56-033c-41be-b46f-e4090775ab5c\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9c27f" Dec 07 16:32:31 crc kubenswrapper[4716]: I1207 16:32:31.603990 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r6jl7\" (UniqueName: \"kubernetes.io/projected/0d4b7b56-033c-41be-b46f-e4090775ab5c-kube-api-access-r6jl7\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-9c27f\" (UID: \"0d4b7b56-033c-41be-b46f-e4090775ab5c\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9c27f" Dec 07 16:32:31 crc kubenswrapper[4716]: I1207 16:32:31.604424 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0d4b7b56-033c-41be-b46f-e4090775ab5c-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-9c27f\" (UID: \"0d4b7b56-033c-41be-b46f-e4090775ab5c\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9c27f" Dec 07 16:32:31 crc kubenswrapper[4716]: I1207 16:32:31.604619 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0d4b7b56-033c-41be-b46f-e4090775ab5c-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-9c27f\" (UID: \"0d4b7b56-033c-41be-b46f-e4090775ab5c\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9c27f" Dec 07 16:32:31 crc kubenswrapper[4716]: I1207 16:32:31.610320 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0d4b7b56-033c-41be-b46f-e4090775ab5c-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-9c27f\" (UID: \"0d4b7b56-033c-41be-b46f-e4090775ab5c\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9c27f" Dec 07 16:32:31 crc kubenswrapper[4716]: I1207 16:32:31.612745 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0d4b7b56-033c-41be-b46f-e4090775ab5c-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-9c27f\" (UID: \"0d4b7b56-033c-41be-b46f-e4090775ab5c\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9c27f" Dec 07 16:32:31 crc kubenswrapper[4716]: I1207 16:32:31.620970 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r6jl7\" (UniqueName: \"kubernetes.io/projected/0d4b7b56-033c-41be-b46f-e4090775ab5c-kube-api-access-r6jl7\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-9c27f\" (UID: \"0d4b7b56-033c-41be-b46f-e4090775ab5c\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9c27f" Dec 07 16:32:31 crc kubenswrapper[4716]: I1207 16:32:31.658023 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9c27f" Dec 07 16:32:32 crc kubenswrapper[4716]: I1207 16:32:32.657534 4716 scope.go:117] "RemoveContainer" containerID="bae1a3e9d80f5b00293cade8b7601799399c2883ae101574459a4c01e523a154" Dec 07 16:32:32 crc kubenswrapper[4716]: E1207 16:32:32.658493 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 16:32:32 crc kubenswrapper[4716]: I1207 16:32:32.686348 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-9c27f"] Dec 07 16:32:33 crc kubenswrapper[4716]: I1207 16:32:33.270908 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9c27f" event={"ID":"0d4b7b56-033c-41be-b46f-e4090775ab5c","Type":"ContainerStarted","Data":"8974682046fcb76b4270d4f461f038af4b70ae28bda25e5b7f8542be54026808"} Dec 07 16:32:34 crc kubenswrapper[4716]: I1207 16:32:34.281573 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9c27f" event={"ID":"0d4b7b56-033c-41be-b46f-e4090775ab5c","Type":"ContainerStarted","Data":"8e4f37b65ff6cadf7a638e7c4a99809b3ef76437dcf4832ad4c0af1792b8bd4e"} Dec 07 16:32:34 crc kubenswrapper[4716]: I1207 16:32:34.306404 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9c27f" podStartSLOduration=2.057361546 podStartE2EDuration="3.306385699s" podCreationTimestamp="2025-12-07 16:32:31 +0000 UTC" firstStartedPulling="2025-12-07 16:32:32.687887848 +0000 UTC m=+1815.378172760" lastFinishedPulling="2025-12-07 16:32:33.936912001 +0000 UTC m=+1816.627196913" observedRunningTime="2025-12-07 16:32:34.300185902 +0000 UTC m=+1816.990470814" watchObservedRunningTime="2025-12-07 16:32:34.306385699 +0000 UTC m=+1816.996670611" Dec 07 16:32:34 crc kubenswrapper[4716]: I1207 16:32:34.398351 4716 scope.go:117] "RemoveContainer" containerID="34dfffe27670489ce32ad670cf9fbdfa109a055adf25d21c955c2e0830783d35" Dec 07 16:32:44 crc kubenswrapper[4716]: I1207 16:32:44.391657 4716 generic.go:334] "Generic (PLEG): container finished" podID="0d4b7b56-033c-41be-b46f-e4090775ab5c" containerID="8e4f37b65ff6cadf7a638e7c4a99809b3ef76437dcf4832ad4c0af1792b8bd4e" exitCode=0 Dec 07 16:32:44 crc kubenswrapper[4716]: I1207 16:32:44.391729 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9c27f" event={"ID":"0d4b7b56-033c-41be-b46f-e4090775ab5c","Type":"ContainerDied","Data":"8e4f37b65ff6cadf7a638e7c4a99809b3ef76437dcf4832ad4c0af1792b8bd4e"} Dec 07 16:32:45 crc kubenswrapper[4716]: I1207 16:32:45.658465 4716 scope.go:117] "RemoveContainer" containerID="bae1a3e9d80f5b00293cade8b7601799399c2883ae101574459a4c01e523a154" Dec 07 16:32:45 crc kubenswrapper[4716]: E1207 16:32:45.658929 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 16:32:45 crc kubenswrapper[4716]: I1207 16:32:45.795366 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9c27f" Dec 07 16:32:45 crc kubenswrapper[4716]: I1207 16:32:45.936411 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0d4b7b56-033c-41be-b46f-e4090775ab5c-inventory\") pod \"0d4b7b56-033c-41be-b46f-e4090775ab5c\" (UID: \"0d4b7b56-033c-41be-b46f-e4090775ab5c\") " Dec 07 16:32:45 crc kubenswrapper[4716]: I1207 16:32:45.936456 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r6jl7\" (UniqueName: \"kubernetes.io/projected/0d4b7b56-033c-41be-b46f-e4090775ab5c-kube-api-access-r6jl7\") pod \"0d4b7b56-033c-41be-b46f-e4090775ab5c\" (UID: \"0d4b7b56-033c-41be-b46f-e4090775ab5c\") " Dec 07 16:32:45 crc kubenswrapper[4716]: I1207 16:32:45.936620 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0d4b7b56-033c-41be-b46f-e4090775ab5c-ssh-key\") pod \"0d4b7b56-033c-41be-b46f-e4090775ab5c\" (UID: \"0d4b7b56-033c-41be-b46f-e4090775ab5c\") " Dec 07 16:32:45 crc kubenswrapper[4716]: I1207 16:32:45.949496 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d4b7b56-033c-41be-b46f-e4090775ab5c-kube-api-access-r6jl7" (OuterVolumeSpecName: "kube-api-access-r6jl7") pod "0d4b7b56-033c-41be-b46f-e4090775ab5c" (UID: "0d4b7b56-033c-41be-b46f-e4090775ab5c"). InnerVolumeSpecName "kube-api-access-r6jl7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:32:45 crc kubenswrapper[4716]: I1207 16:32:45.962262 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d4b7b56-033c-41be-b46f-e4090775ab5c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0d4b7b56-033c-41be-b46f-e4090775ab5c" (UID: "0d4b7b56-033c-41be-b46f-e4090775ab5c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:32:45 crc kubenswrapper[4716]: I1207 16:32:45.971102 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d4b7b56-033c-41be-b46f-e4090775ab5c-inventory" (OuterVolumeSpecName: "inventory") pod "0d4b7b56-033c-41be-b46f-e4090775ab5c" (UID: "0d4b7b56-033c-41be-b46f-e4090775ab5c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:32:46 crc kubenswrapper[4716]: I1207 16:32:46.038387 4716 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0d4b7b56-033c-41be-b46f-e4090775ab5c-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 07 16:32:46 crc kubenswrapper[4716]: I1207 16:32:46.039050 4716 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0d4b7b56-033c-41be-b46f-e4090775ab5c-inventory\") on node \"crc\" DevicePath \"\"" Dec 07 16:32:46 crc kubenswrapper[4716]: I1207 16:32:46.039153 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r6jl7\" (UniqueName: \"kubernetes.io/projected/0d4b7b56-033c-41be-b46f-e4090775ab5c-kube-api-access-r6jl7\") on node \"crc\" DevicePath \"\"" Dec 07 16:32:46 crc kubenswrapper[4716]: I1207 16:32:46.420756 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9c27f" event={"ID":"0d4b7b56-033c-41be-b46f-e4090775ab5c","Type":"ContainerDied","Data":"8974682046fcb76b4270d4f461f038af4b70ae28bda25e5b7f8542be54026808"} Dec 07 16:32:46 crc kubenswrapper[4716]: I1207 16:32:46.420827 4716 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8974682046fcb76b4270d4f461f038af4b70ae28bda25e5b7f8542be54026808" Dec 07 16:32:46 crc kubenswrapper[4716]: I1207 16:32:46.420873 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9c27f" Dec 07 16:32:46 crc kubenswrapper[4716]: I1207 16:32:46.516262 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xsl26"] Dec 07 16:32:46 crc kubenswrapper[4716]: E1207 16:32:46.516672 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d4b7b56-033c-41be-b46f-e4090775ab5c" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 07 16:32:46 crc kubenswrapper[4716]: I1207 16:32:46.516688 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d4b7b56-033c-41be-b46f-e4090775ab5c" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 07 16:32:46 crc kubenswrapper[4716]: I1207 16:32:46.516878 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d4b7b56-033c-41be-b46f-e4090775ab5c" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 07 16:32:46 crc kubenswrapper[4716]: I1207 16:32:46.517509 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xsl26" Dec 07 16:32:46 crc kubenswrapper[4716]: I1207 16:32:46.521249 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 07 16:32:46 crc kubenswrapper[4716]: I1207 16:32:46.521360 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 07 16:32:46 crc kubenswrapper[4716]: I1207 16:32:46.521805 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-h9pt2" Dec 07 16:32:46 crc kubenswrapper[4716]: I1207 16:32:46.522668 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 07 16:32:46 crc kubenswrapper[4716]: I1207 16:32:46.528159 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xsl26"] Dec 07 16:32:46 crc kubenswrapper[4716]: I1207 16:32:46.599317 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3a9def06-fe00-45a7-9599-2fada793ba25-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-xsl26\" (UID: \"3a9def06-fe00-45a7-9599-2fada793ba25\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xsl26" Dec 07 16:32:46 crc kubenswrapper[4716]: I1207 16:32:46.599389 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3a9def06-fe00-45a7-9599-2fada793ba25-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-xsl26\" (UID: \"3a9def06-fe00-45a7-9599-2fada793ba25\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xsl26" Dec 07 16:32:46 crc kubenswrapper[4716]: I1207 16:32:46.599571 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d6hvj\" (UniqueName: \"kubernetes.io/projected/3a9def06-fe00-45a7-9599-2fada793ba25-kube-api-access-d6hvj\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-xsl26\" (UID: \"3a9def06-fe00-45a7-9599-2fada793ba25\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xsl26" Dec 07 16:32:46 crc kubenswrapper[4716]: I1207 16:32:46.701632 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3a9def06-fe00-45a7-9599-2fada793ba25-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-xsl26\" (UID: \"3a9def06-fe00-45a7-9599-2fada793ba25\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xsl26" Dec 07 16:32:46 crc kubenswrapper[4716]: I1207 16:32:46.701685 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3a9def06-fe00-45a7-9599-2fada793ba25-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-xsl26\" (UID: \"3a9def06-fe00-45a7-9599-2fada793ba25\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xsl26" Dec 07 16:32:46 crc kubenswrapper[4716]: I1207 16:32:46.702177 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d6hvj\" (UniqueName: \"kubernetes.io/projected/3a9def06-fe00-45a7-9599-2fada793ba25-kube-api-access-d6hvj\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-xsl26\" (UID: \"3a9def06-fe00-45a7-9599-2fada793ba25\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xsl26" Dec 07 16:32:46 crc kubenswrapper[4716]: I1207 16:32:46.705662 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3a9def06-fe00-45a7-9599-2fada793ba25-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-xsl26\" (UID: \"3a9def06-fe00-45a7-9599-2fada793ba25\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xsl26" Dec 07 16:32:46 crc kubenswrapper[4716]: I1207 16:32:46.705806 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3a9def06-fe00-45a7-9599-2fada793ba25-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-xsl26\" (UID: \"3a9def06-fe00-45a7-9599-2fada793ba25\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xsl26" Dec 07 16:32:46 crc kubenswrapper[4716]: I1207 16:32:46.720124 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d6hvj\" (UniqueName: \"kubernetes.io/projected/3a9def06-fe00-45a7-9599-2fada793ba25-kube-api-access-d6hvj\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-xsl26\" (UID: \"3a9def06-fe00-45a7-9599-2fada793ba25\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xsl26" Dec 07 16:32:46 crc kubenswrapper[4716]: I1207 16:32:46.834889 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xsl26" Dec 07 16:32:47 crc kubenswrapper[4716]: I1207 16:32:47.400924 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xsl26"] Dec 07 16:32:47 crc kubenswrapper[4716]: I1207 16:32:47.431773 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xsl26" event={"ID":"3a9def06-fe00-45a7-9599-2fada793ba25","Type":"ContainerStarted","Data":"939a5827a7ade0d1447a4cd3ad37a06e43b1448f4511c955fd8a0f2efa5ace64"} Dec 07 16:32:48 crc kubenswrapper[4716]: I1207 16:32:48.441290 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xsl26" event={"ID":"3a9def06-fe00-45a7-9599-2fada793ba25","Type":"ContainerStarted","Data":"9b00718afc30ca7f8ca0754ab97ea0b25bf4464d58cdd74e004248aedb73bcbb"} Dec 07 16:32:48 crc kubenswrapper[4716]: I1207 16:32:48.469031 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xsl26" podStartSLOduration=1.973803707 podStartE2EDuration="2.469010579s" podCreationTimestamp="2025-12-07 16:32:46 +0000 UTC" firstStartedPulling="2025-12-07 16:32:47.407064745 +0000 UTC m=+1830.097349657" lastFinishedPulling="2025-12-07 16:32:47.902271607 +0000 UTC m=+1830.592556529" observedRunningTime="2025-12-07 16:32:48.461953848 +0000 UTC m=+1831.152238760" watchObservedRunningTime="2025-12-07 16:32:48.469010579 +0000 UTC m=+1831.159295501" Dec 07 16:32:58 crc kubenswrapper[4716]: I1207 16:32:58.541264 4716 generic.go:334] "Generic (PLEG): container finished" podID="3a9def06-fe00-45a7-9599-2fada793ba25" containerID="9b00718afc30ca7f8ca0754ab97ea0b25bf4464d58cdd74e004248aedb73bcbb" exitCode=0 Dec 07 16:32:58 crc kubenswrapper[4716]: I1207 16:32:58.541396 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xsl26" event={"ID":"3a9def06-fe00-45a7-9599-2fada793ba25","Type":"ContainerDied","Data":"9b00718afc30ca7f8ca0754ab97ea0b25bf4464d58cdd74e004248aedb73bcbb"} Dec 07 16:32:59 crc kubenswrapper[4716]: I1207 16:32:59.962866 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xsl26" Dec 07 16:33:00 crc kubenswrapper[4716]: I1207 16:33:00.062201 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3a9def06-fe00-45a7-9599-2fada793ba25-inventory\") pod \"3a9def06-fe00-45a7-9599-2fada793ba25\" (UID: \"3a9def06-fe00-45a7-9599-2fada793ba25\") " Dec 07 16:33:00 crc kubenswrapper[4716]: I1207 16:33:00.062261 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3a9def06-fe00-45a7-9599-2fada793ba25-ssh-key\") pod \"3a9def06-fe00-45a7-9599-2fada793ba25\" (UID: \"3a9def06-fe00-45a7-9599-2fada793ba25\") " Dec 07 16:33:00 crc kubenswrapper[4716]: I1207 16:33:00.062332 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6hvj\" (UniqueName: \"kubernetes.io/projected/3a9def06-fe00-45a7-9599-2fada793ba25-kube-api-access-d6hvj\") pod \"3a9def06-fe00-45a7-9599-2fada793ba25\" (UID: \"3a9def06-fe00-45a7-9599-2fada793ba25\") " Dec 07 16:33:00 crc kubenswrapper[4716]: I1207 16:33:00.068585 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a9def06-fe00-45a7-9599-2fada793ba25-kube-api-access-d6hvj" (OuterVolumeSpecName: "kube-api-access-d6hvj") pod "3a9def06-fe00-45a7-9599-2fada793ba25" (UID: "3a9def06-fe00-45a7-9599-2fada793ba25"). InnerVolumeSpecName "kube-api-access-d6hvj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:33:00 crc kubenswrapper[4716]: I1207 16:33:00.090117 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a9def06-fe00-45a7-9599-2fada793ba25-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3a9def06-fe00-45a7-9599-2fada793ba25" (UID: "3a9def06-fe00-45a7-9599-2fada793ba25"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:33:00 crc kubenswrapper[4716]: I1207 16:33:00.095059 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a9def06-fe00-45a7-9599-2fada793ba25-inventory" (OuterVolumeSpecName: "inventory") pod "3a9def06-fe00-45a7-9599-2fada793ba25" (UID: "3a9def06-fe00-45a7-9599-2fada793ba25"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:33:00 crc kubenswrapper[4716]: I1207 16:33:00.164645 4716 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3a9def06-fe00-45a7-9599-2fada793ba25-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 07 16:33:00 crc kubenswrapper[4716]: I1207 16:33:00.164694 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6hvj\" (UniqueName: \"kubernetes.io/projected/3a9def06-fe00-45a7-9599-2fada793ba25-kube-api-access-d6hvj\") on node \"crc\" DevicePath \"\"" Dec 07 16:33:00 crc kubenswrapper[4716]: I1207 16:33:00.164715 4716 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3a9def06-fe00-45a7-9599-2fada793ba25-inventory\") on node \"crc\" DevicePath \"\"" Dec 07 16:33:00 crc kubenswrapper[4716]: I1207 16:33:00.562933 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xsl26" event={"ID":"3a9def06-fe00-45a7-9599-2fada793ba25","Type":"ContainerDied","Data":"939a5827a7ade0d1447a4cd3ad37a06e43b1448f4511c955fd8a0f2efa5ace64"} Dec 07 16:33:00 crc kubenswrapper[4716]: I1207 16:33:00.562999 4716 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="939a5827a7ade0d1447a4cd3ad37a06e43b1448f4511c955fd8a0f2efa5ace64" Dec 07 16:33:00 crc kubenswrapper[4716]: I1207 16:33:00.563021 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xsl26" Dec 07 16:33:00 crc kubenswrapper[4716]: I1207 16:33:00.660362 4716 scope.go:117] "RemoveContainer" containerID="bae1a3e9d80f5b00293cade8b7601799399c2883ae101574459a4c01e523a154" Dec 07 16:33:00 crc kubenswrapper[4716]: E1207 16:33:00.660984 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 16:33:00 crc kubenswrapper[4716]: I1207 16:33:00.694486 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-52qx7"] Dec 07 16:33:00 crc kubenswrapper[4716]: E1207 16:33:00.695016 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a9def06-fe00-45a7-9599-2fada793ba25" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 07 16:33:00 crc kubenswrapper[4716]: I1207 16:33:00.695140 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a9def06-fe00-45a7-9599-2fada793ba25" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 07 16:33:00 crc kubenswrapper[4716]: I1207 16:33:00.695389 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a9def06-fe00-45a7-9599-2fada793ba25" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 07 16:33:00 crc kubenswrapper[4716]: I1207 16:33:00.696339 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-52qx7" Dec 07 16:33:00 crc kubenswrapper[4716]: I1207 16:33:00.707562 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Dec 07 16:33:00 crc kubenswrapper[4716]: I1207 16:33:00.707741 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-h9pt2" Dec 07 16:33:00 crc kubenswrapper[4716]: I1207 16:33:00.708012 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Dec 07 16:33:00 crc kubenswrapper[4716]: I1207 16:33:00.708184 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Dec 07 16:33:00 crc kubenswrapper[4716]: I1207 16:33:00.708429 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Dec 07 16:33:00 crc kubenswrapper[4716]: I1207 16:33:00.708559 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 07 16:33:00 crc kubenswrapper[4716]: I1207 16:33:00.708682 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 07 16:33:00 crc kubenswrapper[4716]: I1207 16:33:00.708889 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 07 16:33:00 crc kubenswrapper[4716]: I1207 16:33:00.737424 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-52qx7"] Dec 07 16:33:00 crc kubenswrapper[4716]: I1207 16:33:00.780295 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dd92737-9433-4fe4-ab78-03ebaeb31a24-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-52qx7\" (UID: \"6dd92737-9433-4fe4-ab78-03ebaeb31a24\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-52qx7" Dec 07 16:33:00 crc kubenswrapper[4716]: I1207 16:33:00.780362 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6dd92737-9433-4fe4-ab78-03ebaeb31a24-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-52qx7\" (UID: \"6dd92737-9433-4fe4-ab78-03ebaeb31a24\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-52qx7" Dec 07 16:33:00 crc kubenswrapper[4716]: I1207 16:33:00.780387 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6dd92737-9433-4fe4-ab78-03ebaeb31a24-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-52qx7\" (UID: \"6dd92737-9433-4fe4-ab78-03ebaeb31a24\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-52qx7" Dec 07 16:33:00 crc kubenswrapper[4716]: I1207 16:33:00.780436 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dd92737-9433-4fe4-ab78-03ebaeb31a24-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-52qx7\" (UID: \"6dd92737-9433-4fe4-ab78-03ebaeb31a24\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-52qx7" Dec 07 16:33:00 crc kubenswrapper[4716]: I1207 16:33:00.780486 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/6dd92737-9433-4fe4-ab78-03ebaeb31a24-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-52qx7\" (UID: \"6dd92737-9433-4fe4-ab78-03ebaeb31a24\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-52qx7" Dec 07 16:33:00 crc kubenswrapper[4716]: I1207 16:33:00.780538 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dd92737-9433-4fe4-ab78-03ebaeb31a24-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-52qx7\" (UID: \"6dd92737-9433-4fe4-ab78-03ebaeb31a24\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-52qx7" Dec 07 16:33:00 crc kubenswrapper[4716]: I1207 16:33:00.780560 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mm4ff\" (UniqueName: \"kubernetes.io/projected/6dd92737-9433-4fe4-ab78-03ebaeb31a24-kube-api-access-mm4ff\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-52qx7\" (UID: \"6dd92737-9433-4fe4-ab78-03ebaeb31a24\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-52qx7" Dec 07 16:33:00 crc kubenswrapper[4716]: I1207 16:33:00.780591 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dd92737-9433-4fe4-ab78-03ebaeb31a24-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-52qx7\" (UID: \"6dd92737-9433-4fe4-ab78-03ebaeb31a24\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-52qx7" Dec 07 16:33:00 crc kubenswrapper[4716]: I1207 16:33:00.780633 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dd92737-9433-4fe4-ab78-03ebaeb31a24-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-52qx7\" (UID: \"6dd92737-9433-4fe4-ab78-03ebaeb31a24\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-52qx7" Dec 07 16:33:00 crc kubenswrapper[4716]: I1207 16:33:00.780668 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/6dd92737-9433-4fe4-ab78-03ebaeb31a24-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-52qx7\" (UID: \"6dd92737-9433-4fe4-ab78-03ebaeb31a24\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-52qx7" Dec 07 16:33:00 crc kubenswrapper[4716]: I1207 16:33:00.780694 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dd92737-9433-4fe4-ab78-03ebaeb31a24-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-52qx7\" (UID: \"6dd92737-9433-4fe4-ab78-03ebaeb31a24\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-52qx7" Dec 07 16:33:00 crc kubenswrapper[4716]: I1207 16:33:00.780748 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/6dd92737-9433-4fe4-ab78-03ebaeb31a24-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-52qx7\" (UID: \"6dd92737-9433-4fe4-ab78-03ebaeb31a24\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-52qx7" Dec 07 16:33:00 crc kubenswrapper[4716]: I1207 16:33:00.780782 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dd92737-9433-4fe4-ab78-03ebaeb31a24-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-52qx7\" (UID: \"6dd92737-9433-4fe4-ab78-03ebaeb31a24\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-52qx7" Dec 07 16:33:00 crc kubenswrapper[4716]: I1207 16:33:00.780857 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/6dd92737-9433-4fe4-ab78-03ebaeb31a24-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-52qx7\" (UID: \"6dd92737-9433-4fe4-ab78-03ebaeb31a24\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-52qx7" Dec 07 16:33:00 crc kubenswrapper[4716]: I1207 16:33:00.882372 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/6dd92737-9433-4fe4-ab78-03ebaeb31a24-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-52qx7\" (UID: \"6dd92737-9433-4fe4-ab78-03ebaeb31a24\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-52qx7" Dec 07 16:33:00 crc kubenswrapper[4716]: I1207 16:33:00.882703 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dd92737-9433-4fe4-ab78-03ebaeb31a24-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-52qx7\" (UID: \"6dd92737-9433-4fe4-ab78-03ebaeb31a24\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-52qx7" Dec 07 16:33:00 crc kubenswrapper[4716]: I1207 16:33:00.882753 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/6dd92737-9433-4fe4-ab78-03ebaeb31a24-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-52qx7\" (UID: \"6dd92737-9433-4fe4-ab78-03ebaeb31a24\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-52qx7" Dec 07 16:33:00 crc kubenswrapper[4716]: I1207 16:33:00.882796 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dd92737-9433-4fe4-ab78-03ebaeb31a24-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-52qx7\" (UID: \"6dd92737-9433-4fe4-ab78-03ebaeb31a24\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-52qx7" Dec 07 16:33:00 crc kubenswrapper[4716]: I1207 16:33:00.882824 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6dd92737-9433-4fe4-ab78-03ebaeb31a24-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-52qx7\" (UID: \"6dd92737-9433-4fe4-ab78-03ebaeb31a24\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-52qx7" Dec 07 16:33:00 crc kubenswrapper[4716]: I1207 16:33:00.882843 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6dd92737-9433-4fe4-ab78-03ebaeb31a24-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-52qx7\" (UID: \"6dd92737-9433-4fe4-ab78-03ebaeb31a24\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-52qx7" Dec 07 16:33:00 crc kubenswrapper[4716]: I1207 16:33:00.882880 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dd92737-9433-4fe4-ab78-03ebaeb31a24-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-52qx7\" (UID: \"6dd92737-9433-4fe4-ab78-03ebaeb31a24\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-52qx7" Dec 07 16:33:00 crc kubenswrapper[4716]: I1207 16:33:00.882903 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/6dd92737-9433-4fe4-ab78-03ebaeb31a24-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-52qx7\" (UID: \"6dd92737-9433-4fe4-ab78-03ebaeb31a24\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-52qx7" Dec 07 16:33:00 crc kubenswrapper[4716]: I1207 16:33:00.882942 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dd92737-9433-4fe4-ab78-03ebaeb31a24-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-52qx7\" (UID: \"6dd92737-9433-4fe4-ab78-03ebaeb31a24\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-52qx7" Dec 07 16:33:00 crc kubenswrapper[4716]: I1207 16:33:00.882960 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mm4ff\" (UniqueName: \"kubernetes.io/projected/6dd92737-9433-4fe4-ab78-03ebaeb31a24-kube-api-access-mm4ff\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-52qx7\" (UID: \"6dd92737-9433-4fe4-ab78-03ebaeb31a24\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-52qx7" Dec 07 16:33:00 crc kubenswrapper[4716]: I1207 16:33:00.882984 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dd92737-9433-4fe4-ab78-03ebaeb31a24-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-52qx7\" (UID: \"6dd92737-9433-4fe4-ab78-03ebaeb31a24\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-52qx7" Dec 07 16:33:00 crc kubenswrapper[4716]: I1207 16:33:00.883019 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dd92737-9433-4fe4-ab78-03ebaeb31a24-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-52qx7\" (UID: \"6dd92737-9433-4fe4-ab78-03ebaeb31a24\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-52qx7" Dec 07 16:33:00 crc kubenswrapper[4716]: I1207 16:33:00.883042 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/6dd92737-9433-4fe4-ab78-03ebaeb31a24-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-52qx7\" (UID: \"6dd92737-9433-4fe4-ab78-03ebaeb31a24\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-52qx7" Dec 07 16:33:00 crc kubenswrapper[4716]: I1207 16:33:00.883059 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dd92737-9433-4fe4-ab78-03ebaeb31a24-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-52qx7\" (UID: \"6dd92737-9433-4fe4-ab78-03ebaeb31a24\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-52qx7" Dec 07 16:33:00 crc kubenswrapper[4716]: I1207 16:33:00.886754 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dd92737-9433-4fe4-ab78-03ebaeb31a24-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-52qx7\" (UID: \"6dd92737-9433-4fe4-ab78-03ebaeb31a24\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-52qx7" Dec 07 16:33:00 crc kubenswrapper[4716]: I1207 16:33:00.886859 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dd92737-9433-4fe4-ab78-03ebaeb31a24-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-52qx7\" (UID: \"6dd92737-9433-4fe4-ab78-03ebaeb31a24\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-52qx7" Dec 07 16:33:00 crc kubenswrapper[4716]: I1207 16:33:00.887511 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/6dd92737-9433-4fe4-ab78-03ebaeb31a24-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-52qx7\" (UID: \"6dd92737-9433-4fe4-ab78-03ebaeb31a24\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-52qx7" Dec 07 16:33:00 crc kubenswrapper[4716]: I1207 16:33:00.888263 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6dd92737-9433-4fe4-ab78-03ebaeb31a24-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-52qx7\" (UID: \"6dd92737-9433-4fe4-ab78-03ebaeb31a24\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-52qx7" Dec 07 16:33:00 crc kubenswrapper[4716]: I1207 16:33:00.888560 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dd92737-9433-4fe4-ab78-03ebaeb31a24-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-52qx7\" (UID: \"6dd92737-9433-4fe4-ab78-03ebaeb31a24\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-52qx7" Dec 07 16:33:00 crc kubenswrapper[4716]: I1207 16:33:00.888667 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/6dd92737-9433-4fe4-ab78-03ebaeb31a24-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-52qx7\" (UID: \"6dd92737-9433-4fe4-ab78-03ebaeb31a24\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-52qx7" Dec 07 16:33:00 crc kubenswrapper[4716]: I1207 16:33:00.889272 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dd92737-9433-4fe4-ab78-03ebaeb31a24-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-52qx7\" (UID: \"6dd92737-9433-4fe4-ab78-03ebaeb31a24\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-52qx7" Dec 07 16:33:00 crc kubenswrapper[4716]: I1207 16:33:00.889870 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dd92737-9433-4fe4-ab78-03ebaeb31a24-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-52qx7\" (UID: \"6dd92737-9433-4fe4-ab78-03ebaeb31a24\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-52qx7" Dec 07 16:33:00 crc kubenswrapper[4716]: I1207 16:33:00.890385 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dd92737-9433-4fe4-ab78-03ebaeb31a24-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-52qx7\" (UID: \"6dd92737-9433-4fe4-ab78-03ebaeb31a24\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-52qx7" Dec 07 16:33:00 crc kubenswrapper[4716]: I1207 16:33:00.892688 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/6dd92737-9433-4fe4-ab78-03ebaeb31a24-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-52qx7\" (UID: \"6dd92737-9433-4fe4-ab78-03ebaeb31a24\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-52qx7" Dec 07 16:33:00 crc kubenswrapper[4716]: I1207 16:33:00.895600 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6dd92737-9433-4fe4-ab78-03ebaeb31a24-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-52qx7\" (UID: \"6dd92737-9433-4fe4-ab78-03ebaeb31a24\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-52qx7" Dec 07 16:33:00 crc kubenswrapper[4716]: I1207 16:33:00.896247 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dd92737-9433-4fe4-ab78-03ebaeb31a24-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-52qx7\" (UID: \"6dd92737-9433-4fe4-ab78-03ebaeb31a24\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-52qx7" Dec 07 16:33:00 crc kubenswrapper[4716]: I1207 16:33:00.897589 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/6dd92737-9433-4fe4-ab78-03ebaeb31a24-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-52qx7\" (UID: \"6dd92737-9433-4fe4-ab78-03ebaeb31a24\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-52qx7" Dec 07 16:33:00 crc kubenswrapper[4716]: I1207 16:33:00.898999 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mm4ff\" (UniqueName: \"kubernetes.io/projected/6dd92737-9433-4fe4-ab78-03ebaeb31a24-kube-api-access-mm4ff\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-52qx7\" (UID: \"6dd92737-9433-4fe4-ab78-03ebaeb31a24\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-52qx7" Dec 07 16:33:01 crc kubenswrapper[4716]: I1207 16:33:01.028414 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-52qx7" Dec 07 16:33:01 crc kubenswrapper[4716]: I1207 16:33:01.541805 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-52qx7"] Dec 07 16:33:01 crc kubenswrapper[4716]: W1207 16:33:01.543214 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6dd92737_9433_4fe4_ab78_03ebaeb31a24.slice/crio-54d2ccf8923414937512e202eb51e675d31bc2b3c5f1fedca0a255d3a893eff7 WatchSource:0}: Error finding container 54d2ccf8923414937512e202eb51e675d31bc2b3c5f1fedca0a255d3a893eff7: Status 404 returned error can't find the container with id 54d2ccf8923414937512e202eb51e675d31bc2b3c5f1fedca0a255d3a893eff7 Dec 07 16:33:01 crc kubenswrapper[4716]: I1207 16:33:01.578420 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-52qx7" event={"ID":"6dd92737-9433-4fe4-ab78-03ebaeb31a24","Type":"ContainerStarted","Data":"54d2ccf8923414937512e202eb51e675d31bc2b3c5f1fedca0a255d3a893eff7"} Dec 07 16:33:02 crc kubenswrapper[4716]: I1207 16:33:02.588113 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-52qx7" event={"ID":"6dd92737-9433-4fe4-ab78-03ebaeb31a24","Type":"ContainerStarted","Data":"5547edd90dfca61de4a1176bfe5484760b2069051f03a0fa2ca29ad7753aeef2"} Dec 07 16:33:02 crc kubenswrapper[4716]: I1207 16:33:02.631471 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-52qx7" podStartSLOduration=2.213002535 podStartE2EDuration="2.631450775s" podCreationTimestamp="2025-12-07 16:33:00 +0000 UTC" firstStartedPulling="2025-12-07 16:33:01.548392311 +0000 UTC m=+1844.238677233" lastFinishedPulling="2025-12-07 16:33:01.966840551 +0000 UTC m=+1844.657125473" observedRunningTime="2025-12-07 16:33:02.60937739 +0000 UTC m=+1845.299662342" watchObservedRunningTime="2025-12-07 16:33:02.631450775 +0000 UTC m=+1845.321735697" Dec 07 16:33:12 crc kubenswrapper[4716]: I1207 16:33:12.658479 4716 scope.go:117] "RemoveContainer" containerID="bae1a3e9d80f5b00293cade8b7601799399c2883ae101574459a4c01e523a154" Dec 07 16:33:12 crc kubenswrapper[4716]: E1207 16:33:12.661968 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 16:33:25 crc kubenswrapper[4716]: I1207 16:33:25.657974 4716 scope.go:117] "RemoveContainer" containerID="bae1a3e9d80f5b00293cade8b7601799399c2883ae101574459a4c01e523a154" Dec 07 16:33:25 crc kubenswrapper[4716]: E1207 16:33:25.658816 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 16:33:37 crc kubenswrapper[4716]: I1207 16:33:37.665152 4716 scope.go:117] "RemoveContainer" containerID="bae1a3e9d80f5b00293cade8b7601799399c2883ae101574459a4c01e523a154" Dec 07 16:33:37 crc kubenswrapper[4716]: E1207 16:33:37.665875 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 16:33:48 crc kubenswrapper[4716]: I1207 16:33:48.016587 4716 generic.go:334] "Generic (PLEG): container finished" podID="6dd92737-9433-4fe4-ab78-03ebaeb31a24" containerID="5547edd90dfca61de4a1176bfe5484760b2069051f03a0fa2ca29ad7753aeef2" exitCode=0 Dec 07 16:33:48 crc kubenswrapper[4716]: I1207 16:33:48.016676 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-52qx7" event={"ID":"6dd92737-9433-4fe4-ab78-03ebaeb31a24","Type":"ContainerDied","Data":"5547edd90dfca61de4a1176bfe5484760b2069051f03a0fa2ca29ad7753aeef2"} Dec 07 16:33:49 crc kubenswrapper[4716]: I1207 16:33:49.666435 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-52qx7" Dec 07 16:33:49 crc kubenswrapper[4716]: I1207 16:33:49.832939 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dd92737-9433-4fe4-ab78-03ebaeb31a24-neutron-metadata-combined-ca-bundle\") pod \"6dd92737-9433-4fe4-ab78-03ebaeb31a24\" (UID: \"6dd92737-9433-4fe4-ab78-03ebaeb31a24\") " Dec 07 16:33:49 crc kubenswrapper[4716]: I1207 16:33:49.832994 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6dd92737-9433-4fe4-ab78-03ebaeb31a24-inventory\") pod \"6dd92737-9433-4fe4-ab78-03ebaeb31a24\" (UID: \"6dd92737-9433-4fe4-ab78-03ebaeb31a24\") " Dec 07 16:33:49 crc kubenswrapper[4716]: I1207 16:33:49.833019 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mm4ff\" (UniqueName: \"kubernetes.io/projected/6dd92737-9433-4fe4-ab78-03ebaeb31a24-kube-api-access-mm4ff\") pod \"6dd92737-9433-4fe4-ab78-03ebaeb31a24\" (UID: \"6dd92737-9433-4fe4-ab78-03ebaeb31a24\") " Dec 07 16:33:49 crc kubenswrapper[4716]: I1207 16:33:49.833042 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dd92737-9433-4fe4-ab78-03ebaeb31a24-libvirt-combined-ca-bundle\") pod \"6dd92737-9433-4fe4-ab78-03ebaeb31a24\" (UID: \"6dd92737-9433-4fe4-ab78-03ebaeb31a24\") " Dec 07 16:33:49 crc kubenswrapper[4716]: I1207 16:33:49.833070 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dd92737-9433-4fe4-ab78-03ebaeb31a24-repo-setup-combined-ca-bundle\") pod \"6dd92737-9433-4fe4-ab78-03ebaeb31a24\" (UID: \"6dd92737-9433-4fe4-ab78-03ebaeb31a24\") " Dec 07 16:33:49 crc kubenswrapper[4716]: I1207 16:33:49.833106 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/6dd92737-9433-4fe4-ab78-03ebaeb31a24-openstack-edpm-ipam-ovn-default-certs-0\") pod \"6dd92737-9433-4fe4-ab78-03ebaeb31a24\" (UID: \"6dd92737-9433-4fe4-ab78-03ebaeb31a24\") " Dec 07 16:33:49 crc kubenswrapper[4716]: I1207 16:33:49.833128 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dd92737-9433-4fe4-ab78-03ebaeb31a24-bootstrap-combined-ca-bundle\") pod \"6dd92737-9433-4fe4-ab78-03ebaeb31a24\" (UID: \"6dd92737-9433-4fe4-ab78-03ebaeb31a24\") " Dec 07 16:33:49 crc kubenswrapper[4716]: I1207 16:33:49.833150 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dd92737-9433-4fe4-ab78-03ebaeb31a24-telemetry-combined-ca-bundle\") pod \"6dd92737-9433-4fe4-ab78-03ebaeb31a24\" (UID: \"6dd92737-9433-4fe4-ab78-03ebaeb31a24\") " Dec 07 16:33:49 crc kubenswrapper[4716]: I1207 16:33:49.833186 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dd92737-9433-4fe4-ab78-03ebaeb31a24-ovn-combined-ca-bundle\") pod \"6dd92737-9433-4fe4-ab78-03ebaeb31a24\" (UID: \"6dd92737-9433-4fe4-ab78-03ebaeb31a24\") " Dec 07 16:33:49 crc kubenswrapper[4716]: I1207 16:33:49.833204 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dd92737-9433-4fe4-ab78-03ebaeb31a24-nova-combined-ca-bundle\") pod \"6dd92737-9433-4fe4-ab78-03ebaeb31a24\" (UID: \"6dd92737-9433-4fe4-ab78-03ebaeb31a24\") " Dec 07 16:33:49 crc kubenswrapper[4716]: I1207 16:33:49.833222 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6dd92737-9433-4fe4-ab78-03ebaeb31a24-ssh-key\") pod \"6dd92737-9433-4fe4-ab78-03ebaeb31a24\" (UID: \"6dd92737-9433-4fe4-ab78-03ebaeb31a24\") " Dec 07 16:33:49 crc kubenswrapper[4716]: I1207 16:33:49.833245 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/6dd92737-9433-4fe4-ab78-03ebaeb31a24-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"6dd92737-9433-4fe4-ab78-03ebaeb31a24\" (UID: \"6dd92737-9433-4fe4-ab78-03ebaeb31a24\") " Dec 07 16:33:49 crc kubenswrapper[4716]: I1207 16:33:49.833268 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/6dd92737-9433-4fe4-ab78-03ebaeb31a24-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"6dd92737-9433-4fe4-ab78-03ebaeb31a24\" (UID: \"6dd92737-9433-4fe4-ab78-03ebaeb31a24\") " Dec 07 16:33:49 crc kubenswrapper[4716]: I1207 16:33:49.833333 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/6dd92737-9433-4fe4-ab78-03ebaeb31a24-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"6dd92737-9433-4fe4-ab78-03ebaeb31a24\" (UID: \"6dd92737-9433-4fe4-ab78-03ebaeb31a24\") " Dec 07 16:33:49 crc kubenswrapper[4716]: I1207 16:33:49.840989 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6dd92737-9433-4fe4-ab78-03ebaeb31a24-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "6dd92737-9433-4fe4-ab78-03ebaeb31a24" (UID: "6dd92737-9433-4fe4-ab78-03ebaeb31a24"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:33:49 crc kubenswrapper[4716]: I1207 16:33:49.841096 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6dd92737-9433-4fe4-ab78-03ebaeb31a24-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "6dd92737-9433-4fe4-ab78-03ebaeb31a24" (UID: "6dd92737-9433-4fe4-ab78-03ebaeb31a24"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:33:49 crc kubenswrapper[4716]: I1207 16:33:49.841561 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6dd92737-9433-4fe4-ab78-03ebaeb31a24-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "6dd92737-9433-4fe4-ab78-03ebaeb31a24" (UID: "6dd92737-9433-4fe4-ab78-03ebaeb31a24"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:33:49 crc kubenswrapper[4716]: I1207 16:33:49.841774 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6dd92737-9433-4fe4-ab78-03ebaeb31a24-kube-api-access-mm4ff" (OuterVolumeSpecName: "kube-api-access-mm4ff") pod "6dd92737-9433-4fe4-ab78-03ebaeb31a24" (UID: "6dd92737-9433-4fe4-ab78-03ebaeb31a24"). InnerVolumeSpecName "kube-api-access-mm4ff". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:33:49 crc kubenswrapper[4716]: I1207 16:33:49.841883 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6dd92737-9433-4fe4-ab78-03ebaeb31a24-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "6dd92737-9433-4fe4-ab78-03ebaeb31a24" (UID: "6dd92737-9433-4fe4-ab78-03ebaeb31a24"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:33:49 crc kubenswrapper[4716]: I1207 16:33:49.842193 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6dd92737-9433-4fe4-ab78-03ebaeb31a24-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "6dd92737-9433-4fe4-ab78-03ebaeb31a24" (UID: "6dd92737-9433-4fe4-ab78-03ebaeb31a24"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:33:49 crc kubenswrapper[4716]: I1207 16:33:49.842739 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6dd92737-9433-4fe4-ab78-03ebaeb31a24-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "6dd92737-9433-4fe4-ab78-03ebaeb31a24" (UID: "6dd92737-9433-4fe4-ab78-03ebaeb31a24"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:33:49 crc kubenswrapper[4716]: I1207 16:33:49.843765 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6dd92737-9433-4fe4-ab78-03ebaeb31a24-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "6dd92737-9433-4fe4-ab78-03ebaeb31a24" (UID: "6dd92737-9433-4fe4-ab78-03ebaeb31a24"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:33:49 crc kubenswrapper[4716]: I1207 16:33:49.844492 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6dd92737-9433-4fe4-ab78-03ebaeb31a24-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "6dd92737-9433-4fe4-ab78-03ebaeb31a24" (UID: "6dd92737-9433-4fe4-ab78-03ebaeb31a24"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:33:49 crc kubenswrapper[4716]: I1207 16:33:49.847046 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6dd92737-9433-4fe4-ab78-03ebaeb31a24-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "6dd92737-9433-4fe4-ab78-03ebaeb31a24" (UID: "6dd92737-9433-4fe4-ab78-03ebaeb31a24"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:33:49 crc kubenswrapper[4716]: I1207 16:33:49.849017 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6dd92737-9433-4fe4-ab78-03ebaeb31a24-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "6dd92737-9433-4fe4-ab78-03ebaeb31a24" (UID: "6dd92737-9433-4fe4-ab78-03ebaeb31a24"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:33:49 crc kubenswrapper[4716]: I1207 16:33:49.856214 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6dd92737-9433-4fe4-ab78-03ebaeb31a24-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "6dd92737-9433-4fe4-ab78-03ebaeb31a24" (UID: "6dd92737-9433-4fe4-ab78-03ebaeb31a24"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:33:49 crc kubenswrapper[4716]: I1207 16:33:49.865765 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6dd92737-9433-4fe4-ab78-03ebaeb31a24-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6dd92737-9433-4fe4-ab78-03ebaeb31a24" (UID: "6dd92737-9433-4fe4-ab78-03ebaeb31a24"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:33:49 crc kubenswrapper[4716]: I1207 16:33:49.868734 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6dd92737-9433-4fe4-ab78-03ebaeb31a24-inventory" (OuterVolumeSpecName: "inventory") pod "6dd92737-9433-4fe4-ab78-03ebaeb31a24" (UID: "6dd92737-9433-4fe4-ab78-03ebaeb31a24"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:33:49 crc kubenswrapper[4716]: I1207 16:33:49.934832 4716 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/6dd92737-9433-4fe4-ab78-03ebaeb31a24-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 07 16:33:49 crc kubenswrapper[4716]: I1207 16:33:49.934869 4716 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dd92737-9433-4fe4-ab78-03ebaeb31a24-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 16:33:49 crc kubenswrapper[4716]: I1207 16:33:49.934879 4716 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dd92737-9433-4fe4-ab78-03ebaeb31a24-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 16:33:49 crc kubenswrapper[4716]: I1207 16:33:49.934890 4716 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dd92737-9433-4fe4-ab78-03ebaeb31a24-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 16:33:49 crc kubenswrapper[4716]: I1207 16:33:49.934898 4716 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dd92737-9433-4fe4-ab78-03ebaeb31a24-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 16:33:49 crc kubenswrapper[4716]: I1207 16:33:49.934907 4716 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6dd92737-9433-4fe4-ab78-03ebaeb31a24-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 07 16:33:49 crc kubenswrapper[4716]: I1207 16:33:49.934918 4716 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/6dd92737-9433-4fe4-ab78-03ebaeb31a24-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 07 16:33:49 crc kubenswrapper[4716]: I1207 16:33:49.934928 4716 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/6dd92737-9433-4fe4-ab78-03ebaeb31a24-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 07 16:33:49 crc kubenswrapper[4716]: I1207 16:33:49.934938 4716 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/6dd92737-9433-4fe4-ab78-03ebaeb31a24-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 07 16:33:49 crc kubenswrapper[4716]: I1207 16:33:49.934948 4716 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dd92737-9433-4fe4-ab78-03ebaeb31a24-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 16:33:49 crc kubenswrapper[4716]: I1207 16:33:49.934957 4716 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6dd92737-9433-4fe4-ab78-03ebaeb31a24-inventory\") on node \"crc\" DevicePath \"\"" Dec 07 16:33:49 crc kubenswrapper[4716]: I1207 16:33:49.934965 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mm4ff\" (UniqueName: \"kubernetes.io/projected/6dd92737-9433-4fe4-ab78-03ebaeb31a24-kube-api-access-mm4ff\") on node \"crc\" DevicePath \"\"" Dec 07 16:33:49 crc kubenswrapper[4716]: I1207 16:33:49.934973 4716 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dd92737-9433-4fe4-ab78-03ebaeb31a24-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 16:33:49 crc kubenswrapper[4716]: I1207 16:33:49.934981 4716 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dd92737-9433-4fe4-ab78-03ebaeb31a24-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 16:33:50 crc kubenswrapper[4716]: I1207 16:33:50.176094 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-26hn5"] Dec 07 16:33:50 crc kubenswrapper[4716]: E1207 16:33:50.176927 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6dd92737-9433-4fe4-ab78-03ebaeb31a24" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 07 16:33:50 crc kubenswrapper[4716]: I1207 16:33:50.177040 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="6dd92737-9433-4fe4-ab78-03ebaeb31a24" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 07 16:33:50 crc kubenswrapper[4716]: I1207 16:33:50.177405 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="6dd92737-9433-4fe4-ab78-03ebaeb31a24" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 07 16:33:50 crc kubenswrapper[4716]: I1207 16:33:50.178258 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-26hn5" Dec 07 16:33:50 crc kubenswrapper[4716]: I1207 16:33:50.181404 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Dec 07 16:33:50 crc kubenswrapper[4716]: I1207 16:33:50.188995 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-26hn5"] Dec 07 16:33:50 crc kubenswrapper[4716]: I1207 16:33:50.239373 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-52qx7" event={"ID":"6dd92737-9433-4fe4-ab78-03ebaeb31a24","Type":"ContainerDied","Data":"54d2ccf8923414937512e202eb51e675d31bc2b3c5f1fedca0a255d3a893eff7"} Dec 07 16:33:50 crc kubenswrapper[4716]: I1207 16:33:50.239694 4716 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="54d2ccf8923414937512e202eb51e675d31bc2b3c5f1fedca0a255d3a893eff7" Dec 07 16:33:50 crc kubenswrapper[4716]: I1207 16:33:50.240384 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-52qx7" Dec 07 16:33:50 crc kubenswrapper[4716]: I1207 16:33:50.344609 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/b9e8242a-2734-47b9-9f14-7eea8d9ad3c4-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-26hn5\" (UID: \"b9e8242a-2734-47b9-9f14-7eea8d9ad3c4\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-26hn5" Dec 07 16:33:50 crc kubenswrapper[4716]: I1207 16:33:50.344678 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hp26k\" (UniqueName: \"kubernetes.io/projected/b9e8242a-2734-47b9-9f14-7eea8d9ad3c4-kube-api-access-hp26k\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-26hn5\" (UID: \"b9e8242a-2734-47b9-9f14-7eea8d9ad3c4\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-26hn5" Dec 07 16:33:50 crc kubenswrapper[4716]: I1207 16:33:50.344698 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b9e8242a-2734-47b9-9f14-7eea8d9ad3c4-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-26hn5\" (UID: \"b9e8242a-2734-47b9-9f14-7eea8d9ad3c4\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-26hn5" Dec 07 16:33:50 crc kubenswrapper[4716]: I1207 16:33:50.344771 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b9e8242a-2734-47b9-9f14-7eea8d9ad3c4-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-26hn5\" (UID: \"b9e8242a-2734-47b9-9f14-7eea8d9ad3c4\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-26hn5" Dec 07 16:33:50 crc kubenswrapper[4716]: I1207 16:33:50.344848 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9e8242a-2734-47b9-9f14-7eea8d9ad3c4-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-26hn5\" (UID: \"b9e8242a-2734-47b9-9f14-7eea8d9ad3c4\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-26hn5" Dec 07 16:33:50 crc kubenswrapper[4716]: I1207 16:33:50.446642 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9e8242a-2734-47b9-9f14-7eea8d9ad3c4-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-26hn5\" (UID: \"b9e8242a-2734-47b9-9f14-7eea8d9ad3c4\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-26hn5" Dec 07 16:33:50 crc kubenswrapper[4716]: I1207 16:33:50.446990 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/b9e8242a-2734-47b9-9f14-7eea8d9ad3c4-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-26hn5\" (UID: \"b9e8242a-2734-47b9-9f14-7eea8d9ad3c4\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-26hn5" Dec 07 16:33:50 crc kubenswrapper[4716]: I1207 16:33:50.447049 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hp26k\" (UniqueName: \"kubernetes.io/projected/b9e8242a-2734-47b9-9f14-7eea8d9ad3c4-kube-api-access-hp26k\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-26hn5\" (UID: \"b9e8242a-2734-47b9-9f14-7eea8d9ad3c4\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-26hn5" Dec 07 16:33:50 crc kubenswrapper[4716]: I1207 16:33:50.447097 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b9e8242a-2734-47b9-9f14-7eea8d9ad3c4-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-26hn5\" (UID: \"b9e8242a-2734-47b9-9f14-7eea8d9ad3c4\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-26hn5" Dec 07 16:33:50 crc kubenswrapper[4716]: I1207 16:33:50.447193 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b9e8242a-2734-47b9-9f14-7eea8d9ad3c4-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-26hn5\" (UID: \"b9e8242a-2734-47b9-9f14-7eea8d9ad3c4\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-26hn5" Dec 07 16:33:50 crc kubenswrapper[4716]: I1207 16:33:50.448176 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/b9e8242a-2734-47b9-9f14-7eea8d9ad3c4-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-26hn5\" (UID: \"b9e8242a-2734-47b9-9f14-7eea8d9ad3c4\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-26hn5" Dec 07 16:33:50 crc kubenswrapper[4716]: I1207 16:33:50.453491 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b9e8242a-2734-47b9-9f14-7eea8d9ad3c4-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-26hn5\" (UID: \"b9e8242a-2734-47b9-9f14-7eea8d9ad3c4\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-26hn5" Dec 07 16:33:50 crc kubenswrapper[4716]: I1207 16:33:50.453593 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9e8242a-2734-47b9-9f14-7eea8d9ad3c4-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-26hn5\" (UID: \"b9e8242a-2734-47b9-9f14-7eea8d9ad3c4\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-26hn5" Dec 07 16:33:50 crc kubenswrapper[4716]: I1207 16:33:50.455004 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b9e8242a-2734-47b9-9f14-7eea8d9ad3c4-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-26hn5\" (UID: \"b9e8242a-2734-47b9-9f14-7eea8d9ad3c4\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-26hn5" Dec 07 16:33:50 crc kubenswrapper[4716]: I1207 16:33:50.466927 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hp26k\" (UniqueName: \"kubernetes.io/projected/b9e8242a-2734-47b9-9f14-7eea8d9ad3c4-kube-api-access-hp26k\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-26hn5\" (UID: \"b9e8242a-2734-47b9-9f14-7eea8d9ad3c4\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-26hn5" Dec 07 16:33:50 crc kubenswrapper[4716]: I1207 16:33:50.508780 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-26hn5" Dec 07 16:33:50 crc kubenswrapper[4716]: I1207 16:33:50.935253 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-26hn5"] Dec 07 16:33:51 crc kubenswrapper[4716]: I1207 16:33:51.252063 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-26hn5" event={"ID":"b9e8242a-2734-47b9-9f14-7eea8d9ad3c4","Type":"ContainerStarted","Data":"37ae1d595e95ce1f1869a877cfb7f5e9c61561c9fe95d2f9e326cad01b2221fe"} Dec 07 16:33:52 crc kubenswrapper[4716]: I1207 16:33:52.264217 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-26hn5" event={"ID":"b9e8242a-2734-47b9-9f14-7eea8d9ad3c4","Type":"ContainerStarted","Data":"2abbb8c26fd2a1608dc2a5e2ce122f55bb37987cdbd986867ebe77fb08057648"} Dec 07 16:33:52 crc kubenswrapper[4716]: I1207 16:33:52.292334 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-26hn5" podStartSLOduration=1.7980355810000002 podStartE2EDuration="2.292314648s" podCreationTimestamp="2025-12-07 16:33:50 +0000 UTC" firstStartedPulling="2025-12-07 16:33:50.945499996 +0000 UTC m=+1893.635784908" lastFinishedPulling="2025-12-07 16:33:51.439779053 +0000 UTC m=+1894.130063975" observedRunningTime="2025-12-07 16:33:52.283979413 +0000 UTC m=+1894.974264325" watchObservedRunningTime="2025-12-07 16:33:52.292314648 +0000 UTC m=+1894.982599550" Dec 07 16:33:52 crc kubenswrapper[4716]: I1207 16:33:52.657694 4716 scope.go:117] "RemoveContainer" containerID="bae1a3e9d80f5b00293cade8b7601799399c2883ae101574459a4c01e523a154" Dec 07 16:33:52 crc kubenswrapper[4716]: E1207 16:33:52.657942 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 16:34:05 crc kubenswrapper[4716]: I1207 16:34:05.658203 4716 scope.go:117] "RemoveContainer" containerID="bae1a3e9d80f5b00293cade8b7601799399c2883ae101574459a4c01e523a154" Dec 07 16:34:05 crc kubenswrapper[4716]: E1207 16:34:05.659476 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 16:34:17 crc kubenswrapper[4716]: I1207 16:34:17.664847 4716 scope.go:117] "RemoveContainer" containerID="bae1a3e9d80f5b00293cade8b7601799399c2883ae101574459a4c01e523a154" Dec 07 16:34:17 crc kubenswrapper[4716]: E1207 16:34:17.665687 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 16:34:31 crc kubenswrapper[4716]: I1207 16:34:31.657719 4716 scope.go:117] "RemoveContainer" containerID="bae1a3e9d80f5b00293cade8b7601799399c2883ae101574459a4c01e523a154" Dec 07 16:34:31 crc kubenswrapper[4716]: E1207 16:34:31.658908 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 16:34:45 crc kubenswrapper[4716]: I1207 16:34:45.657957 4716 scope.go:117] "RemoveContainer" containerID="bae1a3e9d80f5b00293cade8b7601799399c2883ae101574459a4c01e523a154" Dec 07 16:34:45 crc kubenswrapper[4716]: E1207 16:34:45.658874 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 16:35:00 crc kubenswrapper[4716]: I1207 16:35:00.658251 4716 scope.go:117] "RemoveContainer" containerID="bae1a3e9d80f5b00293cade8b7601799399c2883ae101574459a4c01e523a154" Dec 07 16:35:00 crc kubenswrapper[4716]: E1207 16:35:00.658934 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 16:35:05 crc kubenswrapper[4716]: I1207 16:35:05.936572 4716 generic.go:334] "Generic (PLEG): container finished" podID="b9e8242a-2734-47b9-9f14-7eea8d9ad3c4" containerID="2abbb8c26fd2a1608dc2a5e2ce122f55bb37987cdbd986867ebe77fb08057648" exitCode=0 Dec 07 16:35:05 crc kubenswrapper[4716]: I1207 16:35:05.936688 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-26hn5" event={"ID":"b9e8242a-2734-47b9-9f14-7eea8d9ad3c4","Type":"ContainerDied","Data":"2abbb8c26fd2a1608dc2a5e2ce122f55bb37987cdbd986867ebe77fb08057648"} Dec 07 16:35:07 crc kubenswrapper[4716]: I1207 16:35:07.415035 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-26hn5" Dec 07 16:35:07 crc kubenswrapper[4716]: I1207 16:35:07.610781 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hp26k\" (UniqueName: \"kubernetes.io/projected/b9e8242a-2734-47b9-9f14-7eea8d9ad3c4-kube-api-access-hp26k\") pod \"b9e8242a-2734-47b9-9f14-7eea8d9ad3c4\" (UID: \"b9e8242a-2734-47b9-9f14-7eea8d9ad3c4\") " Dec 07 16:35:07 crc kubenswrapper[4716]: I1207 16:35:07.611241 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b9e8242a-2734-47b9-9f14-7eea8d9ad3c4-ssh-key\") pod \"b9e8242a-2734-47b9-9f14-7eea8d9ad3c4\" (UID: \"b9e8242a-2734-47b9-9f14-7eea8d9ad3c4\") " Dec 07 16:35:07 crc kubenswrapper[4716]: I1207 16:35:07.611276 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/b9e8242a-2734-47b9-9f14-7eea8d9ad3c4-ovncontroller-config-0\") pod \"b9e8242a-2734-47b9-9f14-7eea8d9ad3c4\" (UID: \"b9e8242a-2734-47b9-9f14-7eea8d9ad3c4\") " Dec 07 16:35:07 crc kubenswrapper[4716]: I1207 16:35:07.611346 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b9e8242a-2734-47b9-9f14-7eea8d9ad3c4-inventory\") pod \"b9e8242a-2734-47b9-9f14-7eea8d9ad3c4\" (UID: \"b9e8242a-2734-47b9-9f14-7eea8d9ad3c4\") " Dec 07 16:35:07 crc kubenswrapper[4716]: I1207 16:35:07.611444 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9e8242a-2734-47b9-9f14-7eea8d9ad3c4-ovn-combined-ca-bundle\") pod \"b9e8242a-2734-47b9-9f14-7eea8d9ad3c4\" (UID: \"b9e8242a-2734-47b9-9f14-7eea8d9ad3c4\") " Dec 07 16:35:07 crc kubenswrapper[4716]: I1207 16:35:07.618932 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9e8242a-2734-47b9-9f14-7eea8d9ad3c4-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "b9e8242a-2734-47b9-9f14-7eea8d9ad3c4" (UID: "b9e8242a-2734-47b9-9f14-7eea8d9ad3c4"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:35:07 crc kubenswrapper[4716]: I1207 16:35:07.619124 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9e8242a-2734-47b9-9f14-7eea8d9ad3c4-kube-api-access-hp26k" (OuterVolumeSpecName: "kube-api-access-hp26k") pod "b9e8242a-2734-47b9-9f14-7eea8d9ad3c4" (UID: "b9e8242a-2734-47b9-9f14-7eea8d9ad3c4"). InnerVolumeSpecName "kube-api-access-hp26k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:35:07 crc kubenswrapper[4716]: I1207 16:35:07.639448 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9e8242a-2734-47b9-9f14-7eea8d9ad3c4-inventory" (OuterVolumeSpecName: "inventory") pod "b9e8242a-2734-47b9-9f14-7eea8d9ad3c4" (UID: "b9e8242a-2734-47b9-9f14-7eea8d9ad3c4"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:35:07 crc kubenswrapper[4716]: I1207 16:35:07.641720 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9e8242a-2734-47b9-9f14-7eea8d9ad3c4-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b9e8242a-2734-47b9-9f14-7eea8d9ad3c4" (UID: "b9e8242a-2734-47b9-9f14-7eea8d9ad3c4"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:35:07 crc kubenswrapper[4716]: I1207 16:35:07.646791 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b9e8242a-2734-47b9-9f14-7eea8d9ad3c4-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "b9e8242a-2734-47b9-9f14-7eea8d9ad3c4" (UID: "b9e8242a-2734-47b9-9f14-7eea8d9ad3c4"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:35:07 crc kubenswrapper[4716]: I1207 16:35:07.714137 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hp26k\" (UniqueName: \"kubernetes.io/projected/b9e8242a-2734-47b9-9f14-7eea8d9ad3c4-kube-api-access-hp26k\") on node \"crc\" DevicePath \"\"" Dec 07 16:35:07 crc kubenswrapper[4716]: I1207 16:35:07.714173 4716 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b9e8242a-2734-47b9-9f14-7eea8d9ad3c4-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 07 16:35:07 crc kubenswrapper[4716]: I1207 16:35:07.714184 4716 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/b9e8242a-2734-47b9-9f14-7eea8d9ad3c4-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Dec 07 16:35:07 crc kubenswrapper[4716]: I1207 16:35:07.714192 4716 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b9e8242a-2734-47b9-9f14-7eea8d9ad3c4-inventory\") on node \"crc\" DevicePath \"\"" Dec 07 16:35:07 crc kubenswrapper[4716]: I1207 16:35:07.714201 4716 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9e8242a-2734-47b9-9f14-7eea8d9ad3c4-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 16:35:07 crc kubenswrapper[4716]: I1207 16:35:07.960178 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-26hn5" event={"ID":"b9e8242a-2734-47b9-9f14-7eea8d9ad3c4","Type":"ContainerDied","Data":"37ae1d595e95ce1f1869a877cfb7f5e9c61561c9fe95d2f9e326cad01b2221fe"} Dec 07 16:35:07 crc kubenswrapper[4716]: I1207 16:35:07.960254 4716 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="37ae1d595e95ce1f1869a877cfb7f5e9c61561c9fe95d2f9e326cad01b2221fe" Dec 07 16:35:07 crc kubenswrapper[4716]: I1207 16:35:07.960276 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-26hn5" Dec 07 16:35:08 crc kubenswrapper[4716]: I1207 16:35:08.055315 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-n265m"] Dec 07 16:35:08 crc kubenswrapper[4716]: E1207 16:35:08.055790 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9e8242a-2734-47b9-9f14-7eea8d9ad3c4" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 07 16:35:08 crc kubenswrapper[4716]: I1207 16:35:08.055810 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9e8242a-2734-47b9-9f14-7eea8d9ad3c4" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 07 16:35:08 crc kubenswrapper[4716]: I1207 16:35:08.056006 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9e8242a-2734-47b9-9f14-7eea8d9ad3c4" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 07 16:35:08 crc kubenswrapper[4716]: I1207 16:35:08.056816 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-n265m" Dec 07 16:35:08 crc kubenswrapper[4716]: I1207 16:35:08.059616 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Dec 07 16:35:08 crc kubenswrapper[4716]: I1207 16:35:08.059625 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Dec 07 16:35:08 crc kubenswrapper[4716]: I1207 16:35:08.059855 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 07 16:35:08 crc kubenswrapper[4716]: I1207 16:35:08.059930 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 07 16:35:08 crc kubenswrapper[4716]: I1207 16:35:08.061542 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 07 16:35:08 crc kubenswrapper[4716]: I1207 16:35:08.061725 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-h9pt2" Dec 07 16:35:08 crc kubenswrapper[4716]: I1207 16:35:08.076330 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-n265m"] Dec 07 16:35:08 crc kubenswrapper[4716]: I1207 16:35:08.224685 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/b4f56131-2845-43e1-84ca-52db74bf2b08-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-n265m\" (UID: \"b4f56131-2845-43e1-84ca-52db74bf2b08\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-n265m" Dec 07 16:35:08 crc kubenswrapper[4716]: I1207 16:35:08.224750 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/b4f56131-2845-43e1-84ca-52db74bf2b08-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-n265m\" (UID: \"b4f56131-2845-43e1-84ca-52db74bf2b08\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-n265m" Dec 07 16:35:08 crc kubenswrapper[4716]: I1207 16:35:08.224799 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qqgz8\" (UniqueName: \"kubernetes.io/projected/b4f56131-2845-43e1-84ca-52db74bf2b08-kube-api-access-qqgz8\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-n265m\" (UID: \"b4f56131-2845-43e1-84ca-52db74bf2b08\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-n265m" Dec 07 16:35:08 crc kubenswrapper[4716]: I1207 16:35:08.224999 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b4f56131-2845-43e1-84ca-52db74bf2b08-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-n265m\" (UID: \"b4f56131-2845-43e1-84ca-52db74bf2b08\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-n265m" Dec 07 16:35:08 crc kubenswrapper[4716]: I1207 16:35:08.225034 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b4f56131-2845-43e1-84ca-52db74bf2b08-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-n265m\" (UID: \"b4f56131-2845-43e1-84ca-52db74bf2b08\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-n265m" Dec 07 16:35:08 crc kubenswrapper[4716]: I1207 16:35:08.225109 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4f56131-2845-43e1-84ca-52db74bf2b08-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-n265m\" (UID: \"b4f56131-2845-43e1-84ca-52db74bf2b08\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-n265m" Dec 07 16:35:08 crc kubenswrapper[4716]: I1207 16:35:08.326218 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4f56131-2845-43e1-84ca-52db74bf2b08-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-n265m\" (UID: \"b4f56131-2845-43e1-84ca-52db74bf2b08\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-n265m" Dec 07 16:35:08 crc kubenswrapper[4716]: I1207 16:35:08.326329 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/b4f56131-2845-43e1-84ca-52db74bf2b08-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-n265m\" (UID: \"b4f56131-2845-43e1-84ca-52db74bf2b08\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-n265m" Dec 07 16:35:08 crc kubenswrapper[4716]: I1207 16:35:08.326371 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/b4f56131-2845-43e1-84ca-52db74bf2b08-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-n265m\" (UID: \"b4f56131-2845-43e1-84ca-52db74bf2b08\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-n265m" Dec 07 16:35:08 crc kubenswrapper[4716]: I1207 16:35:08.326474 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qqgz8\" (UniqueName: \"kubernetes.io/projected/b4f56131-2845-43e1-84ca-52db74bf2b08-kube-api-access-qqgz8\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-n265m\" (UID: \"b4f56131-2845-43e1-84ca-52db74bf2b08\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-n265m" Dec 07 16:35:08 crc kubenswrapper[4716]: I1207 16:35:08.326496 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b4f56131-2845-43e1-84ca-52db74bf2b08-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-n265m\" (UID: \"b4f56131-2845-43e1-84ca-52db74bf2b08\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-n265m" Dec 07 16:35:08 crc kubenswrapper[4716]: I1207 16:35:08.326526 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b4f56131-2845-43e1-84ca-52db74bf2b08-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-n265m\" (UID: \"b4f56131-2845-43e1-84ca-52db74bf2b08\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-n265m" Dec 07 16:35:08 crc kubenswrapper[4716]: I1207 16:35:08.330167 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b4f56131-2845-43e1-84ca-52db74bf2b08-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-n265m\" (UID: \"b4f56131-2845-43e1-84ca-52db74bf2b08\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-n265m" Dec 07 16:35:08 crc kubenswrapper[4716]: I1207 16:35:08.330222 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/b4f56131-2845-43e1-84ca-52db74bf2b08-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-n265m\" (UID: \"b4f56131-2845-43e1-84ca-52db74bf2b08\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-n265m" Dec 07 16:35:08 crc kubenswrapper[4716]: I1207 16:35:08.330490 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b4f56131-2845-43e1-84ca-52db74bf2b08-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-n265m\" (UID: \"b4f56131-2845-43e1-84ca-52db74bf2b08\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-n265m" Dec 07 16:35:08 crc kubenswrapper[4716]: I1207 16:35:08.331204 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4f56131-2845-43e1-84ca-52db74bf2b08-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-n265m\" (UID: \"b4f56131-2845-43e1-84ca-52db74bf2b08\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-n265m" Dec 07 16:35:08 crc kubenswrapper[4716]: I1207 16:35:08.333128 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/b4f56131-2845-43e1-84ca-52db74bf2b08-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-n265m\" (UID: \"b4f56131-2845-43e1-84ca-52db74bf2b08\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-n265m" Dec 07 16:35:08 crc kubenswrapper[4716]: I1207 16:35:08.350346 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qqgz8\" (UniqueName: \"kubernetes.io/projected/b4f56131-2845-43e1-84ca-52db74bf2b08-kube-api-access-qqgz8\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-n265m\" (UID: \"b4f56131-2845-43e1-84ca-52db74bf2b08\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-n265m" Dec 07 16:35:08 crc kubenswrapper[4716]: I1207 16:35:08.382044 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-n265m" Dec 07 16:35:08 crc kubenswrapper[4716]: I1207 16:35:08.886050 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-n265m"] Dec 07 16:35:08 crc kubenswrapper[4716]: I1207 16:35:08.891379 4716 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 07 16:35:08 crc kubenswrapper[4716]: I1207 16:35:08.971025 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-n265m" event={"ID":"b4f56131-2845-43e1-84ca-52db74bf2b08","Type":"ContainerStarted","Data":"b00c2b849971a3595062146cfaf478071771ff1552d854121dd31a297648924b"} Dec 07 16:35:10 crc kubenswrapper[4716]: I1207 16:35:10.986988 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-n265m" event={"ID":"b4f56131-2845-43e1-84ca-52db74bf2b08","Type":"ContainerStarted","Data":"b9b6a789686ff873ac62ef9a7a4bfa1df26f0cf160d85281f04add0647e28cd2"} Dec 07 16:35:11 crc kubenswrapper[4716]: I1207 16:35:11.006010 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-n265m" podStartSLOduration=1.42446135 podStartE2EDuration="3.005990844s" podCreationTimestamp="2025-12-07 16:35:08 +0000 UTC" firstStartedPulling="2025-12-07 16:35:08.891099728 +0000 UTC m=+1971.581384640" lastFinishedPulling="2025-12-07 16:35:10.472629222 +0000 UTC m=+1973.162914134" observedRunningTime="2025-12-07 16:35:10.999888679 +0000 UTC m=+1973.690173591" watchObservedRunningTime="2025-12-07 16:35:11.005990844 +0000 UTC m=+1973.696275756" Dec 07 16:35:12 crc kubenswrapper[4716]: I1207 16:35:12.658156 4716 scope.go:117] "RemoveContainer" containerID="bae1a3e9d80f5b00293cade8b7601799399c2883ae101574459a4c01e523a154" Dec 07 16:35:12 crc kubenswrapper[4716]: E1207 16:35:12.658748 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 16:35:24 crc kubenswrapper[4716]: I1207 16:35:24.657984 4716 scope.go:117] "RemoveContainer" containerID="bae1a3e9d80f5b00293cade8b7601799399c2883ae101574459a4c01e523a154" Dec 07 16:35:25 crc kubenswrapper[4716]: I1207 16:35:25.103487 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" event={"ID":"c15b59eb-565d-4556-a4ce-75afdf159dc8","Type":"ContainerStarted","Data":"8aa74b5db626ce69c5f618707bfc1aab2b0242d5c7b440dfc5c0166d72b0d08c"} Dec 07 16:36:07 crc kubenswrapper[4716]: I1207 16:36:07.530581 4716 generic.go:334] "Generic (PLEG): container finished" podID="b4f56131-2845-43e1-84ca-52db74bf2b08" containerID="b9b6a789686ff873ac62ef9a7a4bfa1df26f0cf160d85281f04add0647e28cd2" exitCode=0 Dec 07 16:36:07 crc kubenswrapper[4716]: I1207 16:36:07.531009 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-n265m" event={"ID":"b4f56131-2845-43e1-84ca-52db74bf2b08","Type":"ContainerDied","Data":"b9b6a789686ff873ac62ef9a7a4bfa1df26f0cf160d85281f04add0647e28cd2"} Dec 07 16:36:08 crc kubenswrapper[4716]: I1207 16:36:08.942872 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-n265m" Dec 07 16:36:08 crc kubenswrapper[4716]: I1207 16:36:08.995551 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/b4f56131-2845-43e1-84ca-52db74bf2b08-neutron-ovn-metadata-agent-neutron-config-0\") pod \"b4f56131-2845-43e1-84ca-52db74bf2b08\" (UID: \"b4f56131-2845-43e1-84ca-52db74bf2b08\") " Dec 07 16:36:08 crc kubenswrapper[4716]: I1207 16:36:08.995628 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4f56131-2845-43e1-84ca-52db74bf2b08-neutron-metadata-combined-ca-bundle\") pod \"b4f56131-2845-43e1-84ca-52db74bf2b08\" (UID: \"b4f56131-2845-43e1-84ca-52db74bf2b08\") " Dec 07 16:36:08 crc kubenswrapper[4716]: I1207 16:36:08.995670 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/b4f56131-2845-43e1-84ca-52db74bf2b08-nova-metadata-neutron-config-0\") pod \"b4f56131-2845-43e1-84ca-52db74bf2b08\" (UID: \"b4f56131-2845-43e1-84ca-52db74bf2b08\") " Dec 07 16:36:08 crc kubenswrapper[4716]: I1207 16:36:08.995717 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qqgz8\" (UniqueName: \"kubernetes.io/projected/b4f56131-2845-43e1-84ca-52db74bf2b08-kube-api-access-qqgz8\") pod \"b4f56131-2845-43e1-84ca-52db74bf2b08\" (UID: \"b4f56131-2845-43e1-84ca-52db74bf2b08\") " Dec 07 16:36:08 crc kubenswrapper[4716]: I1207 16:36:08.995784 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b4f56131-2845-43e1-84ca-52db74bf2b08-inventory\") pod \"b4f56131-2845-43e1-84ca-52db74bf2b08\" (UID: \"b4f56131-2845-43e1-84ca-52db74bf2b08\") " Dec 07 16:36:08 crc kubenswrapper[4716]: I1207 16:36:08.995867 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b4f56131-2845-43e1-84ca-52db74bf2b08-ssh-key\") pod \"b4f56131-2845-43e1-84ca-52db74bf2b08\" (UID: \"b4f56131-2845-43e1-84ca-52db74bf2b08\") " Dec 07 16:36:09 crc kubenswrapper[4716]: I1207 16:36:09.002292 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4f56131-2845-43e1-84ca-52db74bf2b08-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "b4f56131-2845-43e1-84ca-52db74bf2b08" (UID: "b4f56131-2845-43e1-84ca-52db74bf2b08"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:36:09 crc kubenswrapper[4716]: I1207 16:36:09.005504 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4f56131-2845-43e1-84ca-52db74bf2b08-kube-api-access-qqgz8" (OuterVolumeSpecName: "kube-api-access-qqgz8") pod "b4f56131-2845-43e1-84ca-52db74bf2b08" (UID: "b4f56131-2845-43e1-84ca-52db74bf2b08"). InnerVolumeSpecName "kube-api-access-qqgz8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:36:09 crc kubenswrapper[4716]: I1207 16:36:09.024428 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4f56131-2845-43e1-84ca-52db74bf2b08-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b4f56131-2845-43e1-84ca-52db74bf2b08" (UID: "b4f56131-2845-43e1-84ca-52db74bf2b08"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:36:09 crc kubenswrapper[4716]: I1207 16:36:09.026525 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4f56131-2845-43e1-84ca-52db74bf2b08-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "b4f56131-2845-43e1-84ca-52db74bf2b08" (UID: "b4f56131-2845-43e1-84ca-52db74bf2b08"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:36:09 crc kubenswrapper[4716]: I1207 16:36:09.032291 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4f56131-2845-43e1-84ca-52db74bf2b08-inventory" (OuterVolumeSpecName: "inventory") pod "b4f56131-2845-43e1-84ca-52db74bf2b08" (UID: "b4f56131-2845-43e1-84ca-52db74bf2b08"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:36:09 crc kubenswrapper[4716]: I1207 16:36:09.035434 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4f56131-2845-43e1-84ca-52db74bf2b08-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "b4f56131-2845-43e1-84ca-52db74bf2b08" (UID: "b4f56131-2845-43e1-84ca-52db74bf2b08"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:36:09 crc kubenswrapper[4716]: I1207 16:36:09.098372 4716 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b4f56131-2845-43e1-84ca-52db74bf2b08-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 07 16:36:09 crc kubenswrapper[4716]: I1207 16:36:09.098407 4716 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/b4f56131-2845-43e1-84ca-52db74bf2b08-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 07 16:36:09 crc kubenswrapper[4716]: I1207 16:36:09.098422 4716 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4f56131-2845-43e1-84ca-52db74bf2b08-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 16:36:09 crc kubenswrapper[4716]: I1207 16:36:09.098434 4716 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/b4f56131-2845-43e1-84ca-52db74bf2b08-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 07 16:36:09 crc kubenswrapper[4716]: I1207 16:36:09.098444 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qqgz8\" (UniqueName: \"kubernetes.io/projected/b4f56131-2845-43e1-84ca-52db74bf2b08-kube-api-access-qqgz8\") on node \"crc\" DevicePath \"\"" Dec 07 16:36:09 crc kubenswrapper[4716]: I1207 16:36:09.098453 4716 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b4f56131-2845-43e1-84ca-52db74bf2b08-inventory\") on node \"crc\" DevicePath \"\"" Dec 07 16:36:09 crc kubenswrapper[4716]: I1207 16:36:09.548526 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-n265m" event={"ID":"b4f56131-2845-43e1-84ca-52db74bf2b08","Type":"ContainerDied","Data":"b00c2b849971a3595062146cfaf478071771ff1552d854121dd31a297648924b"} Dec 07 16:36:09 crc kubenswrapper[4716]: I1207 16:36:09.548975 4716 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b00c2b849971a3595062146cfaf478071771ff1552d854121dd31a297648924b" Dec 07 16:36:09 crc kubenswrapper[4716]: I1207 16:36:09.548572 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-n265m" Dec 07 16:36:09 crc kubenswrapper[4716]: I1207 16:36:09.643751 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-66mjc"] Dec 07 16:36:09 crc kubenswrapper[4716]: E1207 16:36:09.644207 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4f56131-2845-43e1-84ca-52db74bf2b08" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 07 16:36:09 crc kubenswrapper[4716]: I1207 16:36:09.644227 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4f56131-2845-43e1-84ca-52db74bf2b08" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 07 16:36:09 crc kubenswrapper[4716]: I1207 16:36:09.644421 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4f56131-2845-43e1-84ca-52db74bf2b08" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 07 16:36:09 crc kubenswrapper[4716]: I1207 16:36:09.645220 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-66mjc" Dec 07 16:36:09 crc kubenswrapper[4716]: I1207 16:36:09.646714 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-h9pt2" Dec 07 16:36:09 crc kubenswrapper[4716]: I1207 16:36:09.646945 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 07 16:36:09 crc kubenswrapper[4716]: I1207 16:36:09.646993 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 07 16:36:09 crc kubenswrapper[4716]: I1207 16:36:09.647917 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Dec 07 16:36:09 crc kubenswrapper[4716]: I1207 16:36:09.647934 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 07 16:36:09 crc kubenswrapper[4716]: I1207 16:36:09.653800 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-66mjc"] Dec 07 16:36:09 crc kubenswrapper[4716]: I1207 16:36:09.709888 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-66mjc\" (UID: \"0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-66mjc" Dec 07 16:36:09 crc kubenswrapper[4716]: I1207 16:36:09.710054 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-66mjc\" (UID: \"0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-66mjc" Dec 07 16:36:09 crc kubenswrapper[4716]: I1207 16:36:09.710096 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-66mjc\" (UID: \"0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-66mjc" Dec 07 16:36:09 crc kubenswrapper[4716]: I1207 16:36:09.710158 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k6dgs\" (UniqueName: \"kubernetes.io/projected/0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3-kube-api-access-k6dgs\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-66mjc\" (UID: \"0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-66mjc" Dec 07 16:36:09 crc kubenswrapper[4716]: I1207 16:36:09.710217 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-66mjc\" (UID: \"0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-66mjc" Dec 07 16:36:09 crc kubenswrapper[4716]: I1207 16:36:09.811214 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-66mjc\" (UID: \"0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-66mjc" Dec 07 16:36:09 crc kubenswrapper[4716]: I1207 16:36:09.811305 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-66mjc\" (UID: \"0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-66mjc" Dec 07 16:36:09 crc kubenswrapper[4716]: I1207 16:36:09.811331 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-66mjc\" (UID: \"0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-66mjc" Dec 07 16:36:09 crc kubenswrapper[4716]: I1207 16:36:09.811370 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k6dgs\" (UniqueName: \"kubernetes.io/projected/0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3-kube-api-access-k6dgs\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-66mjc\" (UID: \"0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-66mjc" Dec 07 16:36:09 crc kubenswrapper[4716]: I1207 16:36:09.811410 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-66mjc\" (UID: \"0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-66mjc" Dec 07 16:36:09 crc kubenswrapper[4716]: I1207 16:36:09.816145 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-66mjc\" (UID: \"0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-66mjc" Dec 07 16:36:09 crc kubenswrapper[4716]: I1207 16:36:09.817322 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-66mjc\" (UID: \"0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-66mjc" Dec 07 16:36:09 crc kubenswrapper[4716]: I1207 16:36:09.817796 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-66mjc\" (UID: \"0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-66mjc" Dec 07 16:36:09 crc kubenswrapper[4716]: I1207 16:36:09.823676 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-66mjc\" (UID: \"0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-66mjc" Dec 07 16:36:09 crc kubenswrapper[4716]: I1207 16:36:09.827009 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k6dgs\" (UniqueName: \"kubernetes.io/projected/0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3-kube-api-access-k6dgs\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-66mjc\" (UID: \"0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-66mjc" Dec 07 16:36:10 crc kubenswrapper[4716]: I1207 16:36:10.046671 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-66mjc" Dec 07 16:36:10 crc kubenswrapper[4716]: I1207 16:36:10.541833 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-66mjc"] Dec 07 16:36:10 crc kubenswrapper[4716]: I1207 16:36:10.559702 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-66mjc" event={"ID":"0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3","Type":"ContainerStarted","Data":"7f5eb2b63832959062133120af8bc4d2abfaddb97ce9609dc90b4fa26d0ff39c"} Dec 07 16:36:16 crc kubenswrapper[4716]: I1207 16:36:16.621841 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-66mjc" event={"ID":"0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3","Type":"ContainerStarted","Data":"4ad48961b4f8dd5fe3b04c46db51ce1d65f9a1ac75529d8fdd3541da9f90e012"} Dec 07 16:36:16 crc kubenswrapper[4716]: I1207 16:36:16.643890 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-66mjc" podStartSLOduration=2.644693071 podStartE2EDuration="7.64387302s" podCreationTimestamp="2025-12-07 16:36:09 +0000 UTC" firstStartedPulling="2025-12-07 16:36:10.546582421 +0000 UTC m=+2033.236867333" lastFinishedPulling="2025-12-07 16:36:15.54576235 +0000 UTC m=+2038.236047282" observedRunningTime="2025-12-07 16:36:16.638627139 +0000 UTC m=+2039.328912071" watchObservedRunningTime="2025-12-07 16:36:16.64387302 +0000 UTC m=+2039.334157932" Dec 07 16:37:28 crc kubenswrapper[4716]: I1207 16:37:28.787138 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-tmf6k"] Dec 07 16:37:28 crc kubenswrapper[4716]: I1207 16:37:28.790230 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tmf6k" Dec 07 16:37:28 crc kubenswrapper[4716]: I1207 16:37:28.815937 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tmf6k"] Dec 07 16:37:28 crc kubenswrapper[4716]: I1207 16:37:28.953993 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e66383d-dee5-4bf6-82a8-cfdbc2c60bfa-utilities\") pod \"community-operators-tmf6k\" (UID: \"9e66383d-dee5-4bf6-82a8-cfdbc2c60bfa\") " pod="openshift-marketplace/community-operators-tmf6k" Dec 07 16:37:28 crc kubenswrapper[4716]: I1207 16:37:28.954217 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e66383d-dee5-4bf6-82a8-cfdbc2c60bfa-catalog-content\") pod \"community-operators-tmf6k\" (UID: \"9e66383d-dee5-4bf6-82a8-cfdbc2c60bfa\") " pod="openshift-marketplace/community-operators-tmf6k" Dec 07 16:37:28 crc kubenswrapper[4716]: I1207 16:37:28.954260 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d467s\" (UniqueName: \"kubernetes.io/projected/9e66383d-dee5-4bf6-82a8-cfdbc2c60bfa-kube-api-access-d467s\") pod \"community-operators-tmf6k\" (UID: \"9e66383d-dee5-4bf6-82a8-cfdbc2c60bfa\") " pod="openshift-marketplace/community-operators-tmf6k" Dec 07 16:37:29 crc kubenswrapper[4716]: I1207 16:37:29.055984 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d467s\" (UniqueName: \"kubernetes.io/projected/9e66383d-dee5-4bf6-82a8-cfdbc2c60bfa-kube-api-access-d467s\") pod \"community-operators-tmf6k\" (UID: \"9e66383d-dee5-4bf6-82a8-cfdbc2c60bfa\") " pod="openshift-marketplace/community-operators-tmf6k" Dec 07 16:37:29 crc kubenswrapper[4716]: I1207 16:37:29.056471 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e66383d-dee5-4bf6-82a8-cfdbc2c60bfa-utilities\") pod \"community-operators-tmf6k\" (UID: \"9e66383d-dee5-4bf6-82a8-cfdbc2c60bfa\") " pod="openshift-marketplace/community-operators-tmf6k" Dec 07 16:37:29 crc kubenswrapper[4716]: I1207 16:37:29.056706 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e66383d-dee5-4bf6-82a8-cfdbc2c60bfa-catalog-content\") pod \"community-operators-tmf6k\" (UID: \"9e66383d-dee5-4bf6-82a8-cfdbc2c60bfa\") " pod="openshift-marketplace/community-operators-tmf6k" Dec 07 16:37:29 crc kubenswrapper[4716]: I1207 16:37:29.057668 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e66383d-dee5-4bf6-82a8-cfdbc2c60bfa-utilities\") pod \"community-operators-tmf6k\" (UID: \"9e66383d-dee5-4bf6-82a8-cfdbc2c60bfa\") " pod="openshift-marketplace/community-operators-tmf6k" Dec 07 16:37:29 crc kubenswrapper[4716]: I1207 16:37:29.057773 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e66383d-dee5-4bf6-82a8-cfdbc2c60bfa-catalog-content\") pod \"community-operators-tmf6k\" (UID: \"9e66383d-dee5-4bf6-82a8-cfdbc2c60bfa\") " pod="openshift-marketplace/community-operators-tmf6k" Dec 07 16:37:29 crc kubenswrapper[4716]: I1207 16:37:29.076695 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d467s\" (UniqueName: \"kubernetes.io/projected/9e66383d-dee5-4bf6-82a8-cfdbc2c60bfa-kube-api-access-d467s\") pod \"community-operators-tmf6k\" (UID: \"9e66383d-dee5-4bf6-82a8-cfdbc2c60bfa\") " pod="openshift-marketplace/community-operators-tmf6k" Dec 07 16:37:29 crc kubenswrapper[4716]: I1207 16:37:29.120894 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tmf6k" Dec 07 16:37:29 crc kubenswrapper[4716]: I1207 16:37:29.676553 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tmf6k"] Dec 07 16:37:30 crc kubenswrapper[4716]: I1207 16:37:30.324181 4716 generic.go:334] "Generic (PLEG): container finished" podID="9e66383d-dee5-4bf6-82a8-cfdbc2c60bfa" containerID="5c6fdd31e4d722f3f1eccfb62b25441b4ebbe0b75c59e0028194c5438d2408e2" exitCode=0 Dec 07 16:37:30 crc kubenswrapper[4716]: I1207 16:37:30.324241 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tmf6k" event={"ID":"9e66383d-dee5-4bf6-82a8-cfdbc2c60bfa","Type":"ContainerDied","Data":"5c6fdd31e4d722f3f1eccfb62b25441b4ebbe0b75c59e0028194c5438d2408e2"} Dec 07 16:37:30 crc kubenswrapper[4716]: I1207 16:37:30.324277 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tmf6k" event={"ID":"9e66383d-dee5-4bf6-82a8-cfdbc2c60bfa","Type":"ContainerStarted","Data":"dea92e25a6e3b4b868700610bf114c0be13e56b7c145a027f762072d2841742d"} Dec 07 16:37:31 crc kubenswrapper[4716]: I1207 16:37:31.334699 4716 generic.go:334] "Generic (PLEG): container finished" podID="9e66383d-dee5-4bf6-82a8-cfdbc2c60bfa" containerID="cbb766b75f6083e414a5eea361306f4719c649b377e3a48f1492268f6f088546" exitCode=0 Dec 07 16:37:31 crc kubenswrapper[4716]: I1207 16:37:31.334778 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tmf6k" event={"ID":"9e66383d-dee5-4bf6-82a8-cfdbc2c60bfa","Type":"ContainerDied","Data":"cbb766b75f6083e414a5eea361306f4719c649b377e3a48f1492268f6f088546"} Dec 07 16:37:32 crc kubenswrapper[4716]: I1207 16:37:32.346591 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tmf6k" event={"ID":"9e66383d-dee5-4bf6-82a8-cfdbc2c60bfa","Type":"ContainerStarted","Data":"be76cc568a0e0d2778597903db62d6ca21f979aac0a6d529f839aa819a899924"} Dec 07 16:37:32 crc kubenswrapper[4716]: I1207 16:37:32.364544 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-tmf6k" podStartSLOduration=2.875665062 podStartE2EDuration="4.364519507s" podCreationTimestamp="2025-12-07 16:37:28 +0000 UTC" firstStartedPulling="2025-12-07 16:37:30.328331523 +0000 UTC m=+2113.018616435" lastFinishedPulling="2025-12-07 16:37:31.817185958 +0000 UTC m=+2114.507470880" observedRunningTime="2025-12-07 16:37:32.362065931 +0000 UTC m=+2115.052350843" watchObservedRunningTime="2025-12-07 16:37:32.364519507 +0000 UTC m=+2115.054804419" Dec 07 16:37:39 crc kubenswrapper[4716]: I1207 16:37:39.122114 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-tmf6k" Dec 07 16:37:39 crc kubenswrapper[4716]: I1207 16:37:39.122807 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-tmf6k" Dec 07 16:37:39 crc kubenswrapper[4716]: I1207 16:37:39.173502 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-tmf6k" Dec 07 16:37:39 crc kubenswrapper[4716]: I1207 16:37:39.473389 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-tmf6k" Dec 07 16:37:42 crc kubenswrapper[4716]: I1207 16:37:42.758520 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tmf6k"] Dec 07 16:37:42 crc kubenswrapper[4716]: I1207 16:37:42.759120 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-tmf6k" podUID="9e66383d-dee5-4bf6-82a8-cfdbc2c60bfa" containerName="registry-server" containerID="cri-o://be76cc568a0e0d2778597903db62d6ca21f979aac0a6d529f839aa819a899924" gracePeriod=2 Dec 07 16:37:43 crc kubenswrapper[4716]: I1207 16:37:43.247784 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tmf6k" Dec 07 16:37:43 crc kubenswrapper[4716]: I1207 16:37:43.346508 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e66383d-dee5-4bf6-82a8-cfdbc2c60bfa-utilities\") pod \"9e66383d-dee5-4bf6-82a8-cfdbc2c60bfa\" (UID: \"9e66383d-dee5-4bf6-82a8-cfdbc2c60bfa\") " Dec 07 16:37:43 crc kubenswrapper[4716]: I1207 16:37:43.346649 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e66383d-dee5-4bf6-82a8-cfdbc2c60bfa-catalog-content\") pod \"9e66383d-dee5-4bf6-82a8-cfdbc2c60bfa\" (UID: \"9e66383d-dee5-4bf6-82a8-cfdbc2c60bfa\") " Dec 07 16:37:43 crc kubenswrapper[4716]: I1207 16:37:43.346786 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d467s\" (UniqueName: \"kubernetes.io/projected/9e66383d-dee5-4bf6-82a8-cfdbc2c60bfa-kube-api-access-d467s\") pod \"9e66383d-dee5-4bf6-82a8-cfdbc2c60bfa\" (UID: \"9e66383d-dee5-4bf6-82a8-cfdbc2c60bfa\") " Dec 07 16:37:43 crc kubenswrapper[4716]: I1207 16:37:43.348132 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9e66383d-dee5-4bf6-82a8-cfdbc2c60bfa-utilities" (OuterVolumeSpecName: "utilities") pod "9e66383d-dee5-4bf6-82a8-cfdbc2c60bfa" (UID: "9e66383d-dee5-4bf6-82a8-cfdbc2c60bfa"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:37:43 crc kubenswrapper[4716]: I1207 16:37:43.353385 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e66383d-dee5-4bf6-82a8-cfdbc2c60bfa-kube-api-access-d467s" (OuterVolumeSpecName: "kube-api-access-d467s") pod "9e66383d-dee5-4bf6-82a8-cfdbc2c60bfa" (UID: "9e66383d-dee5-4bf6-82a8-cfdbc2c60bfa"). InnerVolumeSpecName "kube-api-access-d467s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:37:43 crc kubenswrapper[4716]: I1207 16:37:43.396809 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9e66383d-dee5-4bf6-82a8-cfdbc2c60bfa-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9e66383d-dee5-4bf6-82a8-cfdbc2c60bfa" (UID: "9e66383d-dee5-4bf6-82a8-cfdbc2c60bfa"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:37:43 crc kubenswrapper[4716]: I1207 16:37:43.448512 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d467s\" (UniqueName: \"kubernetes.io/projected/9e66383d-dee5-4bf6-82a8-cfdbc2c60bfa-kube-api-access-d467s\") on node \"crc\" DevicePath \"\"" Dec 07 16:37:43 crc kubenswrapper[4716]: I1207 16:37:43.448561 4716 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e66383d-dee5-4bf6-82a8-cfdbc2c60bfa-utilities\") on node \"crc\" DevicePath \"\"" Dec 07 16:37:43 crc kubenswrapper[4716]: I1207 16:37:43.448571 4716 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e66383d-dee5-4bf6-82a8-cfdbc2c60bfa-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 07 16:37:43 crc kubenswrapper[4716]: I1207 16:37:43.453416 4716 generic.go:334] "Generic (PLEG): container finished" podID="9e66383d-dee5-4bf6-82a8-cfdbc2c60bfa" containerID="be76cc568a0e0d2778597903db62d6ca21f979aac0a6d529f839aa819a899924" exitCode=0 Dec 07 16:37:43 crc kubenswrapper[4716]: I1207 16:37:43.453478 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tmf6k" event={"ID":"9e66383d-dee5-4bf6-82a8-cfdbc2c60bfa","Type":"ContainerDied","Data":"be76cc568a0e0d2778597903db62d6ca21f979aac0a6d529f839aa819a899924"} Dec 07 16:37:43 crc kubenswrapper[4716]: I1207 16:37:43.453518 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tmf6k" Dec 07 16:37:43 crc kubenswrapper[4716]: I1207 16:37:43.453669 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tmf6k" event={"ID":"9e66383d-dee5-4bf6-82a8-cfdbc2c60bfa","Type":"ContainerDied","Data":"dea92e25a6e3b4b868700610bf114c0be13e56b7c145a027f762072d2841742d"} Dec 07 16:37:43 crc kubenswrapper[4716]: I1207 16:37:43.453770 4716 scope.go:117] "RemoveContainer" containerID="be76cc568a0e0d2778597903db62d6ca21f979aac0a6d529f839aa819a899924" Dec 07 16:37:43 crc kubenswrapper[4716]: I1207 16:37:43.476150 4716 scope.go:117] "RemoveContainer" containerID="cbb766b75f6083e414a5eea361306f4719c649b377e3a48f1492268f6f088546" Dec 07 16:37:43 crc kubenswrapper[4716]: I1207 16:37:43.493143 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tmf6k"] Dec 07 16:37:43 crc kubenswrapper[4716]: I1207 16:37:43.500336 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-tmf6k"] Dec 07 16:37:43 crc kubenswrapper[4716]: I1207 16:37:43.520785 4716 scope.go:117] "RemoveContainer" containerID="5c6fdd31e4d722f3f1eccfb62b25441b4ebbe0b75c59e0028194c5438d2408e2" Dec 07 16:37:43 crc kubenswrapper[4716]: I1207 16:37:43.541532 4716 scope.go:117] "RemoveContainer" containerID="be76cc568a0e0d2778597903db62d6ca21f979aac0a6d529f839aa819a899924" Dec 07 16:37:43 crc kubenswrapper[4716]: E1207 16:37:43.541950 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be76cc568a0e0d2778597903db62d6ca21f979aac0a6d529f839aa819a899924\": container with ID starting with be76cc568a0e0d2778597903db62d6ca21f979aac0a6d529f839aa819a899924 not found: ID does not exist" containerID="be76cc568a0e0d2778597903db62d6ca21f979aac0a6d529f839aa819a899924" Dec 07 16:37:43 crc kubenswrapper[4716]: I1207 16:37:43.542009 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be76cc568a0e0d2778597903db62d6ca21f979aac0a6d529f839aa819a899924"} err="failed to get container status \"be76cc568a0e0d2778597903db62d6ca21f979aac0a6d529f839aa819a899924\": rpc error: code = NotFound desc = could not find container \"be76cc568a0e0d2778597903db62d6ca21f979aac0a6d529f839aa819a899924\": container with ID starting with be76cc568a0e0d2778597903db62d6ca21f979aac0a6d529f839aa819a899924 not found: ID does not exist" Dec 07 16:37:43 crc kubenswrapper[4716]: I1207 16:37:43.542048 4716 scope.go:117] "RemoveContainer" containerID="cbb766b75f6083e414a5eea361306f4719c649b377e3a48f1492268f6f088546" Dec 07 16:37:43 crc kubenswrapper[4716]: E1207 16:37:43.542581 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cbb766b75f6083e414a5eea361306f4719c649b377e3a48f1492268f6f088546\": container with ID starting with cbb766b75f6083e414a5eea361306f4719c649b377e3a48f1492268f6f088546 not found: ID does not exist" containerID="cbb766b75f6083e414a5eea361306f4719c649b377e3a48f1492268f6f088546" Dec 07 16:37:43 crc kubenswrapper[4716]: I1207 16:37:43.542606 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cbb766b75f6083e414a5eea361306f4719c649b377e3a48f1492268f6f088546"} err="failed to get container status \"cbb766b75f6083e414a5eea361306f4719c649b377e3a48f1492268f6f088546\": rpc error: code = NotFound desc = could not find container \"cbb766b75f6083e414a5eea361306f4719c649b377e3a48f1492268f6f088546\": container with ID starting with cbb766b75f6083e414a5eea361306f4719c649b377e3a48f1492268f6f088546 not found: ID does not exist" Dec 07 16:37:43 crc kubenswrapper[4716]: I1207 16:37:43.542620 4716 scope.go:117] "RemoveContainer" containerID="5c6fdd31e4d722f3f1eccfb62b25441b4ebbe0b75c59e0028194c5438d2408e2" Dec 07 16:37:43 crc kubenswrapper[4716]: E1207 16:37:43.542935 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c6fdd31e4d722f3f1eccfb62b25441b4ebbe0b75c59e0028194c5438d2408e2\": container with ID starting with 5c6fdd31e4d722f3f1eccfb62b25441b4ebbe0b75c59e0028194c5438d2408e2 not found: ID does not exist" containerID="5c6fdd31e4d722f3f1eccfb62b25441b4ebbe0b75c59e0028194c5438d2408e2" Dec 07 16:37:43 crc kubenswrapper[4716]: I1207 16:37:43.542975 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c6fdd31e4d722f3f1eccfb62b25441b4ebbe0b75c59e0028194c5438d2408e2"} err="failed to get container status \"5c6fdd31e4d722f3f1eccfb62b25441b4ebbe0b75c59e0028194c5438d2408e2\": rpc error: code = NotFound desc = could not find container \"5c6fdd31e4d722f3f1eccfb62b25441b4ebbe0b75c59e0028194c5438d2408e2\": container with ID starting with 5c6fdd31e4d722f3f1eccfb62b25441b4ebbe0b75c59e0028194c5438d2408e2 not found: ID does not exist" Dec 07 16:37:43 crc kubenswrapper[4716]: I1207 16:37:43.670909 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e66383d-dee5-4bf6-82a8-cfdbc2c60bfa" path="/var/lib/kubelet/pods/9e66383d-dee5-4bf6-82a8-cfdbc2c60bfa/volumes" Dec 07 16:37:46 crc kubenswrapper[4716]: I1207 16:37:46.768641 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-pnjtg"] Dec 07 16:37:46 crc kubenswrapper[4716]: E1207 16:37:46.769945 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e66383d-dee5-4bf6-82a8-cfdbc2c60bfa" containerName="extract-utilities" Dec 07 16:37:46 crc kubenswrapper[4716]: I1207 16:37:46.769966 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e66383d-dee5-4bf6-82a8-cfdbc2c60bfa" containerName="extract-utilities" Dec 07 16:37:46 crc kubenswrapper[4716]: E1207 16:37:46.769996 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e66383d-dee5-4bf6-82a8-cfdbc2c60bfa" containerName="registry-server" Dec 07 16:37:46 crc kubenswrapper[4716]: I1207 16:37:46.770008 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e66383d-dee5-4bf6-82a8-cfdbc2c60bfa" containerName="registry-server" Dec 07 16:37:46 crc kubenswrapper[4716]: E1207 16:37:46.770028 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e66383d-dee5-4bf6-82a8-cfdbc2c60bfa" containerName="extract-content" Dec 07 16:37:46 crc kubenswrapper[4716]: I1207 16:37:46.770038 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e66383d-dee5-4bf6-82a8-cfdbc2c60bfa" containerName="extract-content" Dec 07 16:37:46 crc kubenswrapper[4716]: I1207 16:37:46.770359 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e66383d-dee5-4bf6-82a8-cfdbc2c60bfa" containerName="registry-server" Dec 07 16:37:46 crc kubenswrapper[4716]: I1207 16:37:46.772604 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pnjtg" Dec 07 16:37:46 crc kubenswrapper[4716]: I1207 16:37:46.784841 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pnjtg"] Dec 07 16:37:46 crc kubenswrapper[4716]: I1207 16:37:46.915209 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c0ad3b2-97b0-4c37-85da-ab61e86ef418-catalog-content\") pod \"certified-operators-pnjtg\" (UID: \"2c0ad3b2-97b0-4c37-85da-ab61e86ef418\") " pod="openshift-marketplace/certified-operators-pnjtg" Dec 07 16:37:46 crc kubenswrapper[4716]: I1207 16:37:46.915286 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c0ad3b2-97b0-4c37-85da-ab61e86ef418-utilities\") pod \"certified-operators-pnjtg\" (UID: \"2c0ad3b2-97b0-4c37-85da-ab61e86ef418\") " pod="openshift-marketplace/certified-operators-pnjtg" Dec 07 16:37:46 crc kubenswrapper[4716]: I1207 16:37:46.915375 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwcwz\" (UniqueName: \"kubernetes.io/projected/2c0ad3b2-97b0-4c37-85da-ab61e86ef418-kube-api-access-fwcwz\") pod \"certified-operators-pnjtg\" (UID: \"2c0ad3b2-97b0-4c37-85da-ab61e86ef418\") " pod="openshift-marketplace/certified-operators-pnjtg" Dec 07 16:37:47 crc kubenswrapper[4716]: I1207 16:37:47.016998 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwcwz\" (UniqueName: \"kubernetes.io/projected/2c0ad3b2-97b0-4c37-85da-ab61e86ef418-kube-api-access-fwcwz\") pod \"certified-operators-pnjtg\" (UID: \"2c0ad3b2-97b0-4c37-85da-ab61e86ef418\") " pod="openshift-marketplace/certified-operators-pnjtg" Dec 07 16:37:47 crc kubenswrapper[4716]: I1207 16:37:47.017117 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c0ad3b2-97b0-4c37-85da-ab61e86ef418-catalog-content\") pod \"certified-operators-pnjtg\" (UID: \"2c0ad3b2-97b0-4c37-85da-ab61e86ef418\") " pod="openshift-marketplace/certified-operators-pnjtg" Dec 07 16:37:47 crc kubenswrapper[4716]: I1207 16:37:47.017176 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c0ad3b2-97b0-4c37-85da-ab61e86ef418-utilities\") pod \"certified-operators-pnjtg\" (UID: \"2c0ad3b2-97b0-4c37-85da-ab61e86ef418\") " pod="openshift-marketplace/certified-operators-pnjtg" Dec 07 16:37:47 crc kubenswrapper[4716]: I1207 16:37:47.017735 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c0ad3b2-97b0-4c37-85da-ab61e86ef418-utilities\") pod \"certified-operators-pnjtg\" (UID: \"2c0ad3b2-97b0-4c37-85da-ab61e86ef418\") " pod="openshift-marketplace/certified-operators-pnjtg" Dec 07 16:37:47 crc kubenswrapper[4716]: I1207 16:37:47.017864 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c0ad3b2-97b0-4c37-85da-ab61e86ef418-catalog-content\") pod \"certified-operators-pnjtg\" (UID: \"2c0ad3b2-97b0-4c37-85da-ab61e86ef418\") " pod="openshift-marketplace/certified-operators-pnjtg" Dec 07 16:37:47 crc kubenswrapper[4716]: I1207 16:37:47.039841 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwcwz\" (UniqueName: \"kubernetes.io/projected/2c0ad3b2-97b0-4c37-85da-ab61e86ef418-kube-api-access-fwcwz\") pod \"certified-operators-pnjtg\" (UID: \"2c0ad3b2-97b0-4c37-85da-ab61e86ef418\") " pod="openshift-marketplace/certified-operators-pnjtg" Dec 07 16:37:47 crc kubenswrapper[4716]: I1207 16:37:47.100507 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pnjtg" Dec 07 16:37:47 crc kubenswrapper[4716]: I1207 16:37:47.589999 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pnjtg"] Dec 07 16:37:48 crc kubenswrapper[4716]: I1207 16:37:48.507423 4716 generic.go:334] "Generic (PLEG): container finished" podID="2c0ad3b2-97b0-4c37-85da-ab61e86ef418" containerID="f7c9f5749af69e1bebeaf8993dac87f8ba53edf646812ee6ef6daa3786b16704" exitCode=0 Dec 07 16:37:48 crc kubenswrapper[4716]: I1207 16:37:48.507636 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pnjtg" event={"ID":"2c0ad3b2-97b0-4c37-85da-ab61e86ef418","Type":"ContainerDied","Data":"f7c9f5749af69e1bebeaf8993dac87f8ba53edf646812ee6ef6daa3786b16704"} Dec 07 16:37:48 crc kubenswrapper[4716]: I1207 16:37:48.507757 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pnjtg" event={"ID":"2c0ad3b2-97b0-4c37-85da-ab61e86ef418","Type":"ContainerStarted","Data":"712e66fdf5206f0fe77121a10baa2a125c9c3090b90e8daa736278112d1889e4"} Dec 07 16:37:50 crc kubenswrapper[4716]: I1207 16:37:50.523669 4716 generic.go:334] "Generic (PLEG): container finished" podID="2c0ad3b2-97b0-4c37-85da-ab61e86ef418" containerID="24cf28fe02990feccec64f3129bc206ea89e3ca05b8aafdc5f3f11d12e34d3ce" exitCode=0 Dec 07 16:37:50 crc kubenswrapper[4716]: I1207 16:37:50.523876 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pnjtg" event={"ID":"2c0ad3b2-97b0-4c37-85da-ab61e86ef418","Type":"ContainerDied","Data":"24cf28fe02990feccec64f3129bc206ea89e3ca05b8aafdc5f3f11d12e34d3ce"} Dec 07 16:37:51 crc kubenswrapper[4716]: I1207 16:37:51.556473 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pnjtg" event={"ID":"2c0ad3b2-97b0-4c37-85da-ab61e86ef418","Type":"ContainerStarted","Data":"d925ae6ace86e5385cea11dfab9747ab35a07fbedd1b95fcfb55cf7e8c975310"} Dec 07 16:37:51 crc kubenswrapper[4716]: I1207 16:37:51.580117 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-pnjtg" podStartSLOduration=3.143570448 podStartE2EDuration="5.580097833s" podCreationTimestamp="2025-12-07 16:37:46 +0000 UTC" firstStartedPulling="2025-12-07 16:37:48.509353839 +0000 UTC m=+2131.199638751" lastFinishedPulling="2025-12-07 16:37:50.945881234 +0000 UTC m=+2133.636166136" observedRunningTime="2025-12-07 16:37:51.578810348 +0000 UTC m=+2134.269095270" watchObservedRunningTime="2025-12-07 16:37:51.580097833 +0000 UTC m=+2134.270382745" Dec 07 16:37:52 crc kubenswrapper[4716]: I1207 16:37:52.760925 4716 patch_prober.go:28] interesting pod/machine-config-daemon-zcxxp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 16:37:52 crc kubenswrapper[4716]: I1207 16:37:52.760983 4716 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 16:37:54 crc kubenswrapper[4716]: I1207 16:37:54.162832 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-bdmpf"] Dec 07 16:37:54 crc kubenswrapper[4716]: I1207 16:37:54.164787 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bdmpf" Dec 07 16:37:54 crc kubenswrapper[4716]: I1207 16:37:54.218868 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bdmpf"] Dec 07 16:37:54 crc kubenswrapper[4716]: I1207 16:37:54.246852 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thz24\" (UniqueName: \"kubernetes.io/projected/03951083-dea9-4d8c-a83d-4dc0fc99cbff-kube-api-access-thz24\") pod \"redhat-marketplace-bdmpf\" (UID: \"03951083-dea9-4d8c-a83d-4dc0fc99cbff\") " pod="openshift-marketplace/redhat-marketplace-bdmpf" Dec 07 16:37:54 crc kubenswrapper[4716]: I1207 16:37:54.247059 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03951083-dea9-4d8c-a83d-4dc0fc99cbff-catalog-content\") pod \"redhat-marketplace-bdmpf\" (UID: \"03951083-dea9-4d8c-a83d-4dc0fc99cbff\") " pod="openshift-marketplace/redhat-marketplace-bdmpf" Dec 07 16:37:54 crc kubenswrapper[4716]: I1207 16:37:54.247178 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03951083-dea9-4d8c-a83d-4dc0fc99cbff-utilities\") pod \"redhat-marketplace-bdmpf\" (UID: \"03951083-dea9-4d8c-a83d-4dc0fc99cbff\") " pod="openshift-marketplace/redhat-marketplace-bdmpf" Dec 07 16:37:54 crc kubenswrapper[4716]: I1207 16:37:54.349099 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03951083-dea9-4d8c-a83d-4dc0fc99cbff-utilities\") pod \"redhat-marketplace-bdmpf\" (UID: \"03951083-dea9-4d8c-a83d-4dc0fc99cbff\") " pod="openshift-marketplace/redhat-marketplace-bdmpf" Dec 07 16:37:54 crc kubenswrapper[4716]: I1207 16:37:54.349247 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-thz24\" (UniqueName: \"kubernetes.io/projected/03951083-dea9-4d8c-a83d-4dc0fc99cbff-kube-api-access-thz24\") pod \"redhat-marketplace-bdmpf\" (UID: \"03951083-dea9-4d8c-a83d-4dc0fc99cbff\") " pod="openshift-marketplace/redhat-marketplace-bdmpf" Dec 07 16:37:54 crc kubenswrapper[4716]: I1207 16:37:54.349300 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03951083-dea9-4d8c-a83d-4dc0fc99cbff-catalog-content\") pod \"redhat-marketplace-bdmpf\" (UID: \"03951083-dea9-4d8c-a83d-4dc0fc99cbff\") " pod="openshift-marketplace/redhat-marketplace-bdmpf" Dec 07 16:37:54 crc kubenswrapper[4716]: I1207 16:37:54.350145 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03951083-dea9-4d8c-a83d-4dc0fc99cbff-utilities\") pod \"redhat-marketplace-bdmpf\" (UID: \"03951083-dea9-4d8c-a83d-4dc0fc99cbff\") " pod="openshift-marketplace/redhat-marketplace-bdmpf" Dec 07 16:37:54 crc kubenswrapper[4716]: I1207 16:37:54.350165 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03951083-dea9-4d8c-a83d-4dc0fc99cbff-catalog-content\") pod \"redhat-marketplace-bdmpf\" (UID: \"03951083-dea9-4d8c-a83d-4dc0fc99cbff\") " pod="openshift-marketplace/redhat-marketplace-bdmpf" Dec 07 16:37:54 crc kubenswrapper[4716]: I1207 16:37:54.380950 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-thz24\" (UniqueName: \"kubernetes.io/projected/03951083-dea9-4d8c-a83d-4dc0fc99cbff-kube-api-access-thz24\") pod \"redhat-marketplace-bdmpf\" (UID: \"03951083-dea9-4d8c-a83d-4dc0fc99cbff\") " pod="openshift-marketplace/redhat-marketplace-bdmpf" Dec 07 16:37:54 crc kubenswrapper[4716]: I1207 16:37:54.485768 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bdmpf" Dec 07 16:37:55 crc kubenswrapper[4716]: I1207 16:37:55.014907 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bdmpf"] Dec 07 16:37:55 crc kubenswrapper[4716]: I1207 16:37:55.594419 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bdmpf" event={"ID":"03951083-dea9-4d8c-a83d-4dc0fc99cbff","Type":"ContainerStarted","Data":"80dc1b6b88d57140bbb078e5065b10e07e3cf4bbd6b328d429d52951747a885b"} Dec 07 16:37:55 crc kubenswrapper[4716]: I1207 16:37:55.594825 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bdmpf" event={"ID":"03951083-dea9-4d8c-a83d-4dc0fc99cbff","Type":"ContainerStarted","Data":"c5ccebff22ab683fe76bbd83ffb0a337cdec69db0c424756af0f2d096cd9a9bd"} Dec 07 16:37:56 crc kubenswrapper[4716]: I1207 16:37:56.364158 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-v2xw2"] Dec 07 16:37:56 crc kubenswrapper[4716]: I1207 16:37:56.366690 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-v2xw2" Dec 07 16:37:56 crc kubenswrapper[4716]: I1207 16:37:56.374896 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-v2xw2"] Dec 07 16:37:56 crc kubenswrapper[4716]: I1207 16:37:56.488339 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f30ae9ca-c2da-4ed4-89ac-55970d000fd8-catalog-content\") pod \"redhat-operators-v2xw2\" (UID: \"f30ae9ca-c2da-4ed4-89ac-55970d000fd8\") " pod="openshift-marketplace/redhat-operators-v2xw2" Dec 07 16:37:56 crc kubenswrapper[4716]: I1207 16:37:56.488436 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f30ae9ca-c2da-4ed4-89ac-55970d000fd8-utilities\") pod \"redhat-operators-v2xw2\" (UID: \"f30ae9ca-c2da-4ed4-89ac-55970d000fd8\") " pod="openshift-marketplace/redhat-operators-v2xw2" Dec 07 16:37:56 crc kubenswrapper[4716]: I1207 16:37:56.488485 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vkm8\" (UniqueName: \"kubernetes.io/projected/f30ae9ca-c2da-4ed4-89ac-55970d000fd8-kube-api-access-7vkm8\") pod \"redhat-operators-v2xw2\" (UID: \"f30ae9ca-c2da-4ed4-89ac-55970d000fd8\") " pod="openshift-marketplace/redhat-operators-v2xw2" Dec 07 16:37:56 crc kubenswrapper[4716]: I1207 16:37:56.590153 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vkm8\" (UniqueName: \"kubernetes.io/projected/f30ae9ca-c2da-4ed4-89ac-55970d000fd8-kube-api-access-7vkm8\") pod \"redhat-operators-v2xw2\" (UID: \"f30ae9ca-c2da-4ed4-89ac-55970d000fd8\") " pod="openshift-marketplace/redhat-operators-v2xw2" Dec 07 16:37:56 crc kubenswrapper[4716]: I1207 16:37:56.590262 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f30ae9ca-c2da-4ed4-89ac-55970d000fd8-catalog-content\") pod \"redhat-operators-v2xw2\" (UID: \"f30ae9ca-c2da-4ed4-89ac-55970d000fd8\") " pod="openshift-marketplace/redhat-operators-v2xw2" Dec 07 16:37:56 crc kubenswrapper[4716]: I1207 16:37:56.590341 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f30ae9ca-c2da-4ed4-89ac-55970d000fd8-utilities\") pod \"redhat-operators-v2xw2\" (UID: \"f30ae9ca-c2da-4ed4-89ac-55970d000fd8\") " pod="openshift-marketplace/redhat-operators-v2xw2" Dec 07 16:37:56 crc kubenswrapper[4716]: I1207 16:37:56.590754 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f30ae9ca-c2da-4ed4-89ac-55970d000fd8-catalog-content\") pod \"redhat-operators-v2xw2\" (UID: \"f30ae9ca-c2da-4ed4-89ac-55970d000fd8\") " pod="openshift-marketplace/redhat-operators-v2xw2" Dec 07 16:37:56 crc kubenswrapper[4716]: I1207 16:37:56.590777 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f30ae9ca-c2da-4ed4-89ac-55970d000fd8-utilities\") pod \"redhat-operators-v2xw2\" (UID: \"f30ae9ca-c2da-4ed4-89ac-55970d000fd8\") " pod="openshift-marketplace/redhat-operators-v2xw2" Dec 07 16:37:56 crc kubenswrapper[4716]: I1207 16:37:56.604874 4716 generic.go:334] "Generic (PLEG): container finished" podID="03951083-dea9-4d8c-a83d-4dc0fc99cbff" containerID="80dc1b6b88d57140bbb078e5065b10e07e3cf4bbd6b328d429d52951747a885b" exitCode=0 Dec 07 16:37:56 crc kubenswrapper[4716]: I1207 16:37:56.604929 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bdmpf" event={"ID":"03951083-dea9-4d8c-a83d-4dc0fc99cbff","Type":"ContainerDied","Data":"80dc1b6b88d57140bbb078e5065b10e07e3cf4bbd6b328d429d52951747a885b"} Dec 07 16:37:56 crc kubenswrapper[4716]: I1207 16:37:56.613488 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vkm8\" (UniqueName: \"kubernetes.io/projected/f30ae9ca-c2da-4ed4-89ac-55970d000fd8-kube-api-access-7vkm8\") pod \"redhat-operators-v2xw2\" (UID: \"f30ae9ca-c2da-4ed4-89ac-55970d000fd8\") " pod="openshift-marketplace/redhat-operators-v2xw2" Dec 07 16:37:56 crc kubenswrapper[4716]: I1207 16:37:56.686314 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-v2xw2" Dec 07 16:37:57 crc kubenswrapper[4716]: I1207 16:37:57.101071 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-pnjtg" Dec 07 16:37:57 crc kubenswrapper[4716]: I1207 16:37:57.101468 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-pnjtg" Dec 07 16:37:57 crc kubenswrapper[4716]: I1207 16:37:57.155753 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-v2xw2"] Dec 07 16:37:57 crc kubenswrapper[4716]: I1207 16:37:57.159198 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-pnjtg" Dec 07 16:37:57 crc kubenswrapper[4716]: W1207 16:37:57.166407 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf30ae9ca_c2da_4ed4_89ac_55970d000fd8.slice/crio-041aa6dc87e523d6d985b6937e1d159d7c14f34d3f8fdf9bb1a683d3ccfaa949 WatchSource:0}: Error finding container 041aa6dc87e523d6d985b6937e1d159d7c14f34d3f8fdf9bb1a683d3ccfaa949: Status 404 returned error can't find the container with id 041aa6dc87e523d6d985b6937e1d159d7c14f34d3f8fdf9bb1a683d3ccfaa949 Dec 07 16:37:57 crc kubenswrapper[4716]: I1207 16:37:57.614349 4716 generic.go:334] "Generic (PLEG): container finished" podID="03951083-dea9-4d8c-a83d-4dc0fc99cbff" containerID="763fb8102e7be00d7a6dc5c4521c998713f0e94083b0ef21965cac83206d94ee" exitCode=0 Dec 07 16:37:57 crc kubenswrapper[4716]: I1207 16:37:57.614416 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bdmpf" event={"ID":"03951083-dea9-4d8c-a83d-4dc0fc99cbff","Type":"ContainerDied","Data":"763fb8102e7be00d7a6dc5c4521c998713f0e94083b0ef21965cac83206d94ee"} Dec 07 16:37:57 crc kubenswrapper[4716]: I1207 16:37:57.617507 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v2xw2" event={"ID":"f30ae9ca-c2da-4ed4-89ac-55970d000fd8","Type":"ContainerStarted","Data":"041aa6dc87e523d6d985b6937e1d159d7c14f34d3f8fdf9bb1a683d3ccfaa949"} Dec 07 16:37:57 crc kubenswrapper[4716]: I1207 16:37:57.668169 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-pnjtg" Dec 07 16:37:58 crc kubenswrapper[4716]: I1207 16:37:58.628420 4716 generic.go:334] "Generic (PLEG): container finished" podID="f30ae9ca-c2da-4ed4-89ac-55970d000fd8" containerID="98319ef0f68e3aa4631550a4fdfb8e901bc46d3035904197f987ae897dc4c250" exitCode=0 Dec 07 16:37:58 crc kubenswrapper[4716]: I1207 16:37:58.628950 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v2xw2" event={"ID":"f30ae9ca-c2da-4ed4-89ac-55970d000fd8","Type":"ContainerDied","Data":"98319ef0f68e3aa4631550a4fdfb8e901bc46d3035904197f987ae897dc4c250"} Dec 07 16:37:59 crc kubenswrapper[4716]: I1207 16:37:59.639870 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v2xw2" event={"ID":"f30ae9ca-c2da-4ed4-89ac-55970d000fd8","Type":"ContainerStarted","Data":"63e609dace34cb195151c54d0c1a3993488e04cd66588790df22f3e7ceb59ff8"} Dec 07 16:37:59 crc kubenswrapper[4716]: I1207 16:37:59.641923 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bdmpf" event={"ID":"03951083-dea9-4d8c-a83d-4dc0fc99cbff","Type":"ContainerStarted","Data":"55d6361dc3a754e3e22df84fde267577b39c357c9982c5ee4f6b67d38cb7eea0"} Dec 07 16:37:59 crc kubenswrapper[4716]: I1207 16:37:59.691438 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-bdmpf" podStartSLOduration=2.586972986 podStartE2EDuration="5.691421008s" podCreationTimestamp="2025-12-07 16:37:54 +0000 UTC" firstStartedPulling="2025-12-07 16:37:55.596327188 +0000 UTC m=+2138.286612100" lastFinishedPulling="2025-12-07 16:37:58.70077521 +0000 UTC m=+2141.391060122" observedRunningTime="2025-12-07 16:37:59.682671372 +0000 UTC m=+2142.372956284" watchObservedRunningTime="2025-12-07 16:37:59.691421008 +0000 UTC m=+2142.381705920" Dec 07 16:38:00 crc kubenswrapper[4716]: I1207 16:38:00.758248 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pnjtg"] Dec 07 16:38:00 crc kubenswrapper[4716]: I1207 16:38:00.758963 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-pnjtg" podUID="2c0ad3b2-97b0-4c37-85da-ab61e86ef418" containerName="registry-server" containerID="cri-o://d925ae6ace86e5385cea11dfab9747ab35a07fbedd1b95fcfb55cf7e8c975310" gracePeriod=2 Dec 07 16:38:02 crc kubenswrapper[4716]: I1207 16:38:02.680829 4716 generic.go:334] "Generic (PLEG): container finished" podID="f30ae9ca-c2da-4ed4-89ac-55970d000fd8" containerID="63e609dace34cb195151c54d0c1a3993488e04cd66588790df22f3e7ceb59ff8" exitCode=0 Dec 07 16:38:02 crc kubenswrapper[4716]: I1207 16:38:02.680955 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v2xw2" event={"ID":"f30ae9ca-c2da-4ed4-89ac-55970d000fd8","Type":"ContainerDied","Data":"63e609dace34cb195151c54d0c1a3993488e04cd66588790df22f3e7ceb59ff8"} Dec 07 16:38:03 crc kubenswrapper[4716]: I1207 16:38:03.699614 4716 generic.go:334] "Generic (PLEG): container finished" podID="2c0ad3b2-97b0-4c37-85da-ab61e86ef418" containerID="d925ae6ace86e5385cea11dfab9747ab35a07fbedd1b95fcfb55cf7e8c975310" exitCode=0 Dec 07 16:38:03 crc kubenswrapper[4716]: I1207 16:38:03.699728 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pnjtg" event={"ID":"2c0ad3b2-97b0-4c37-85da-ab61e86ef418","Type":"ContainerDied","Data":"d925ae6ace86e5385cea11dfab9747ab35a07fbedd1b95fcfb55cf7e8c975310"} Dec 07 16:38:03 crc kubenswrapper[4716]: I1207 16:38:03.910998 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pnjtg" Dec 07 16:38:03 crc kubenswrapper[4716]: I1207 16:38:03.926189 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fwcwz\" (UniqueName: \"kubernetes.io/projected/2c0ad3b2-97b0-4c37-85da-ab61e86ef418-kube-api-access-fwcwz\") pod \"2c0ad3b2-97b0-4c37-85da-ab61e86ef418\" (UID: \"2c0ad3b2-97b0-4c37-85da-ab61e86ef418\") " Dec 07 16:38:03 crc kubenswrapper[4716]: I1207 16:38:03.926330 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c0ad3b2-97b0-4c37-85da-ab61e86ef418-catalog-content\") pod \"2c0ad3b2-97b0-4c37-85da-ab61e86ef418\" (UID: \"2c0ad3b2-97b0-4c37-85da-ab61e86ef418\") " Dec 07 16:38:03 crc kubenswrapper[4716]: I1207 16:38:03.926521 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c0ad3b2-97b0-4c37-85da-ab61e86ef418-utilities\") pod \"2c0ad3b2-97b0-4c37-85da-ab61e86ef418\" (UID: \"2c0ad3b2-97b0-4c37-85da-ab61e86ef418\") " Dec 07 16:38:03 crc kubenswrapper[4716]: I1207 16:38:03.927252 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2c0ad3b2-97b0-4c37-85da-ab61e86ef418-utilities" (OuterVolumeSpecName: "utilities") pod "2c0ad3b2-97b0-4c37-85da-ab61e86ef418" (UID: "2c0ad3b2-97b0-4c37-85da-ab61e86ef418"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:38:03 crc kubenswrapper[4716]: I1207 16:38:03.927746 4716 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c0ad3b2-97b0-4c37-85da-ab61e86ef418-utilities\") on node \"crc\" DevicePath \"\"" Dec 07 16:38:03 crc kubenswrapper[4716]: I1207 16:38:03.933542 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c0ad3b2-97b0-4c37-85da-ab61e86ef418-kube-api-access-fwcwz" (OuterVolumeSpecName: "kube-api-access-fwcwz") pod "2c0ad3b2-97b0-4c37-85da-ab61e86ef418" (UID: "2c0ad3b2-97b0-4c37-85da-ab61e86ef418"). InnerVolumeSpecName "kube-api-access-fwcwz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:38:03 crc kubenswrapper[4716]: I1207 16:38:03.979336 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2c0ad3b2-97b0-4c37-85da-ab61e86ef418-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2c0ad3b2-97b0-4c37-85da-ab61e86ef418" (UID: "2c0ad3b2-97b0-4c37-85da-ab61e86ef418"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:38:04 crc kubenswrapper[4716]: I1207 16:38:04.029528 4716 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c0ad3b2-97b0-4c37-85da-ab61e86ef418-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 07 16:38:04 crc kubenswrapper[4716]: I1207 16:38:04.029571 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fwcwz\" (UniqueName: \"kubernetes.io/projected/2c0ad3b2-97b0-4c37-85da-ab61e86ef418-kube-api-access-fwcwz\") on node \"crc\" DevicePath \"\"" Dec 07 16:38:04 crc kubenswrapper[4716]: I1207 16:38:04.487101 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-bdmpf" Dec 07 16:38:04 crc kubenswrapper[4716]: I1207 16:38:04.487350 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-bdmpf" Dec 07 16:38:04 crc kubenswrapper[4716]: I1207 16:38:04.544208 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-bdmpf" Dec 07 16:38:04 crc kubenswrapper[4716]: I1207 16:38:04.711979 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pnjtg" event={"ID":"2c0ad3b2-97b0-4c37-85da-ab61e86ef418","Type":"ContainerDied","Data":"712e66fdf5206f0fe77121a10baa2a125c9c3090b90e8daa736278112d1889e4"} Dec 07 16:38:04 crc kubenswrapper[4716]: I1207 16:38:04.712006 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pnjtg" Dec 07 16:38:04 crc kubenswrapper[4716]: I1207 16:38:04.712068 4716 scope.go:117] "RemoveContainer" containerID="d925ae6ace86e5385cea11dfab9747ab35a07fbedd1b95fcfb55cf7e8c975310" Dec 07 16:38:04 crc kubenswrapper[4716]: I1207 16:38:04.714152 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v2xw2" event={"ID":"f30ae9ca-c2da-4ed4-89ac-55970d000fd8","Type":"ContainerStarted","Data":"208c7925c76f2b82ae3ab3315599a38317ce8160e5fcf6a17300b612ac328562"} Dec 07 16:38:04 crc kubenswrapper[4716]: I1207 16:38:04.745816 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-v2xw2" podStartSLOduration=3.907422754 podStartE2EDuration="8.745798879s" podCreationTimestamp="2025-12-07 16:37:56 +0000 UTC" firstStartedPulling="2025-12-07 16:37:58.651112203 +0000 UTC m=+2141.341397115" lastFinishedPulling="2025-12-07 16:38:03.489488328 +0000 UTC m=+2146.179773240" observedRunningTime="2025-12-07 16:38:04.73987025 +0000 UTC m=+2147.430155162" watchObservedRunningTime="2025-12-07 16:38:04.745798879 +0000 UTC m=+2147.436083791" Dec 07 16:38:04 crc kubenswrapper[4716]: I1207 16:38:04.746172 4716 scope.go:117] "RemoveContainer" containerID="24cf28fe02990feccec64f3129bc206ea89e3ca05b8aafdc5f3f11d12e34d3ce" Dec 07 16:38:04 crc kubenswrapper[4716]: I1207 16:38:04.783851 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pnjtg"] Dec 07 16:38:04 crc kubenswrapper[4716]: I1207 16:38:04.790103 4716 scope.go:117] "RemoveContainer" containerID="f7c9f5749af69e1bebeaf8993dac87f8ba53edf646812ee6ef6daa3786b16704" Dec 07 16:38:04 crc kubenswrapper[4716]: I1207 16:38:04.793807 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-pnjtg"] Dec 07 16:38:04 crc kubenswrapper[4716]: I1207 16:38:04.809917 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-bdmpf" Dec 07 16:38:05 crc kubenswrapper[4716]: I1207 16:38:05.671397 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c0ad3b2-97b0-4c37-85da-ab61e86ef418" path="/var/lib/kubelet/pods/2c0ad3b2-97b0-4c37-85da-ab61e86ef418/volumes" Dec 07 16:38:06 crc kubenswrapper[4716]: I1207 16:38:06.687255 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-v2xw2" Dec 07 16:38:06 crc kubenswrapper[4716]: I1207 16:38:06.689046 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-v2xw2" Dec 07 16:38:07 crc kubenswrapper[4716]: I1207 16:38:07.157154 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bdmpf"] Dec 07 16:38:07 crc kubenswrapper[4716]: I1207 16:38:07.735241 4716 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-v2xw2" podUID="f30ae9ca-c2da-4ed4-89ac-55970d000fd8" containerName="registry-server" probeResult="failure" output=< Dec 07 16:38:07 crc kubenswrapper[4716]: timeout: failed to connect service ":50051" within 1s Dec 07 16:38:07 crc kubenswrapper[4716]: > Dec 07 16:38:07 crc kubenswrapper[4716]: I1207 16:38:07.740090 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-bdmpf" podUID="03951083-dea9-4d8c-a83d-4dc0fc99cbff" containerName="registry-server" containerID="cri-o://55d6361dc3a754e3e22df84fde267577b39c357c9982c5ee4f6b67d38cb7eea0" gracePeriod=2 Dec 07 16:38:08 crc kubenswrapper[4716]: I1207 16:38:08.193826 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bdmpf" Dec 07 16:38:08 crc kubenswrapper[4716]: I1207 16:38:08.204378 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-thz24\" (UniqueName: \"kubernetes.io/projected/03951083-dea9-4d8c-a83d-4dc0fc99cbff-kube-api-access-thz24\") pod \"03951083-dea9-4d8c-a83d-4dc0fc99cbff\" (UID: \"03951083-dea9-4d8c-a83d-4dc0fc99cbff\") " Dec 07 16:38:08 crc kubenswrapper[4716]: I1207 16:38:08.204498 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03951083-dea9-4d8c-a83d-4dc0fc99cbff-utilities\") pod \"03951083-dea9-4d8c-a83d-4dc0fc99cbff\" (UID: \"03951083-dea9-4d8c-a83d-4dc0fc99cbff\") " Dec 07 16:38:08 crc kubenswrapper[4716]: I1207 16:38:08.204654 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03951083-dea9-4d8c-a83d-4dc0fc99cbff-catalog-content\") pod \"03951083-dea9-4d8c-a83d-4dc0fc99cbff\" (UID: \"03951083-dea9-4d8c-a83d-4dc0fc99cbff\") " Dec 07 16:38:08 crc kubenswrapper[4716]: I1207 16:38:08.205020 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03951083-dea9-4d8c-a83d-4dc0fc99cbff-utilities" (OuterVolumeSpecName: "utilities") pod "03951083-dea9-4d8c-a83d-4dc0fc99cbff" (UID: "03951083-dea9-4d8c-a83d-4dc0fc99cbff"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:38:08 crc kubenswrapper[4716]: I1207 16:38:08.205165 4716 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03951083-dea9-4d8c-a83d-4dc0fc99cbff-utilities\") on node \"crc\" DevicePath \"\"" Dec 07 16:38:08 crc kubenswrapper[4716]: I1207 16:38:08.211230 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03951083-dea9-4d8c-a83d-4dc0fc99cbff-kube-api-access-thz24" (OuterVolumeSpecName: "kube-api-access-thz24") pod "03951083-dea9-4d8c-a83d-4dc0fc99cbff" (UID: "03951083-dea9-4d8c-a83d-4dc0fc99cbff"). InnerVolumeSpecName "kube-api-access-thz24". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:38:08 crc kubenswrapper[4716]: I1207 16:38:08.238974 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03951083-dea9-4d8c-a83d-4dc0fc99cbff-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "03951083-dea9-4d8c-a83d-4dc0fc99cbff" (UID: "03951083-dea9-4d8c-a83d-4dc0fc99cbff"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:38:08 crc kubenswrapper[4716]: I1207 16:38:08.307197 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-thz24\" (UniqueName: \"kubernetes.io/projected/03951083-dea9-4d8c-a83d-4dc0fc99cbff-kube-api-access-thz24\") on node \"crc\" DevicePath \"\"" Dec 07 16:38:08 crc kubenswrapper[4716]: I1207 16:38:08.307233 4716 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03951083-dea9-4d8c-a83d-4dc0fc99cbff-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 07 16:38:08 crc kubenswrapper[4716]: I1207 16:38:08.749841 4716 generic.go:334] "Generic (PLEG): container finished" podID="03951083-dea9-4d8c-a83d-4dc0fc99cbff" containerID="55d6361dc3a754e3e22df84fde267577b39c357c9982c5ee4f6b67d38cb7eea0" exitCode=0 Dec 07 16:38:08 crc kubenswrapper[4716]: I1207 16:38:08.749896 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bdmpf" event={"ID":"03951083-dea9-4d8c-a83d-4dc0fc99cbff","Type":"ContainerDied","Data":"55d6361dc3a754e3e22df84fde267577b39c357c9982c5ee4f6b67d38cb7eea0"} Dec 07 16:38:08 crc kubenswrapper[4716]: I1207 16:38:08.750314 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bdmpf" event={"ID":"03951083-dea9-4d8c-a83d-4dc0fc99cbff","Type":"ContainerDied","Data":"c5ccebff22ab683fe76bbd83ffb0a337cdec69db0c424756af0f2d096cd9a9bd"} Dec 07 16:38:08 crc kubenswrapper[4716]: I1207 16:38:08.749929 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bdmpf" Dec 07 16:38:08 crc kubenswrapper[4716]: I1207 16:38:08.750334 4716 scope.go:117] "RemoveContainer" containerID="55d6361dc3a754e3e22df84fde267577b39c357c9982c5ee4f6b67d38cb7eea0" Dec 07 16:38:08 crc kubenswrapper[4716]: I1207 16:38:08.775905 4716 scope.go:117] "RemoveContainer" containerID="763fb8102e7be00d7a6dc5c4521c998713f0e94083b0ef21965cac83206d94ee" Dec 07 16:38:08 crc kubenswrapper[4716]: I1207 16:38:08.788383 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bdmpf"] Dec 07 16:38:08 crc kubenswrapper[4716]: I1207 16:38:08.797811 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-bdmpf"] Dec 07 16:38:08 crc kubenswrapper[4716]: I1207 16:38:08.808303 4716 scope.go:117] "RemoveContainer" containerID="80dc1b6b88d57140bbb078e5065b10e07e3cf4bbd6b328d429d52951747a885b" Dec 07 16:38:08 crc kubenswrapper[4716]: E1207 16:38:08.837935 4716 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod03951083_dea9_4d8c_a83d_4dc0fc99cbff.slice/crio-c5ccebff22ab683fe76bbd83ffb0a337cdec69db0c424756af0f2d096cd9a9bd\": RecentStats: unable to find data in memory cache]" Dec 07 16:38:08 crc kubenswrapper[4716]: I1207 16:38:08.847766 4716 scope.go:117] "RemoveContainer" containerID="55d6361dc3a754e3e22df84fde267577b39c357c9982c5ee4f6b67d38cb7eea0" Dec 07 16:38:08 crc kubenswrapper[4716]: E1207 16:38:08.848126 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"55d6361dc3a754e3e22df84fde267577b39c357c9982c5ee4f6b67d38cb7eea0\": container with ID starting with 55d6361dc3a754e3e22df84fde267577b39c357c9982c5ee4f6b67d38cb7eea0 not found: ID does not exist" containerID="55d6361dc3a754e3e22df84fde267577b39c357c9982c5ee4f6b67d38cb7eea0" Dec 07 16:38:08 crc kubenswrapper[4716]: I1207 16:38:08.848162 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55d6361dc3a754e3e22df84fde267577b39c357c9982c5ee4f6b67d38cb7eea0"} err="failed to get container status \"55d6361dc3a754e3e22df84fde267577b39c357c9982c5ee4f6b67d38cb7eea0\": rpc error: code = NotFound desc = could not find container \"55d6361dc3a754e3e22df84fde267577b39c357c9982c5ee4f6b67d38cb7eea0\": container with ID starting with 55d6361dc3a754e3e22df84fde267577b39c357c9982c5ee4f6b67d38cb7eea0 not found: ID does not exist" Dec 07 16:38:08 crc kubenswrapper[4716]: I1207 16:38:08.848185 4716 scope.go:117] "RemoveContainer" containerID="763fb8102e7be00d7a6dc5c4521c998713f0e94083b0ef21965cac83206d94ee" Dec 07 16:38:08 crc kubenswrapper[4716]: E1207 16:38:08.848567 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"763fb8102e7be00d7a6dc5c4521c998713f0e94083b0ef21965cac83206d94ee\": container with ID starting with 763fb8102e7be00d7a6dc5c4521c998713f0e94083b0ef21965cac83206d94ee not found: ID does not exist" containerID="763fb8102e7be00d7a6dc5c4521c998713f0e94083b0ef21965cac83206d94ee" Dec 07 16:38:08 crc kubenswrapper[4716]: I1207 16:38:08.848589 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"763fb8102e7be00d7a6dc5c4521c998713f0e94083b0ef21965cac83206d94ee"} err="failed to get container status \"763fb8102e7be00d7a6dc5c4521c998713f0e94083b0ef21965cac83206d94ee\": rpc error: code = NotFound desc = could not find container \"763fb8102e7be00d7a6dc5c4521c998713f0e94083b0ef21965cac83206d94ee\": container with ID starting with 763fb8102e7be00d7a6dc5c4521c998713f0e94083b0ef21965cac83206d94ee not found: ID does not exist" Dec 07 16:38:08 crc kubenswrapper[4716]: I1207 16:38:08.848602 4716 scope.go:117] "RemoveContainer" containerID="80dc1b6b88d57140bbb078e5065b10e07e3cf4bbd6b328d429d52951747a885b" Dec 07 16:38:08 crc kubenswrapper[4716]: E1207 16:38:08.848819 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80dc1b6b88d57140bbb078e5065b10e07e3cf4bbd6b328d429d52951747a885b\": container with ID starting with 80dc1b6b88d57140bbb078e5065b10e07e3cf4bbd6b328d429d52951747a885b not found: ID does not exist" containerID="80dc1b6b88d57140bbb078e5065b10e07e3cf4bbd6b328d429d52951747a885b" Dec 07 16:38:08 crc kubenswrapper[4716]: I1207 16:38:08.848842 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80dc1b6b88d57140bbb078e5065b10e07e3cf4bbd6b328d429d52951747a885b"} err="failed to get container status \"80dc1b6b88d57140bbb078e5065b10e07e3cf4bbd6b328d429d52951747a885b\": rpc error: code = NotFound desc = could not find container \"80dc1b6b88d57140bbb078e5065b10e07e3cf4bbd6b328d429d52951747a885b\": container with ID starting with 80dc1b6b88d57140bbb078e5065b10e07e3cf4bbd6b328d429d52951747a885b not found: ID does not exist" Dec 07 16:38:09 crc kubenswrapper[4716]: I1207 16:38:09.667238 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03951083-dea9-4d8c-a83d-4dc0fc99cbff" path="/var/lib/kubelet/pods/03951083-dea9-4d8c-a83d-4dc0fc99cbff/volumes" Dec 07 16:38:16 crc kubenswrapper[4716]: I1207 16:38:16.742031 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-v2xw2" Dec 07 16:38:16 crc kubenswrapper[4716]: I1207 16:38:16.815351 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-v2xw2" Dec 07 16:38:17 crc kubenswrapper[4716]: I1207 16:38:17.981334 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-v2xw2"] Dec 07 16:38:17 crc kubenswrapper[4716]: I1207 16:38:17.981592 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-v2xw2" podUID="f30ae9ca-c2da-4ed4-89ac-55970d000fd8" containerName="registry-server" containerID="cri-o://208c7925c76f2b82ae3ab3315599a38317ce8160e5fcf6a17300b612ac328562" gracePeriod=2 Dec 07 16:38:18 crc kubenswrapper[4716]: I1207 16:38:18.482218 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-v2xw2" Dec 07 16:38:18 crc kubenswrapper[4716]: I1207 16:38:18.590164 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7vkm8\" (UniqueName: \"kubernetes.io/projected/f30ae9ca-c2da-4ed4-89ac-55970d000fd8-kube-api-access-7vkm8\") pod \"f30ae9ca-c2da-4ed4-89ac-55970d000fd8\" (UID: \"f30ae9ca-c2da-4ed4-89ac-55970d000fd8\") " Dec 07 16:38:18 crc kubenswrapper[4716]: I1207 16:38:18.590322 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f30ae9ca-c2da-4ed4-89ac-55970d000fd8-utilities\") pod \"f30ae9ca-c2da-4ed4-89ac-55970d000fd8\" (UID: \"f30ae9ca-c2da-4ed4-89ac-55970d000fd8\") " Dec 07 16:38:18 crc kubenswrapper[4716]: I1207 16:38:18.590411 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f30ae9ca-c2da-4ed4-89ac-55970d000fd8-catalog-content\") pod \"f30ae9ca-c2da-4ed4-89ac-55970d000fd8\" (UID: \"f30ae9ca-c2da-4ed4-89ac-55970d000fd8\") " Dec 07 16:38:18 crc kubenswrapper[4716]: I1207 16:38:18.591896 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f30ae9ca-c2da-4ed4-89ac-55970d000fd8-utilities" (OuterVolumeSpecName: "utilities") pod "f30ae9ca-c2da-4ed4-89ac-55970d000fd8" (UID: "f30ae9ca-c2da-4ed4-89ac-55970d000fd8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:38:18 crc kubenswrapper[4716]: I1207 16:38:18.597328 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f30ae9ca-c2da-4ed4-89ac-55970d000fd8-kube-api-access-7vkm8" (OuterVolumeSpecName: "kube-api-access-7vkm8") pod "f30ae9ca-c2da-4ed4-89ac-55970d000fd8" (UID: "f30ae9ca-c2da-4ed4-89ac-55970d000fd8"). InnerVolumeSpecName "kube-api-access-7vkm8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:38:18 crc kubenswrapper[4716]: I1207 16:38:18.692331 4716 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f30ae9ca-c2da-4ed4-89ac-55970d000fd8-utilities\") on node \"crc\" DevicePath \"\"" Dec 07 16:38:18 crc kubenswrapper[4716]: I1207 16:38:18.692372 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7vkm8\" (UniqueName: \"kubernetes.io/projected/f30ae9ca-c2da-4ed4-89ac-55970d000fd8-kube-api-access-7vkm8\") on node \"crc\" DevicePath \"\"" Dec 07 16:38:18 crc kubenswrapper[4716]: I1207 16:38:18.727645 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f30ae9ca-c2da-4ed4-89ac-55970d000fd8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f30ae9ca-c2da-4ed4-89ac-55970d000fd8" (UID: "f30ae9ca-c2da-4ed4-89ac-55970d000fd8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:38:18 crc kubenswrapper[4716]: I1207 16:38:18.794346 4716 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f30ae9ca-c2da-4ed4-89ac-55970d000fd8-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 07 16:38:18 crc kubenswrapper[4716]: I1207 16:38:18.844454 4716 generic.go:334] "Generic (PLEG): container finished" podID="f30ae9ca-c2da-4ed4-89ac-55970d000fd8" containerID="208c7925c76f2b82ae3ab3315599a38317ce8160e5fcf6a17300b612ac328562" exitCode=0 Dec 07 16:38:18 crc kubenswrapper[4716]: I1207 16:38:18.844501 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v2xw2" event={"ID":"f30ae9ca-c2da-4ed4-89ac-55970d000fd8","Type":"ContainerDied","Data":"208c7925c76f2b82ae3ab3315599a38317ce8160e5fcf6a17300b612ac328562"} Dec 07 16:38:18 crc kubenswrapper[4716]: I1207 16:38:18.844526 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v2xw2" event={"ID":"f30ae9ca-c2da-4ed4-89ac-55970d000fd8","Type":"ContainerDied","Data":"041aa6dc87e523d6d985b6937e1d159d7c14f34d3f8fdf9bb1a683d3ccfaa949"} Dec 07 16:38:18 crc kubenswrapper[4716]: I1207 16:38:18.844543 4716 scope.go:117] "RemoveContainer" containerID="208c7925c76f2b82ae3ab3315599a38317ce8160e5fcf6a17300b612ac328562" Dec 07 16:38:18 crc kubenswrapper[4716]: I1207 16:38:18.844540 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-v2xw2" Dec 07 16:38:18 crc kubenswrapper[4716]: I1207 16:38:18.876071 4716 scope.go:117] "RemoveContainer" containerID="63e609dace34cb195151c54d0c1a3993488e04cd66588790df22f3e7ceb59ff8" Dec 07 16:38:18 crc kubenswrapper[4716]: I1207 16:38:18.890638 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-v2xw2"] Dec 07 16:38:18 crc kubenswrapper[4716]: I1207 16:38:18.898374 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-v2xw2"] Dec 07 16:38:18 crc kubenswrapper[4716]: I1207 16:38:18.909364 4716 scope.go:117] "RemoveContainer" containerID="98319ef0f68e3aa4631550a4fdfb8e901bc46d3035904197f987ae897dc4c250" Dec 07 16:38:18 crc kubenswrapper[4716]: I1207 16:38:18.947586 4716 scope.go:117] "RemoveContainer" containerID="208c7925c76f2b82ae3ab3315599a38317ce8160e5fcf6a17300b612ac328562" Dec 07 16:38:18 crc kubenswrapper[4716]: E1207 16:38:18.948045 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"208c7925c76f2b82ae3ab3315599a38317ce8160e5fcf6a17300b612ac328562\": container with ID starting with 208c7925c76f2b82ae3ab3315599a38317ce8160e5fcf6a17300b612ac328562 not found: ID does not exist" containerID="208c7925c76f2b82ae3ab3315599a38317ce8160e5fcf6a17300b612ac328562" Dec 07 16:38:18 crc kubenswrapper[4716]: I1207 16:38:18.948121 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"208c7925c76f2b82ae3ab3315599a38317ce8160e5fcf6a17300b612ac328562"} err="failed to get container status \"208c7925c76f2b82ae3ab3315599a38317ce8160e5fcf6a17300b612ac328562\": rpc error: code = NotFound desc = could not find container \"208c7925c76f2b82ae3ab3315599a38317ce8160e5fcf6a17300b612ac328562\": container with ID starting with 208c7925c76f2b82ae3ab3315599a38317ce8160e5fcf6a17300b612ac328562 not found: ID does not exist" Dec 07 16:38:18 crc kubenswrapper[4716]: I1207 16:38:18.948147 4716 scope.go:117] "RemoveContainer" containerID="63e609dace34cb195151c54d0c1a3993488e04cd66588790df22f3e7ceb59ff8" Dec 07 16:38:18 crc kubenswrapper[4716]: E1207 16:38:18.948536 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"63e609dace34cb195151c54d0c1a3993488e04cd66588790df22f3e7ceb59ff8\": container with ID starting with 63e609dace34cb195151c54d0c1a3993488e04cd66588790df22f3e7ceb59ff8 not found: ID does not exist" containerID="63e609dace34cb195151c54d0c1a3993488e04cd66588790df22f3e7ceb59ff8" Dec 07 16:38:18 crc kubenswrapper[4716]: I1207 16:38:18.948557 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"63e609dace34cb195151c54d0c1a3993488e04cd66588790df22f3e7ceb59ff8"} err="failed to get container status \"63e609dace34cb195151c54d0c1a3993488e04cd66588790df22f3e7ceb59ff8\": rpc error: code = NotFound desc = could not find container \"63e609dace34cb195151c54d0c1a3993488e04cd66588790df22f3e7ceb59ff8\": container with ID starting with 63e609dace34cb195151c54d0c1a3993488e04cd66588790df22f3e7ceb59ff8 not found: ID does not exist" Dec 07 16:38:18 crc kubenswrapper[4716]: I1207 16:38:18.948586 4716 scope.go:117] "RemoveContainer" containerID="98319ef0f68e3aa4631550a4fdfb8e901bc46d3035904197f987ae897dc4c250" Dec 07 16:38:18 crc kubenswrapper[4716]: E1207 16:38:18.948950 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98319ef0f68e3aa4631550a4fdfb8e901bc46d3035904197f987ae897dc4c250\": container with ID starting with 98319ef0f68e3aa4631550a4fdfb8e901bc46d3035904197f987ae897dc4c250 not found: ID does not exist" containerID="98319ef0f68e3aa4631550a4fdfb8e901bc46d3035904197f987ae897dc4c250" Dec 07 16:38:18 crc kubenswrapper[4716]: I1207 16:38:18.949005 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98319ef0f68e3aa4631550a4fdfb8e901bc46d3035904197f987ae897dc4c250"} err="failed to get container status \"98319ef0f68e3aa4631550a4fdfb8e901bc46d3035904197f987ae897dc4c250\": rpc error: code = NotFound desc = could not find container \"98319ef0f68e3aa4631550a4fdfb8e901bc46d3035904197f987ae897dc4c250\": container with ID starting with 98319ef0f68e3aa4631550a4fdfb8e901bc46d3035904197f987ae897dc4c250 not found: ID does not exist" Dec 07 16:38:19 crc kubenswrapper[4716]: I1207 16:38:19.678517 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f30ae9ca-c2da-4ed4-89ac-55970d000fd8" path="/var/lib/kubelet/pods/f30ae9ca-c2da-4ed4-89ac-55970d000fd8/volumes" Dec 07 16:38:22 crc kubenswrapper[4716]: I1207 16:38:22.761044 4716 patch_prober.go:28] interesting pod/machine-config-daemon-zcxxp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 16:38:22 crc kubenswrapper[4716]: I1207 16:38:22.761520 4716 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 16:38:52 crc kubenswrapper[4716]: I1207 16:38:52.762562 4716 patch_prober.go:28] interesting pod/machine-config-daemon-zcxxp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 16:38:52 crc kubenswrapper[4716]: I1207 16:38:52.763308 4716 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 16:38:52 crc kubenswrapper[4716]: I1207 16:38:52.763364 4716 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" Dec 07 16:38:52 crc kubenswrapper[4716]: I1207 16:38:52.764146 4716 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8aa74b5db626ce69c5f618707bfc1aab2b0242d5c7b440dfc5c0166d72b0d08c"} pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 07 16:38:52 crc kubenswrapper[4716]: I1207 16:38:52.764209 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" containerName="machine-config-daemon" containerID="cri-o://8aa74b5db626ce69c5f618707bfc1aab2b0242d5c7b440dfc5c0166d72b0d08c" gracePeriod=600 Dec 07 16:38:53 crc kubenswrapper[4716]: I1207 16:38:53.178734 4716 generic.go:334] "Generic (PLEG): container finished" podID="c15b59eb-565d-4556-a4ce-75afdf159dc8" containerID="8aa74b5db626ce69c5f618707bfc1aab2b0242d5c7b440dfc5c0166d72b0d08c" exitCode=0 Dec 07 16:38:53 crc kubenswrapper[4716]: I1207 16:38:53.178811 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" event={"ID":"c15b59eb-565d-4556-a4ce-75afdf159dc8","Type":"ContainerDied","Data":"8aa74b5db626ce69c5f618707bfc1aab2b0242d5c7b440dfc5c0166d72b0d08c"} Dec 07 16:38:53 crc kubenswrapper[4716]: I1207 16:38:53.179169 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" event={"ID":"c15b59eb-565d-4556-a4ce-75afdf159dc8","Type":"ContainerStarted","Data":"3eb63b7962def203a31ba32ba16ac93470e254c2a4edb4a4dd2033b290e2f217"} Dec 07 16:38:53 crc kubenswrapper[4716]: I1207 16:38:53.179195 4716 scope.go:117] "RemoveContainer" containerID="bae1a3e9d80f5b00293cade8b7601799399c2883ae101574459a4c01e523a154" Dec 07 16:41:06 crc kubenswrapper[4716]: I1207 16:41:06.391860 4716 generic.go:334] "Generic (PLEG): container finished" podID="0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3" containerID="4ad48961b4f8dd5fe3b04c46db51ce1d65f9a1ac75529d8fdd3541da9f90e012" exitCode=0 Dec 07 16:41:06 crc kubenswrapper[4716]: I1207 16:41:06.392017 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-66mjc" event={"ID":"0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3","Type":"ContainerDied","Data":"4ad48961b4f8dd5fe3b04c46db51ce1d65f9a1ac75529d8fdd3541da9f90e012"} Dec 07 16:41:07 crc kubenswrapper[4716]: I1207 16:41:07.794231 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-66mjc" Dec 07 16:41:07 crc kubenswrapper[4716]: I1207 16:41:07.898707 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3-ssh-key\") pod \"0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3\" (UID: \"0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3\") " Dec 07 16:41:07 crc kubenswrapper[4716]: I1207 16:41:07.898817 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k6dgs\" (UniqueName: \"kubernetes.io/projected/0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3-kube-api-access-k6dgs\") pod \"0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3\" (UID: \"0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3\") " Dec 07 16:41:07 crc kubenswrapper[4716]: I1207 16:41:07.898863 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3-inventory\") pod \"0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3\" (UID: \"0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3\") " Dec 07 16:41:07 crc kubenswrapper[4716]: I1207 16:41:07.898911 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3-libvirt-secret-0\") pod \"0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3\" (UID: \"0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3\") " Dec 07 16:41:07 crc kubenswrapper[4716]: I1207 16:41:07.898998 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3-libvirt-combined-ca-bundle\") pod \"0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3\" (UID: \"0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3\") " Dec 07 16:41:07 crc kubenswrapper[4716]: I1207 16:41:07.914965 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3-kube-api-access-k6dgs" (OuterVolumeSpecName: "kube-api-access-k6dgs") pod "0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3" (UID: "0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3"). InnerVolumeSpecName "kube-api-access-k6dgs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:41:07 crc kubenswrapper[4716]: I1207 16:41:07.927528 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3" (UID: "0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:41:07 crc kubenswrapper[4716]: I1207 16:41:07.937419 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3" (UID: "0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:41:07 crc kubenswrapper[4716]: E1207 16:41:07.946444 4716 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3-ssh-key podName:0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3 nodeName:}" failed. No retries permitted until 2025-12-07 16:41:08.446411603 +0000 UTC m=+2331.136696515 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "ssh-key" (UniqueName: "kubernetes.io/secret/0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3-ssh-key") pod "0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3" (UID: "0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3") : error deleting /var/lib/kubelet/pods/0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3/volume-subpaths: remove /var/lib/kubelet/pods/0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3/volume-subpaths: no such file or directory Dec 07 16:41:07 crc kubenswrapper[4716]: I1207 16:41:07.950264 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3-inventory" (OuterVolumeSpecName: "inventory") pod "0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3" (UID: "0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:41:08 crc kubenswrapper[4716]: I1207 16:41:08.001743 4716 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Dec 07 16:41:08 crc kubenswrapper[4716]: I1207 16:41:08.001967 4716 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 16:41:08 crc kubenswrapper[4716]: I1207 16:41:08.002178 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k6dgs\" (UniqueName: \"kubernetes.io/projected/0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3-kube-api-access-k6dgs\") on node \"crc\" DevicePath \"\"" Dec 07 16:41:08 crc kubenswrapper[4716]: I1207 16:41:08.002305 4716 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3-inventory\") on node \"crc\" DevicePath \"\"" Dec 07 16:41:08 crc kubenswrapper[4716]: I1207 16:41:08.419254 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-66mjc" event={"ID":"0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3","Type":"ContainerDied","Data":"7f5eb2b63832959062133120af8bc4d2abfaddb97ce9609dc90b4fa26d0ff39c"} Dec 07 16:41:08 crc kubenswrapper[4716]: I1207 16:41:08.419292 4716 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7f5eb2b63832959062133120af8bc4d2abfaddb97ce9609dc90b4fa26d0ff39c" Dec 07 16:41:08 crc kubenswrapper[4716]: I1207 16:41:08.419303 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-66mjc" Dec 07 16:41:08 crc kubenswrapper[4716]: I1207 16:41:08.512394 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3-ssh-key\") pod \"0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3\" (UID: \"0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3\") " Dec 07 16:41:08 crc kubenswrapper[4716]: I1207 16:41:08.517256 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3" (UID: "0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:41:08 crc kubenswrapper[4716]: I1207 16:41:08.519390 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-8jm5n"] Dec 07 16:41:08 crc kubenswrapper[4716]: E1207 16:41:08.520060 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03951083-dea9-4d8c-a83d-4dc0fc99cbff" containerName="registry-server" Dec 07 16:41:08 crc kubenswrapper[4716]: I1207 16:41:08.520094 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="03951083-dea9-4d8c-a83d-4dc0fc99cbff" containerName="registry-server" Dec 07 16:41:08 crc kubenswrapper[4716]: E1207 16:41:08.520109 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c0ad3b2-97b0-4c37-85da-ab61e86ef418" containerName="extract-utilities" Dec 07 16:41:08 crc kubenswrapper[4716]: I1207 16:41:08.520117 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c0ad3b2-97b0-4c37-85da-ab61e86ef418" containerName="extract-utilities" Dec 07 16:41:08 crc kubenswrapper[4716]: E1207 16:41:08.520126 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f30ae9ca-c2da-4ed4-89ac-55970d000fd8" containerName="extract-utilities" Dec 07 16:41:08 crc kubenswrapper[4716]: I1207 16:41:08.520134 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="f30ae9ca-c2da-4ed4-89ac-55970d000fd8" containerName="extract-utilities" Dec 07 16:41:08 crc kubenswrapper[4716]: E1207 16:41:08.520149 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f30ae9ca-c2da-4ed4-89ac-55970d000fd8" containerName="extract-content" Dec 07 16:41:08 crc kubenswrapper[4716]: I1207 16:41:08.520156 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="f30ae9ca-c2da-4ed4-89ac-55970d000fd8" containerName="extract-content" Dec 07 16:41:08 crc kubenswrapper[4716]: E1207 16:41:08.520180 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c0ad3b2-97b0-4c37-85da-ab61e86ef418" containerName="registry-server" Dec 07 16:41:08 crc kubenswrapper[4716]: I1207 16:41:08.520187 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c0ad3b2-97b0-4c37-85da-ab61e86ef418" containerName="registry-server" Dec 07 16:41:08 crc kubenswrapper[4716]: E1207 16:41:08.520205 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c0ad3b2-97b0-4c37-85da-ab61e86ef418" containerName="extract-content" Dec 07 16:41:08 crc kubenswrapper[4716]: I1207 16:41:08.520213 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c0ad3b2-97b0-4c37-85da-ab61e86ef418" containerName="extract-content" Dec 07 16:41:08 crc kubenswrapper[4716]: E1207 16:41:08.520231 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03951083-dea9-4d8c-a83d-4dc0fc99cbff" containerName="extract-content" Dec 07 16:41:08 crc kubenswrapper[4716]: I1207 16:41:08.520237 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="03951083-dea9-4d8c-a83d-4dc0fc99cbff" containerName="extract-content" Dec 07 16:41:08 crc kubenswrapper[4716]: E1207 16:41:08.520262 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f30ae9ca-c2da-4ed4-89ac-55970d000fd8" containerName="registry-server" Dec 07 16:41:08 crc kubenswrapper[4716]: I1207 16:41:08.520269 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="f30ae9ca-c2da-4ed4-89ac-55970d000fd8" containerName="registry-server" Dec 07 16:41:08 crc kubenswrapper[4716]: E1207 16:41:08.520279 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 07 16:41:08 crc kubenswrapper[4716]: I1207 16:41:08.520288 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 07 16:41:08 crc kubenswrapper[4716]: E1207 16:41:08.520301 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03951083-dea9-4d8c-a83d-4dc0fc99cbff" containerName="extract-utilities" Dec 07 16:41:08 crc kubenswrapper[4716]: I1207 16:41:08.520308 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="03951083-dea9-4d8c-a83d-4dc0fc99cbff" containerName="extract-utilities" Dec 07 16:41:08 crc kubenswrapper[4716]: I1207 16:41:08.520512 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="03951083-dea9-4d8c-a83d-4dc0fc99cbff" containerName="registry-server" Dec 07 16:41:08 crc kubenswrapper[4716]: I1207 16:41:08.520531 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c0ad3b2-97b0-4c37-85da-ab61e86ef418" containerName="registry-server" Dec 07 16:41:08 crc kubenswrapper[4716]: I1207 16:41:08.520545 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 07 16:41:08 crc kubenswrapper[4716]: I1207 16:41:08.520571 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="f30ae9ca-c2da-4ed4-89ac-55970d000fd8" containerName="registry-server" Dec 07 16:41:08 crc kubenswrapper[4716]: I1207 16:41:08.521564 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8jm5n" Dec 07 16:41:08 crc kubenswrapper[4716]: I1207 16:41:08.523526 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Dec 07 16:41:08 crc kubenswrapper[4716]: I1207 16:41:08.524103 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Dec 07 16:41:08 crc kubenswrapper[4716]: I1207 16:41:08.524925 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Dec 07 16:41:08 crc kubenswrapper[4716]: I1207 16:41:08.532020 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-8jm5n"] Dec 07 16:41:08 crc kubenswrapper[4716]: I1207 16:41:08.614576 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/01472d87-913d-4565-8d83-40966b88a630-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8jm5n\" (UID: \"01472d87-913d-4565-8d83-40966b88a630\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8jm5n" Dec 07 16:41:08 crc kubenswrapper[4716]: I1207 16:41:08.614633 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/01472d87-913d-4565-8d83-40966b88a630-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8jm5n\" (UID: \"01472d87-913d-4565-8d83-40966b88a630\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8jm5n" Dec 07 16:41:08 crc kubenswrapper[4716]: I1207 16:41:08.614662 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/01472d87-913d-4565-8d83-40966b88a630-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8jm5n\" (UID: \"01472d87-913d-4565-8d83-40966b88a630\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8jm5n" Dec 07 16:41:08 crc kubenswrapper[4716]: I1207 16:41:08.614700 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/01472d87-913d-4565-8d83-40966b88a630-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8jm5n\" (UID: \"01472d87-913d-4565-8d83-40966b88a630\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8jm5n" Dec 07 16:41:08 crc kubenswrapper[4716]: I1207 16:41:08.614757 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bnfj9\" (UniqueName: \"kubernetes.io/projected/01472d87-913d-4565-8d83-40966b88a630-kube-api-access-bnfj9\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8jm5n\" (UID: \"01472d87-913d-4565-8d83-40966b88a630\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8jm5n" Dec 07 16:41:08 crc kubenswrapper[4716]: I1207 16:41:08.614789 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01472d87-913d-4565-8d83-40966b88a630-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8jm5n\" (UID: \"01472d87-913d-4565-8d83-40966b88a630\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8jm5n" Dec 07 16:41:08 crc kubenswrapper[4716]: I1207 16:41:08.614812 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/01472d87-913d-4565-8d83-40966b88a630-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8jm5n\" (UID: \"01472d87-913d-4565-8d83-40966b88a630\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8jm5n" Dec 07 16:41:08 crc kubenswrapper[4716]: I1207 16:41:08.614843 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/01472d87-913d-4565-8d83-40966b88a630-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8jm5n\" (UID: \"01472d87-913d-4565-8d83-40966b88a630\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8jm5n" Dec 07 16:41:08 crc kubenswrapper[4716]: I1207 16:41:08.614891 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/01472d87-913d-4565-8d83-40966b88a630-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8jm5n\" (UID: \"01472d87-913d-4565-8d83-40966b88a630\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8jm5n" Dec 07 16:41:08 crc kubenswrapper[4716]: I1207 16:41:08.615112 4716 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 07 16:41:08 crc kubenswrapper[4716]: I1207 16:41:08.716161 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bnfj9\" (UniqueName: \"kubernetes.io/projected/01472d87-913d-4565-8d83-40966b88a630-kube-api-access-bnfj9\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8jm5n\" (UID: \"01472d87-913d-4565-8d83-40966b88a630\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8jm5n" Dec 07 16:41:08 crc kubenswrapper[4716]: I1207 16:41:08.716471 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01472d87-913d-4565-8d83-40966b88a630-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8jm5n\" (UID: \"01472d87-913d-4565-8d83-40966b88a630\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8jm5n" Dec 07 16:41:08 crc kubenswrapper[4716]: I1207 16:41:08.716557 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/01472d87-913d-4565-8d83-40966b88a630-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8jm5n\" (UID: \"01472d87-913d-4565-8d83-40966b88a630\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8jm5n" Dec 07 16:41:08 crc kubenswrapper[4716]: I1207 16:41:08.716647 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/01472d87-913d-4565-8d83-40966b88a630-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8jm5n\" (UID: \"01472d87-913d-4565-8d83-40966b88a630\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8jm5n" Dec 07 16:41:08 crc kubenswrapper[4716]: I1207 16:41:08.716746 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/01472d87-913d-4565-8d83-40966b88a630-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8jm5n\" (UID: \"01472d87-913d-4565-8d83-40966b88a630\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8jm5n" Dec 07 16:41:08 crc kubenswrapper[4716]: I1207 16:41:08.716868 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/01472d87-913d-4565-8d83-40966b88a630-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8jm5n\" (UID: \"01472d87-913d-4565-8d83-40966b88a630\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8jm5n" Dec 07 16:41:08 crc kubenswrapper[4716]: I1207 16:41:08.716939 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/01472d87-913d-4565-8d83-40966b88a630-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8jm5n\" (UID: \"01472d87-913d-4565-8d83-40966b88a630\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8jm5n" Dec 07 16:41:08 crc kubenswrapper[4716]: I1207 16:41:08.717015 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/01472d87-913d-4565-8d83-40966b88a630-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8jm5n\" (UID: \"01472d87-913d-4565-8d83-40966b88a630\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8jm5n" Dec 07 16:41:08 crc kubenswrapper[4716]: I1207 16:41:08.717121 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/01472d87-913d-4565-8d83-40966b88a630-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8jm5n\" (UID: \"01472d87-913d-4565-8d83-40966b88a630\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8jm5n" Dec 07 16:41:08 crc kubenswrapper[4716]: I1207 16:41:08.718032 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/01472d87-913d-4565-8d83-40966b88a630-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8jm5n\" (UID: \"01472d87-913d-4565-8d83-40966b88a630\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8jm5n" Dec 07 16:41:08 crc kubenswrapper[4716]: I1207 16:41:08.723474 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/01472d87-913d-4565-8d83-40966b88a630-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8jm5n\" (UID: \"01472d87-913d-4565-8d83-40966b88a630\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8jm5n" Dec 07 16:41:08 crc kubenswrapper[4716]: I1207 16:41:08.723991 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/01472d87-913d-4565-8d83-40966b88a630-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8jm5n\" (UID: \"01472d87-913d-4565-8d83-40966b88a630\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8jm5n" Dec 07 16:41:08 crc kubenswrapper[4716]: I1207 16:41:08.724866 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/01472d87-913d-4565-8d83-40966b88a630-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8jm5n\" (UID: \"01472d87-913d-4565-8d83-40966b88a630\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8jm5n" Dec 07 16:41:08 crc kubenswrapper[4716]: I1207 16:41:08.724990 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/01472d87-913d-4565-8d83-40966b88a630-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8jm5n\" (UID: \"01472d87-913d-4565-8d83-40966b88a630\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8jm5n" Dec 07 16:41:08 crc kubenswrapper[4716]: I1207 16:41:08.726319 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/01472d87-913d-4565-8d83-40966b88a630-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8jm5n\" (UID: \"01472d87-913d-4565-8d83-40966b88a630\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8jm5n" Dec 07 16:41:08 crc kubenswrapper[4716]: I1207 16:41:08.726531 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/01472d87-913d-4565-8d83-40966b88a630-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8jm5n\" (UID: \"01472d87-913d-4565-8d83-40966b88a630\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8jm5n" Dec 07 16:41:08 crc kubenswrapper[4716]: I1207 16:41:08.728288 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01472d87-913d-4565-8d83-40966b88a630-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8jm5n\" (UID: \"01472d87-913d-4565-8d83-40966b88a630\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8jm5n" Dec 07 16:41:08 crc kubenswrapper[4716]: I1207 16:41:08.740986 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bnfj9\" (UniqueName: \"kubernetes.io/projected/01472d87-913d-4565-8d83-40966b88a630-kube-api-access-bnfj9\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8jm5n\" (UID: \"01472d87-913d-4565-8d83-40966b88a630\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8jm5n" Dec 07 16:41:08 crc kubenswrapper[4716]: I1207 16:41:08.885441 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8jm5n" Dec 07 16:41:09 crc kubenswrapper[4716]: I1207 16:41:09.399707 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-8jm5n"] Dec 07 16:41:09 crc kubenswrapper[4716]: I1207 16:41:09.401783 4716 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 07 16:41:09 crc kubenswrapper[4716]: I1207 16:41:09.429031 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8jm5n" event={"ID":"01472d87-913d-4565-8d83-40966b88a630","Type":"ContainerStarted","Data":"e3e9bd09d057efe377eb51db361f9034c04cf3a9e32aab503c4e70582b4f40c3"} Dec 07 16:41:10 crc kubenswrapper[4716]: I1207 16:41:10.440095 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8jm5n" event={"ID":"01472d87-913d-4565-8d83-40966b88a630","Type":"ContainerStarted","Data":"b899f821798054c6ba0a01aa6cb58b3e2c9c6c18193b927c8e5a1de898a0538e"} Dec 07 16:41:10 crc kubenswrapper[4716]: I1207 16:41:10.462539 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8jm5n" podStartSLOduration=1.960639484 podStartE2EDuration="2.462518856s" podCreationTimestamp="2025-12-07 16:41:08 +0000 UTC" firstStartedPulling="2025-12-07 16:41:09.401568689 +0000 UTC m=+2332.091853601" lastFinishedPulling="2025-12-07 16:41:09.903448061 +0000 UTC m=+2332.593732973" observedRunningTime="2025-12-07 16:41:10.455050185 +0000 UTC m=+2333.145335097" watchObservedRunningTime="2025-12-07 16:41:10.462518856 +0000 UTC m=+2333.152803768" Dec 07 16:41:22 crc kubenswrapper[4716]: I1207 16:41:22.761162 4716 patch_prober.go:28] interesting pod/machine-config-daemon-zcxxp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 16:41:22 crc kubenswrapper[4716]: I1207 16:41:22.762021 4716 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 16:41:52 crc kubenswrapper[4716]: I1207 16:41:52.761167 4716 patch_prober.go:28] interesting pod/machine-config-daemon-zcxxp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 16:41:52 crc kubenswrapper[4716]: I1207 16:41:52.761924 4716 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 16:42:22 crc kubenswrapper[4716]: I1207 16:42:22.761523 4716 patch_prober.go:28] interesting pod/machine-config-daemon-zcxxp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 16:42:22 crc kubenswrapper[4716]: I1207 16:42:22.762296 4716 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 16:42:22 crc kubenswrapper[4716]: I1207 16:42:22.762355 4716 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" Dec 07 16:42:22 crc kubenswrapper[4716]: I1207 16:42:22.763415 4716 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3eb63b7962def203a31ba32ba16ac93470e254c2a4edb4a4dd2033b290e2f217"} pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 07 16:42:22 crc kubenswrapper[4716]: I1207 16:42:22.763488 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" containerName="machine-config-daemon" containerID="cri-o://3eb63b7962def203a31ba32ba16ac93470e254c2a4edb4a4dd2033b290e2f217" gracePeriod=600 Dec 07 16:42:22 crc kubenswrapper[4716]: E1207 16:42:22.891681 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 16:42:23 crc kubenswrapper[4716]: I1207 16:42:23.165748 4716 generic.go:334] "Generic (PLEG): container finished" podID="c15b59eb-565d-4556-a4ce-75afdf159dc8" containerID="3eb63b7962def203a31ba32ba16ac93470e254c2a4edb4a4dd2033b290e2f217" exitCode=0 Dec 07 16:42:23 crc kubenswrapper[4716]: I1207 16:42:23.165841 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" event={"ID":"c15b59eb-565d-4556-a4ce-75afdf159dc8","Type":"ContainerDied","Data":"3eb63b7962def203a31ba32ba16ac93470e254c2a4edb4a4dd2033b290e2f217"} Dec 07 16:42:23 crc kubenswrapper[4716]: I1207 16:42:23.166177 4716 scope.go:117] "RemoveContainer" containerID="8aa74b5db626ce69c5f618707bfc1aab2b0242d5c7b440dfc5c0166d72b0d08c" Dec 07 16:42:23 crc kubenswrapper[4716]: I1207 16:42:23.166929 4716 scope.go:117] "RemoveContainer" containerID="3eb63b7962def203a31ba32ba16ac93470e254c2a4edb4a4dd2033b290e2f217" Dec 07 16:42:23 crc kubenswrapper[4716]: E1207 16:42:23.167290 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 16:42:36 crc kubenswrapper[4716]: I1207 16:42:36.659056 4716 scope.go:117] "RemoveContainer" containerID="3eb63b7962def203a31ba32ba16ac93470e254c2a4edb4a4dd2033b290e2f217" Dec 07 16:42:36 crc kubenswrapper[4716]: E1207 16:42:36.661060 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 16:42:48 crc kubenswrapper[4716]: I1207 16:42:48.657815 4716 scope.go:117] "RemoveContainer" containerID="3eb63b7962def203a31ba32ba16ac93470e254c2a4edb4a4dd2033b290e2f217" Dec 07 16:42:48 crc kubenswrapper[4716]: E1207 16:42:48.658726 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 16:42:59 crc kubenswrapper[4716]: I1207 16:42:59.658072 4716 scope.go:117] "RemoveContainer" containerID="3eb63b7962def203a31ba32ba16ac93470e254c2a4edb4a4dd2033b290e2f217" Dec 07 16:42:59 crc kubenswrapper[4716]: E1207 16:42:59.658980 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 16:43:14 crc kubenswrapper[4716]: I1207 16:43:14.663792 4716 scope.go:117] "RemoveContainer" containerID="3eb63b7962def203a31ba32ba16ac93470e254c2a4edb4a4dd2033b290e2f217" Dec 07 16:43:14 crc kubenswrapper[4716]: E1207 16:43:14.665783 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 16:43:29 crc kubenswrapper[4716]: I1207 16:43:29.658274 4716 scope.go:117] "RemoveContainer" containerID="3eb63b7962def203a31ba32ba16ac93470e254c2a4edb4a4dd2033b290e2f217" Dec 07 16:43:29 crc kubenswrapper[4716]: E1207 16:43:29.659280 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 16:43:41 crc kubenswrapper[4716]: I1207 16:43:41.657815 4716 scope.go:117] "RemoveContainer" containerID="3eb63b7962def203a31ba32ba16ac93470e254c2a4edb4a4dd2033b290e2f217" Dec 07 16:43:41 crc kubenswrapper[4716]: E1207 16:43:41.658775 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 16:43:56 crc kubenswrapper[4716]: I1207 16:43:56.658221 4716 scope.go:117] "RemoveContainer" containerID="3eb63b7962def203a31ba32ba16ac93470e254c2a4edb4a4dd2033b290e2f217" Dec 07 16:43:56 crc kubenswrapper[4716]: E1207 16:43:56.659177 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 16:44:11 crc kubenswrapper[4716]: I1207 16:44:11.657771 4716 scope.go:117] "RemoveContainer" containerID="3eb63b7962def203a31ba32ba16ac93470e254c2a4edb4a4dd2033b290e2f217" Dec 07 16:44:11 crc kubenswrapper[4716]: E1207 16:44:11.658747 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 16:44:19 crc kubenswrapper[4716]: I1207 16:44:19.232818 4716 generic.go:334] "Generic (PLEG): container finished" podID="01472d87-913d-4565-8d83-40966b88a630" containerID="b899f821798054c6ba0a01aa6cb58b3e2c9c6c18193b927c8e5a1de898a0538e" exitCode=0 Dec 07 16:44:19 crc kubenswrapper[4716]: I1207 16:44:19.232923 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8jm5n" event={"ID":"01472d87-913d-4565-8d83-40966b88a630","Type":"ContainerDied","Data":"b899f821798054c6ba0a01aa6cb58b3e2c9c6c18193b927c8e5a1de898a0538e"} Dec 07 16:44:20 crc kubenswrapper[4716]: I1207 16:44:20.692863 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8jm5n" Dec 07 16:44:20 crc kubenswrapper[4716]: I1207 16:44:20.752756 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01472d87-913d-4565-8d83-40966b88a630-nova-combined-ca-bundle\") pod \"01472d87-913d-4565-8d83-40966b88a630\" (UID: \"01472d87-913d-4565-8d83-40966b88a630\") " Dec 07 16:44:20 crc kubenswrapper[4716]: I1207 16:44:20.752810 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/01472d87-913d-4565-8d83-40966b88a630-nova-cell1-compute-config-1\") pod \"01472d87-913d-4565-8d83-40966b88a630\" (UID: \"01472d87-913d-4565-8d83-40966b88a630\") " Dec 07 16:44:20 crc kubenswrapper[4716]: I1207 16:44:20.752858 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/01472d87-913d-4565-8d83-40966b88a630-inventory\") pod \"01472d87-913d-4565-8d83-40966b88a630\" (UID: \"01472d87-913d-4565-8d83-40966b88a630\") " Dec 07 16:44:20 crc kubenswrapper[4716]: I1207 16:44:20.752940 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/01472d87-913d-4565-8d83-40966b88a630-nova-extra-config-0\") pod \"01472d87-913d-4565-8d83-40966b88a630\" (UID: \"01472d87-913d-4565-8d83-40966b88a630\") " Dec 07 16:44:20 crc kubenswrapper[4716]: I1207 16:44:20.753779 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/01472d87-913d-4565-8d83-40966b88a630-nova-cell1-compute-config-0\") pod \"01472d87-913d-4565-8d83-40966b88a630\" (UID: \"01472d87-913d-4565-8d83-40966b88a630\") " Dec 07 16:44:20 crc kubenswrapper[4716]: I1207 16:44:20.753804 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/01472d87-913d-4565-8d83-40966b88a630-ssh-key\") pod \"01472d87-913d-4565-8d83-40966b88a630\" (UID: \"01472d87-913d-4565-8d83-40966b88a630\") " Dec 07 16:44:20 crc kubenswrapper[4716]: I1207 16:44:20.753822 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/01472d87-913d-4565-8d83-40966b88a630-nova-migration-ssh-key-1\") pod \"01472d87-913d-4565-8d83-40966b88a630\" (UID: \"01472d87-913d-4565-8d83-40966b88a630\") " Dec 07 16:44:20 crc kubenswrapper[4716]: I1207 16:44:20.753872 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bnfj9\" (UniqueName: \"kubernetes.io/projected/01472d87-913d-4565-8d83-40966b88a630-kube-api-access-bnfj9\") pod \"01472d87-913d-4565-8d83-40966b88a630\" (UID: \"01472d87-913d-4565-8d83-40966b88a630\") " Dec 07 16:44:20 crc kubenswrapper[4716]: I1207 16:44:20.753928 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/01472d87-913d-4565-8d83-40966b88a630-nova-migration-ssh-key-0\") pod \"01472d87-913d-4565-8d83-40966b88a630\" (UID: \"01472d87-913d-4565-8d83-40966b88a630\") " Dec 07 16:44:20 crc kubenswrapper[4716]: I1207 16:44:20.759330 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01472d87-913d-4565-8d83-40966b88a630-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "01472d87-913d-4565-8d83-40966b88a630" (UID: "01472d87-913d-4565-8d83-40966b88a630"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:44:20 crc kubenswrapper[4716]: I1207 16:44:20.759397 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01472d87-913d-4565-8d83-40966b88a630-kube-api-access-bnfj9" (OuterVolumeSpecName: "kube-api-access-bnfj9") pod "01472d87-913d-4565-8d83-40966b88a630" (UID: "01472d87-913d-4565-8d83-40966b88a630"). InnerVolumeSpecName "kube-api-access-bnfj9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:44:20 crc kubenswrapper[4716]: I1207 16:44:20.784096 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01472d87-913d-4565-8d83-40966b88a630-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "01472d87-913d-4565-8d83-40966b88a630" (UID: "01472d87-913d-4565-8d83-40966b88a630"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:44:20 crc kubenswrapper[4716]: I1207 16:44:20.786143 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01472d87-913d-4565-8d83-40966b88a630-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "01472d87-913d-4565-8d83-40966b88a630" (UID: "01472d87-913d-4565-8d83-40966b88a630"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:44:20 crc kubenswrapper[4716]: I1207 16:44:20.787219 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01472d87-913d-4565-8d83-40966b88a630-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "01472d87-913d-4565-8d83-40966b88a630" (UID: "01472d87-913d-4565-8d83-40966b88a630"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:44:20 crc kubenswrapper[4716]: I1207 16:44:20.788366 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01472d87-913d-4565-8d83-40966b88a630-inventory" (OuterVolumeSpecName: "inventory") pod "01472d87-913d-4565-8d83-40966b88a630" (UID: "01472d87-913d-4565-8d83-40966b88a630"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:44:20 crc kubenswrapper[4716]: I1207 16:44:20.790699 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01472d87-913d-4565-8d83-40966b88a630-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "01472d87-913d-4565-8d83-40966b88a630" (UID: "01472d87-913d-4565-8d83-40966b88a630"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:44:20 crc kubenswrapper[4716]: I1207 16:44:20.792138 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01472d87-913d-4565-8d83-40966b88a630-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "01472d87-913d-4565-8d83-40966b88a630" (UID: "01472d87-913d-4565-8d83-40966b88a630"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:44:20 crc kubenswrapper[4716]: I1207 16:44:20.806208 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01472d87-913d-4565-8d83-40966b88a630-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "01472d87-913d-4565-8d83-40966b88a630" (UID: "01472d87-913d-4565-8d83-40966b88a630"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:44:20 crc kubenswrapper[4716]: I1207 16:44:20.855732 4716 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/01472d87-913d-4565-8d83-40966b88a630-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Dec 07 16:44:20 crc kubenswrapper[4716]: I1207 16:44:20.855795 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bnfj9\" (UniqueName: \"kubernetes.io/projected/01472d87-913d-4565-8d83-40966b88a630-kube-api-access-bnfj9\") on node \"crc\" DevicePath \"\"" Dec 07 16:44:20 crc kubenswrapper[4716]: I1207 16:44:20.855804 4716 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/01472d87-913d-4565-8d83-40966b88a630-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Dec 07 16:44:20 crc kubenswrapper[4716]: I1207 16:44:20.855813 4716 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01472d87-913d-4565-8d83-40966b88a630-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 16:44:20 crc kubenswrapper[4716]: I1207 16:44:20.855822 4716 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/01472d87-913d-4565-8d83-40966b88a630-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Dec 07 16:44:20 crc kubenswrapper[4716]: I1207 16:44:20.855831 4716 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/01472d87-913d-4565-8d83-40966b88a630-inventory\") on node \"crc\" DevicePath \"\"" Dec 07 16:44:20 crc kubenswrapper[4716]: I1207 16:44:20.855840 4716 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/01472d87-913d-4565-8d83-40966b88a630-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Dec 07 16:44:20 crc kubenswrapper[4716]: I1207 16:44:20.855849 4716 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/01472d87-913d-4565-8d83-40966b88a630-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Dec 07 16:44:20 crc kubenswrapper[4716]: I1207 16:44:20.855857 4716 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/01472d87-913d-4565-8d83-40966b88a630-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 07 16:44:21 crc kubenswrapper[4716]: I1207 16:44:21.260207 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8jm5n" event={"ID":"01472d87-913d-4565-8d83-40966b88a630","Type":"ContainerDied","Data":"e3e9bd09d057efe377eb51db361f9034c04cf3a9e32aab503c4e70582b4f40c3"} Dec 07 16:44:21 crc kubenswrapper[4716]: I1207 16:44:21.260255 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8jm5n" Dec 07 16:44:21 crc kubenswrapper[4716]: I1207 16:44:21.260280 4716 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e3e9bd09d057efe377eb51db361f9034c04cf3a9e32aab503c4e70582b4f40c3" Dec 07 16:44:21 crc kubenswrapper[4716]: I1207 16:44:21.369696 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-69qz6"] Dec 07 16:44:21 crc kubenswrapper[4716]: E1207 16:44:21.370183 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01472d87-913d-4565-8d83-40966b88a630" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 07 16:44:21 crc kubenswrapper[4716]: I1207 16:44:21.370206 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="01472d87-913d-4565-8d83-40966b88a630" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 07 16:44:21 crc kubenswrapper[4716]: I1207 16:44:21.370431 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="01472d87-913d-4565-8d83-40966b88a630" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 07 16:44:21 crc kubenswrapper[4716]: I1207 16:44:21.371030 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-69qz6" Dec 07 16:44:21 crc kubenswrapper[4716]: I1207 16:44:21.373647 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Dec 07 16:44:21 crc kubenswrapper[4716]: I1207 16:44:21.373874 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 07 16:44:21 crc kubenswrapper[4716]: I1207 16:44:21.374224 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 07 16:44:21 crc kubenswrapper[4716]: I1207 16:44:21.374474 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 07 16:44:21 crc kubenswrapper[4716]: I1207 16:44:21.378558 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-h9pt2" Dec 07 16:44:21 crc kubenswrapper[4716]: I1207 16:44:21.384310 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-69qz6"] Dec 07 16:44:21 crc kubenswrapper[4716]: I1207 16:44:21.463923 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fct8b\" (UniqueName: \"kubernetes.io/projected/c9735df5-27ef-41b1-84c1-4cab55f23b3a-kube-api-access-fct8b\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-69qz6\" (UID: \"c9735df5-27ef-41b1-84c1-4cab55f23b3a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-69qz6" Dec 07 16:44:21 crc kubenswrapper[4716]: I1207 16:44:21.464012 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/c9735df5-27ef-41b1-84c1-4cab55f23b3a-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-69qz6\" (UID: \"c9735df5-27ef-41b1-84c1-4cab55f23b3a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-69qz6" Dec 07 16:44:21 crc kubenswrapper[4716]: I1207 16:44:21.464044 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c9735df5-27ef-41b1-84c1-4cab55f23b3a-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-69qz6\" (UID: \"c9735df5-27ef-41b1-84c1-4cab55f23b3a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-69qz6" Dec 07 16:44:21 crc kubenswrapper[4716]: I1207 16:44:21.464108 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9735df5-27ef-41b1-84c1-4cab55f23b3a-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-69qz6\" (UID: \"c9735df5-27ef-41b1-84c1-4cab55f23b3a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-69qz6" Dec 07 16:44:21 crc kubenswrapper[4716]: I1207 16:44:21.464133 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/c9735df5-27ef-41b1-84c1-4cab55f23b3a-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-69qz6\" (UID: \"c9735df5-27ef-41b1-84c1-4cab55f23b3a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-69qz6" Dec 07 16:44:21 crc kubenswrapper[4716]: I1207 16:44:21.464197 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/c9735df5-27ef-41b1-84c1-4cab55f23b3a-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-69qz6\" (UID: \"c9735df5-27ef-41b1-84c1-4cab55f23b3a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-69qz6" Dec 07 16:44:21 crc kubenswrapper[4716]: I1207 16:44:21.464224 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c9735df5-27ef-41b1-84c1-4cab55f23b3a-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-69qz6\" (UID: \"c9735df5-27ef-41b1-84c1-4cab55f23b3a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-69qz6" Dec 07 16:44:21 crc kubenswrapper[4716]: I1207 16:44:21.564955 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fct8b\" (UniqueName: \"kubernetes.io/projected/c9735df5-27ef-41b1-84c1-4cab55f23b3a-kube-api-access-fct8b\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-69qz6\" (UID: \"c9735df5-27ef-41b1-84c1-4cab55f23b3a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-69qz6" Dec 07 16:44:21 crc kubenswrapper[4716]: I1207 16:44:21.565058 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/c9735df5-27ef-41b1-84c1-4cab55f23b3a-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-69qz6\" (UID: \"c9735df5-27ef-41b1-84c1-4cab55f23b3a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-69qz6" Dec 07 16:44:21 crc kubenswrapper[4716]: I1207 16:44:21.565113 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c9735df5-27ef-41b1-84c1-4cab55f23b3a-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-69qz6\" (UID: \"c9735df5-27ef-41b1-84c1-4cab55f23b3a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-69qz6" Dec 07 16:44:21 crc kubenswrapper[4716]: I1207 16:44:21.565166 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9735df5-27ef-41b1-84c1-4cab55f23b3a-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-69qz6\" (UID: \"c9735df5-27ef-41b1-84c1-4cab55f23b3a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-69qz6" Dec 07 16:44:21 crc kubenswrapper[4716]: I1207 16:44:21.565194 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/c9735df5-27ef-41b1-84c1-4cab55f23b3a-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-69qz6\" (UID: \"c9735df5-27ef-41b1-84c1-4cab55f23b3a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-69qz6" Dec 07 16:44:21 crc kubenswrapper[4716]: I1207 16:44:21.565261 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/c9735df5-27ef-41b1-84c1-4cab55f23b3a-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-69qz6\" (UID: \"c9735df5-27ef-41b1-84c1-4cab55f23b3a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-69qz6" Dec 07 16:44:21 crc kubenswrapper[4716]: I1207 16:44:21.565289 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c9735df5-27ef-41b1-84c1-4cab55f23b3a-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-69qz6\" (UID: \"c9735df5-27ef-41b1-84c1-4cab55f23b3a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-69qz6" Dec 07 16:44:21 crc kubenswrapper[4716]: I1207 16:44:21.570726 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/c9735df5-27ef-41b1-84c1-4cab55f23b3a-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-69qz6\" (UID: \"c9735df5-27ef-41b1-84c1-4cab55f23b3a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-69qz6" Dec 07 16:44:21 crc kubenswrapper[4716]: I1207 16:44:21.570944 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9735df5-27ef-41b1-84c1-4cab55f23b3a-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-69qz6\" (UID: \"c9735df5-27ef-41b1-84c1-4cab55f23b3a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-69qz6" Dec 07 16:44:21 crc kubenswrapper[4716]: I1207 16:44:21.571098 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c9735df5-27ef-41b1-84c1-4cab55f23b3a-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-69qz6\" (UID: \"c9735df5-27ef-41b1-84c1-4cab55f23b3a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-69qz6" Dec 07 16:44:21 crc kubenswrapper[4716]: I1207 16:44:21.587120 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/c9735df5-27ef-41b1-84c1-4cab55f23b3a-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-69qz6\" (UID: \"c9735df5-27ef-41b1-84c1-4cab55f23b3a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-69qz6" Dec 07 16:44:21 crc kubenswrapper[4716]: I1207 16:44:21.587709 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c9735df5-27ef-41b1-84c1-4cab55f23b3a-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-69qz6\" (UID: \"c9735df5-27ef-41b1-84c1-4cab55f23b3a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-69qz6" Dec 07 16:44:21 crc kubenswrapper[4716]: I1207 16:44:21.589989 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/c9735df5-27ef-41b1-84c1-4cab55f23b3a-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-69qz6\" (UID: \"c9735df5-27ef-41b1-84c1-4cab55f23b3a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-69qz6" Dec 07 16:44:21 crc kubenswrapper[4716]: I1207 16:44:21.594808 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fct8b\" (UniqueName: \"kubernetes.io/projected/c9735df5-27ef-41b1-84c1-4cab55f23b3a-kube-api-access-fct8b\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-69qz6\" (UID: \"c9735df5-27ef-41b1-84c1-4cab55f23b3a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-69qz6" Dec 07 16:44:21 crc kubenswrapper[4716]: I1207 16:44:21.699131 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-69qz6" Dec 07 16:44:22 crc kubenswrapper[4716]: I1207 16:44:22.018445 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-69qz6"] Dec 07 16:44:22 crc kubenswrapper[4716]: I1207 16:44:22.269525 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-69qz6" event={"ID":"c9735df5-27ef-41b1-84c1-4cab55f23b3a","Type":"ContainerStarted","Data":"ebe10360ff3c44deaf764b47072771041b95118a0aa0d1e55124789d03b95271"} Dec 07 16:44:23 crc kubenswrapper[4716]: I1207 16:44:23.286165 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-69qz6" event={"ID":"c9735df5-27ef-41b1-84c1-4cab55f23b3a","Type":"ContainerStarted","Data":"f6eecc554d89c956aad8dbb5c308bc5e9c56a401f0a33d8a3a15e756910184d6"} Dec 07 16:44:23 crc kubenswrapper[4716]: I1207 16:44:23.312208 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-69qz6" podStartSLOduration=1.85881783 podStartE2EDuration="2.312182375s" podCreationTimestamp="2025-12-07 16:44:21 +0000 UTC" firstStartedPulling="2025-12-07 16:44:22.030252868 +0000 UTC m=+2524.720537780" lastFinishedPulling="2025-12-07 16:44:22.483617403 +0000 UTC m=+2525.173902325" observedRunningTime="2025-12-07 16:44:23.302721669 +0000 UTC m=+2525.993006581" watchObservedRunningTime="2025-12-07 16:44:23.312182375 +0000 UTC m=+2526.002467277" Dec 07 16:44:25 crc kubenswrapper[4716]: I1207 16:44:25.658389 4716 scope.go:117] "RemoveContainer" containerID="3eb63b7962def203a31ba32ba16ac93470e254c2a4edb4a4dd2033b290e2f217" Dec 07 16:44:25 crc kubenswrapper[4716]: E1207 16:44:25.659087 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 16:44:37 crc kubenswrapper[4716]: I1207 16:44:37.664058 4716 scope.go:117] "RemoveContainer" containerID="3eb63b7962def203a31ba32ba16ac93470e254c2a4edb4a4dd2033b290e2f217" Dec 07 16:44:37 crc kubenswrapper[4716]: E1207 16:44:37.664900 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 16:44:52 crc kubenswrapper[4716]: I1207 16:44:52.658235 4716 scope.go:117] "RemoveContainer" containerID="3eb63b7962def203a31ba32ba16ac93470e254c2a4edb4a4dd2033b290e2f217" Dec 07 16:44:52 crc kubenswrapper[4716]: E1207 16:44:52.659283 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 16:45:00 crc kubenswrapper[4716]: I1207 16:45:00.143028 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29418765-57x25"] Dec 07 16:45:00 crc kubenswrapper[4716]: I1207 16:45:00.145287 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29418765-57x25" Dec 07 16:45:00 crc kubenswrapper[4716]: I1207 16:45:00.147771 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 07 16:45:00 crc kubenswrapper[4716]: I1207 16:45:00.148826 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 07 16:45:00 crc kubenswrapper[4716]: I1207 16:45:00.158424 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29418765-57x25"] Dec 07 16:45:00 crc kubenswrapper[4716]: I1207 16:45:00.160022 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/93fa4e0c-38c0-4793-80b2-d39258b416e3-secret-volume\") pod \"collect-profiles-29418765-57x25\" (UID: \"93fa4e0c-38c0-4793-80b2-d39258b416e3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418765-57x25" Dec 07 16:45:00 crc kubenswrapper[4716]: I1207 16:45:00.160413 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5dm8n\" (UniqueName: \"kubernetes.io/projected/93fa4e0c-38c0-4793-80b2-d39258b416e3-kube-api-access-5dm8n\") pod \"collect-profiles-29418765-57x25\" (UID: \"93fa4e0c-38c0-4793-80b2-d39258b416e3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418765-57x25" Dec 07 16:45:00 crc kubenswrapper[4716]: I1207 16:45:00.160843 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/93fa4e0c-38c0-4793-80b2-d39258b416e3-config-volume\") pod \"collect-profiles-29418765-57x25\" (UID: \"93fa4e0c-38c0-4793-80b2-d39258b416e3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418765-57x25" Dec 07 16:45:00 crc kubenswrapper[4716]: I1207 16:45:00.262759 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/93fa4e0c-38c0-4793-80b2-d39258b416e3-config-volume\") pod \"collect-profiles-29418765-57x25\" (UID: \"93fa4e0c-38c0-4793-80b2-d39258b416e3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418765-57x25" Dec 07 16:45:00 crc kubenswrapper[4716]: I1207 16:45:00.262828 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/93fa4e0c-38c0-4793-80b2-d39258b416e3-secret-volume\") pod \"collect-profiles-29418765-57x25\" (UID: \"93fa4e0c-38c0-4793-80b2-d39258b416e3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418765-57x25" Dec 07 16:45:00 crc kubenswrapper[4716]: I1207 16:45:00.263054 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5dm8n\" (UniqueName: \"kubernetes.io/projected/93fa4e0c-38c0-4793-80b2-d39258b416e3-kube-api-access-5dm8n\") pod \"collect-profiles-29418765-57x25\" (UID: \"93fa4e0c-38c0-4793-80b2-d39258b416e3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418765-57x25" Dec 07 16:45:00 crc kubenswrapper[4716]: I1207 16:45:00.263768 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/93fa4e0c-38c0-4793-80b2-d39258b416e3-config-volume\") pod \"collect-profiles-29418765-57x25\" (UID: \"93fa4e0c-38c0-4793-80b2-d39258b416e3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418765-57x25" Dec 07 16:45:00 crc kubenswrapper[4716]: I1207 16:45:00.275864 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/93fa4e0c-38c0-4793-80b2-d39258b416e3-secret-volume\") pod \"collect-profiles-29418765-57x25\" (UID: \"93fa4e0c-38c0-4793-80b2-d39258b416e3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418765-57x25" Dec 07 16:45:00 crc kubenswrapper[4716]: I1207 16:45:00.282110 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5dm8n\" (UniqueName: \"kubernetes.io/projected/93fa4e0c-38c0-4793-80b2-d39258b416e3-kube-api-access-5dm8n\") pod \"collect-profiles-29418765-57x25\" (UID: \"93fa4e0c-38c0-4793-80b2-d39258b416e3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418765-57x25" Dec 07 16:45:00 crc kubenswrapper[4716]: I1207 16:45:00.474900 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29418765-57x25" Dec 07 16:45:00 crc kubenswrapper[4716]: I1207 16:45:00.918226 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29418765-57x25"] Dec 07 16:45:01 crc kubenswrapper[4716]: I1207 16:45:01.607227 4716 generic.go:334] "Generic (PLEG): container finished" podID="93fa4e0c-38c0-4793-80b2-d39258b416e3" containerID="8ee780bf15cb5710a3631ecf226d61451334be28fa484899d45fbafbdce81125" exitCode=0 Dec 07 16:45:01 crc kubenswrapper[4716]: I1207 16:45:01.607328 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29418765-57x25" event={"ID":"93fa4e0c-38c0-4793-80b2-d39258b416e3","Type":"ContainerDied","Data":"8ee780bf15cb5710a3631ecf226d61451334be28fa484899d45fbafbdce81125"} Dec 07 16:45:01 crc kubenswrapper[4716]: I1207 16:45:01.607672 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29418765-57x25" event={"ID":"93fa4e0c-38c0-4793-80b2-d39258b416e3","Type":"ContainerStarted","Data":"babe0ae634f8f4deb3e14c7072cbb07783ac62e8f5e1a09e33258f4e317919da"} Dec 07 16:45:02 crc kubenswrapper[4716]: I1207 16:45:02.950112 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29418765-57x25" Dec 07 16:45:03 crc kubenswrapper[4716]: I1207 16:45:03.114129 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5dm8n\" (UniqueName: \"kubernetes.io/projected/93fa4e0c-38c0-4793-80b2-d39258b416e3-kube-api-access-5dm8n\") pod \"93fa4e0c-38c0-4793-80b2-d39258b416e3\" (UID: \"93fa4e0c-38c0-4793-80b2-d39258b416e3\") " Dec 07 16:45:03 crc kubenswrapper[4716]: I1207 16:45:03.114685 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/93fa4e0c-38c0-4793-80b2-d39258b416e3-config-volume\") pod \"93fa4e0c-38c0-4793-80b2-d39258b416e3\" (UID: \"93fa4e0c-38c0-4793-80b2-d39258b416e3\") " Dec 07 16:45:03 crc kubenswrapper[4716]: I1207 16:45:03.114784 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/93fa4e0c-38c0-4793-80b2-d39258b416e3-secret-volume\") pod \"93fa4e0c-38c0-4793-80b2-d39258b416e3\" (UID: \"93fa4e0c-38c0-4793-80b2-d39258b416e3\") " Dec 07 16:45:03 crc kubenswrapper[4716]: I1207 16:45:03.115353 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/93fa4e0c-38c0-4793-80b2-d39258b416e3-config-volume" (OuterVolumeSpecName: "config-volume") pod "93fa4e0c-38c0-4793-80b2-d39258b416e3" (UID: "93fa4e0c-38c0-4793-80b2-d39258b416e3"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:45:03 crc kubenswrapper[4716]: I1207 16:45:03.119733 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93fa4e0c-38c0-4793-80b2-d39258b416e3-kube-api-access-5dm8n" (OuterVolumeSpecName: "kube-api-access-5dm8n") pod "93fa4e0c-38c0-4793-80b2-d39258b416e3" (UID: "93fa4e0c-38c0-4793-80b2-d39258b416e3"). InnerVolumeSpecName "kube-api-access-5dm8n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:45:03 crc kubenswrapper[4716]: I1207 16:45:03.123259 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93fa4e0c-38c0-4793-80b2-d39258b416e3-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "93fa4e0c-38c0-4793-80b2-d39258b416e3" (UID: "93fa4e0c-38c0-4793-80b2-d39258b416e3"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:45:03 crc kubenswrapper[4716]: I1207 16:45:03.217495 4716 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/93fa4e0c-38c0-4793-80b2-d39258b416e3-config-volume\") on node \"crc\" DevicePath \"\"" Dec 07 16:45:03 crc kubenswrapper[4716]: I1207 16:45:03.217538 4716 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/93fa4e0c-38c0-4793-80b2-d39258b416e3-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 07 16:45:03 crc kubenswrapper[4716]: I1207 16:45:03.217551 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5dm8n\" (UniqueName: \"kubernetes.io/projected/93fa4e0c-38c0-4793-80b2-d39258b416e3-kube-api-access-5dm8n\") on node \"crc\" DevicePath \"\"" Dec 07 16:45:03 crc kubenswrapper[4716]: I1207 16:45:03.625806 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29418765-57x25" event={"ID":"93fa4e0c-38c0-4793-80b2-d39258b416e3","Type":"ContainerDied","Data":"babe0ae634f8f4deb3e14c7072cbb07783ac62e8f5e1a09e33258f4e317919da"} Dec 07 16:45:03 crc kubenswrapper[4716]: I1207 16:45:03.625852 4716 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="babe0ae634f8f4deb3e14c7072cbb07783ac62e8f5e1a09e33258f4e317919da" Dec 07 16:45:03 crc kubenswrapper[4716]: I1207 16:45:03.625870 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29418765-57x25" Dec 07 16:45:04 crc kubenswrapper[4716]: I1207 16:45:04.031718 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29418720-x5rqz"] Dec 07 16:45:04 crc kubenswrapper[4716]: I1207 16:45:04.040849 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29418720-x5rqz"] Dec 07 16:45:05 crc kubenswrapper[4716]: I1207 16:45:05.657629 4716 scope.go:117] "RemoveContainer" containerID="3eb63b7962def203a31ba32ba16ac93470e254c2a4edb4a4dd2033b290e2f217" Dec 07 16:45:05 crc kubenswrapper[4716]: E1207 16:45:05.658187 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 16:45:05 crc kubenswrapper[4716]: I1207 16:45:05.668899 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98efd33b-de1f-4414-b9cc-55fb0d423784" path="/var/lib/kubelet/pods/98efd33b-de1f-4414-b9cc-55fb0d423784/volumes" Dec 07 16:45:20 crc kubenswrapper[4716]: I1207 16:45:20.657793 4716 scope.go:117] "RemoveContainer" containerID="3eb63b7962def203a31ba32ba16ac93470e254c2a4edb4a4dd2033b290e2f217" Dec 07 16:45:20 crc kubenswrapper[4716]: E1207 16:45:20.658670 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 16:45:34 crc kubenswrapper[4716]: I1207 16:45:34.658405 4716 scope.go:117] "RemoveContainer" containerID="3eb63b7962def203a31ba32ba16ac93470e254c2a4edb4a4dd2033b290e2f217" Dec 07 16:45:34 crc kubenswrapper[4716]: E1207 16:45:34.659258 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 16:45:34 crc kubenswrapper[4716]: I1207 16:45:34.792901 4716 scope.go:117] "RemoveContainer" containerID="efc3cdf3d0d1568d81c77d3121f4fc0f37b2c7bfba87382c2b7631532f2d134b" Dec 07 16:45:47 crc kubenswrapper[4716]: I1207 16:45:47.664486 4716 scope.go:117] "RemoveContainer" containerID="3eb63b7962def203a31ba32ba16ac93470e254c2a4edb4a4dd2033b290e2f217" Dec 07 16:45:47 crc kubenswrapper[4716]: E1207 16:45:47.665474 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 16:46:01 crc kubenswrapper[4716]: I1207 16:46:01.657858 4716 scope.go:117] "RemoveContainer" containerID="3eb63b7962def203a31ba32ba16ac93470e254c2a4edb4a4dd2033b290e2f217" Dec 07 16:46:01 crc kubenswrapper[4716]: E1207 16:46:01.658994 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 16:46:13 crc kubenswrapper[4716]: I1207 16:46:13.658342 4716 scope.go:117] "RemoveContainer" containerID="3eb63b7962def203a31ba32ba16ac93470e254c2a4edb4a4dd2033b290e2f217" Dec 07 16:46:13 crc kubenswrapper[4716]: E1207 16:46:13.659610 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 16:46:25 crc kubenswrapper[4716]: I1207 16:46:25.658021 4716 scope.go:117] "RemoveContainer" containerID="3eb63b7962def203a31ba32ba16ac93470e254c2a4edb4a4dd2033b290e2f217" Dec 07 16:46:25 crc kubenswrapper[4716]: E1207 16:46:25.658937 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 16:46:39 crc kubenswrapper[4716]: I1207 16:46:39.659488 4716 scope.go:117] "RemoveContainer" containerID="3eb63b7962def203a31ba32ba16ac93470e254c2a4edb4a4dd2033b290e2f217" Dec 07 16:46:39 crc kubenswrapper[4716]: E1207 16:46:39.660552 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 16:46:50 crc kubenswrapper[4716]: I1207 16:46:50.657970 4716 scope.go:117] "RemoveContainer" containerID="3eb63b7962def203a31ba32ba16ac93470e254c2a4edb4a4dd2033b290e2f217" Dec 07 16:46:50 crc kubenswrapper[4716]: E1207 16:46:50.659305 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 16:47:02 crc kubenswrapper[4716]: I1207 16:47:02.657951 4716 scope.go:117] "RemoveContainer" containerID="3eb63b7962def203a31ba32ba16ac93470e254c2a4edb4a4dd2033b290e2f217" Dec 07 16:47:02 crc kubenswrapper[4716]: E1207 16:47:02.658857 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 16:47:03 crc kubenswrapper[4716]: I1207 16:47:03.716650 4716 generic.go:334] "Generic (PLEG): container finished" podID="c9735df5-27ef-41b1-84c1-4cab55f23b3a" containerID="f6eecc554d89c956aad8dbb5c308bc5e9c56a401f0a33d8a3a15e756910184d6" exitCode=0 Dec 07 16:47:03 crc kubenswrapper[4716]: I1207 16:47:03.716694 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-69qz6" event={"ID":"c9735df5-27ef-41b1-84c1-4cab55f23b3a","Type":"ContainerDied","Data":"f6eecc554d89c956aad8dbb5c308bc5e9c56a401f0a33d8a3a15e756910184d6"} Dec 07 16:47:05 crc kubenswrapper[4716]: I1207 16:47:05.125480 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-69qz6" Dec 07 16:47:05 crc kubenswrapper[4716]: I1207 16:47:05.232241 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/c9735df5-27ef-41b1-84c1-4cab55f23b3a-ceilometer-compute-config-data-0\") pod \"c9735df5-27ef-41b1-84c1-4cab55f23b3a\" (UID: \"c9735df5-27ef-41b1-84c1-4cab55f23b3a\") " Dec 07 16:47:05 crc kubenswrapper[4716]: I1207 16:47:05.232345 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fct8b\" (UniqueName: \"kubernetes.io/projected/c9735df5-27ef-41b1-84c1-4cab55f23b3a-kube-api-access-fct8b\") pod \"c9735df5-27ef-41b1-84c1-4cab55f23b3a\" (UID: \"c9735df5-27ef-41b1-84c1-4cab55f23b3a\") " Dec 07 16:47:05 crc kubenswrapper[4716]: I1207 16:47:05.232373 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/c9735df5-27ef-41b1-84c1-4cab55f23b3a-ceilometer-compute-config-data-2\") pod \"c9735df5-27ef-41b1-84c1-4cab55f23b3a\" (UID: \"c9735df5-27ef-41b1-84c1-4cab55f23b3a\") " Dec 07 16:47:05 crc kubenswrapper[4716]: I1207 16:47:05.232405 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c9735df5-27ef-41b1-84c1-4cab55f23b3a-ssh-key\") pod \"c9735df5-27ef-41b1-84c1-4cab55f23b3a\" (UID: \"c9735df5-27ef-41b1-84c1-4cab55f23b3a\") " Dec 07 16:47:05 crc kubenswrapper[4716]: I1207 16:47:05.232442 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9735df5-27ef-41b1-84c1-4cab55f23b3a-telemetry-combined-ca-bundle\") pod \"c9735df5-27ef-41b1-84c1-4cab55f23b3a\" (UID: \"c9735df5-27ef-41b1-84c1-4cab55f23b3a\") " Dec 07 16:47:05 crc kubenswrapper[4716]: I1207 16:47:05.232508 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c9735df5-27ef-41b1-84c1-4cab55f23b3a-inventory\") pod \"c9735df5-27ef-41b1-84c1-4cab55f23b3a\" (UID: \"c9735df5-27ef-41b1-84c1-4cab55f23b3a\") " Dec 07 16:47:05 crc kubenswrapper[4716]: I1207 16:47:05.232556 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/c9735df5-27ef-41b1-84c1-4cab55f23b3a-ceilometer-compute-config-data-1\") pod \"c9735df5-27ef-41b1-84c1-4cab55f23b3a\" (UID: \"c9735df5-27ef-41b1-84c1-4cab55f23b3a\") " Dec 07 16:47:05 crc kubenswrapper[4716]: I1207 16:47:05.238350 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9735df5-27ef-41b1-84c1-4cab55f23b3a-kube-api-access-fct8b" (OuterVolumeSpecName: "kube-api-access-fct8b") pod "c9735df5-27ef-41b1-84c1-4cab55f23b3a" (UID: "c9735df5-27ef-41b1-84c1-4cab55f23b3a"). InnerVolumeSpecName "kube-api-access-fct8b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:47:05 crc kubenswrapper[4716]: I1207 16:47:05.243926 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9735df5-27ef-41b1-84c1-4cab55f23b3a-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "c9735df5-27ef-41b1-84c1-4cab55f23b3a" (UID: "c9735df5-27ef-41b1-84c1-4cab55f23b3a"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:47:05 crc kubenswrapper[4716]: I1207 16:47:05.261297 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9735df5-27ef-41b1-84c1-4cab55f23b3a-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "c9735df5-27ef-41b1-84c1-4cab55f23b3a" (UID: "c9735df5-27ef-41b1-84c1-4cab55f23b3a"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:47:05 crc kubenswrapper[4716]: I1207 16:47:05.261605 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9735df5-27ef-41b1-84c1-4cab55f23b3a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c9735df5-27ef-41b1-84c1-4cab55f23b3a" (UID: "c9735df5-27ef-41b1-84c1-4cab55f23b3a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:47:05 crc kubenswrapper[4716]: I1207 16:47:05.270372 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9735df5-27ef-41b1-84c1-4cab55f23b3a-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "c9735df5-27ef-41b1-84c1-4cab55f23b3a" (UID: "c9735df5-27ef-41b1-84c1-4cab55f23b3a"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:47:05 crc kubenswrapper[4716]: I1207 16:47:05.270944 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9735df5-27ef-41b1-84c1-4cab55f23b3a-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "c9735df5-27ef-41b1-84c1-4cab55f23b3a" (UID: "c9735df5-27ef-41b1-84c1-4cab55f23b3a"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:47:05 crc kubenswrapper[4716]: I1207 16:47:05.281015 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9735df5-27ef-41b1-84c1-4cab55f23b3a-inventory" (OuterVolumeSpecName: "inventory") pod "c9735df5-27ef-41b1-84c1-4cab55f23b3a" (UID: "c9735df5-27ef-41b1-84c1-4cab55f23b3a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:47:05 crc kubenswrapper[4716]: I1207 16:47:05.334159 4716 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c9735df5-27ef-41b1-84c1-4cab55f23b3a-inventory\") on node \"crc\" DevicePath \"\"" Dec 07 16:47:05 crc kubenswrapper[4716]: I1207 16:47:05.334189 4716 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/c9735df5-27ef-41b1-84c1-4cab55f23b3a-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 07 16:47:05 crc kubenswrapper[4716]: I1207 16:47:05.334203 4716 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/c9735df5-27ef-41b1-84c1-4cab55f23b3a-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 07 16:47:05 crc kubenswrapper[4716]: I1207 16:47:05.334212 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fct8b\" (UniqueName: \"kubernetes.io/projected/c9735df5-27ef-41b1-84c1-4cab55f23b3a-kube-api-access-fct8b\") on node \"crc\" DevicePath \"\"" Dec 07 16:47:05 crc kubenswrapper[4716]: I1207 16:47:05.334223 4716 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/c9735df5-27ef-41b1-84c1-4cab55f23b3a-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Dec 07 16:47:05 crc kubenswrapper[4716]: I1207 16:47:05.334232 4716 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c9735df5-27ef-41b1-84c1-4cab55f23b3a-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 07 16:47:05 crc kubenswrapper[4716]: I1207 16:47:05.334240 4716 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9735df5-27ef-41b1-84c1-4cab55f23b3a-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 16:47:05 crc kubenswrapper[4716]: I1207 16:47:05.734814 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-69qz6" event={"ID":"c9735df5-27ef-41b1-84c1-4cab55f23b3a","Type":"ContainerDied","Data":"ebe10360ff3c44deaf764b47072771041b95118a0aa0d1e55124789d03b95271"} Dec 07 16:47:05 crc kubenswrapper[4716]: I1207 16:47:05.734853 4716 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ebe10360ff3c44deaf764b47072771041b95118a0aa0d1e55124789d03b95271" Dec 07 16:47:05 crc kubenswrapper[4716]: I1207 16:47:05.734864 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-69qz6" Dec 07 16:47:13 crc kubenswrapper[4716]: I1207 16:47:13.658306 4716 scope.go:117] "RemoveContainer" containerID="3eb63b7962def203a31ba32ba16ac93470e254c2a4edb4a4dd2033b290e2f217" Dec 07 16:47:13 crc kubenswrapper[4716]: E1207 16:47:13.659197 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 16:47:26 crc kubenswrapper[4716]: I1207 16:47:26.657710 4716 scope.go:117] "RemoveContainer" containerID="3eb63b7962def203a31ba32ba16ac93470e254c2a4edb4a4dd2033b290e2f217" Dec 07 16:47:26 crc kubenswrapper[4716]: I1207 16:47:26.925306 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" event={"ID":"c15b59eb-565d-4556-a4ce-75afdf159dc8","Type":"ContainerStarted","Data":"be6a29d905d404203a7ce5b0088583826cc3981f161994b9094c7227407ce5aa"} Dec 07 16:47:55 crc kubenswrapper[4716]: I1207 16:47:55.505457 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Dec 07 16:47:55 crc kubenswrapper[4716]: E1207 16:47:55.506709 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9735df5-27ef-41b1-84c1-4cab55f23b3a" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 07 16:47:55 crc kubenswrapper[4716]: I1207 16:47:55.506731 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9735df5-27ef-41b1-84c1-4cab55f23b3a" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 07 16:47:55 crc kubenswrapper[4716]: E1207 16:47:55.506795 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93fa4e0c-38c0-4793-80b2-d39258b416e3" containerName="collect-profiles" Dec 07 16:47:55 crc kubenswrapper[4716]: I1207 16:47:55.506807 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="93fa4e0c-38c0-4793-80b2-d39258b416e3" containerName="collect-profiles" Dec 07 16:47:55 crc kubenswrapper[4716]: I1207 16:47:55.507022 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9735df5-27ef-41b1-84c1-4cab55f23b3a" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 07 16:47:55 crc kubenswrapper[4716]: I1207 16:47:55.507053 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="93fa4e0c-38c0-4793-80b2-d39258b416e3" containerName="collect-profiles" Dec 07 16:47:55 crc kubenswrapper[4716]: I1207 16:47:55.507917 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 07 16:47:55 crc kubenswrapper[4716]: I1207 16:47:55.510406 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Dec 07 16:47:55 crc kubenswrapper[4716]: I1207 16:47:55.510422 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-mq8wt" Dec 07 16:47:55 crc kubenswrapper[4716]: I1207 16:47:55.510959 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 07 16:47:55 crc kubenswrapper[4716]: I1207 16:47:55.513071 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Dec 07 16:47:55 crc kubenswrapper[4716]: I1207 16:47:55.538514 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 07 16:47:55 crc kubenswrapper[4716]: I1207 16:47:55.663729 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/7c4bb2e7-29d0-4548-863b-ea89d11d68a7-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"7c4bb2e7-29d0-4548-863b-ea89d11d68a7\") " pod="openstack/tempest-tests-tempest" Dec 07 16:47:55 crc kubenswrapper[4716]: I1207 16:47:55.663802 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/7c4bb2e7-29d0-4548-863b-ea89d11d68a7-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"7c4bb2e7-29d0-4548-863b-ea89d11d68a7\") " pod="openstack/tempest-tests-tempest" Dec 07 16:47:55 crc kubenswrapper[4716]: I1207 16:47:55.663830 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/7c4bb2e7-29d0-4548-863b-ea89d11d68a7-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"7c4bb2e7-29d0-4548-863b-ea89d11d68a7\") " pod="openstack/tempest-tests-tempest" Dec 07 16:47:55 crc kubenswrapper[4716]: I1207 16:47:55.664313 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-476jr\" (UniqueName: \"kubernetes.io/projected/7c4bb2e7-29d0-4548-863b-ea89d11d68a7-kube-api-access-476jr\") pod \"tempest-tests-tempest\" (UID: \"7c4bb2e7-29d0-4548-863b-ea89d11d68a7\") " pod="openstack/tempest-tests-tempest" Dec 07 16:47:55 crc kubenswrapper[4716]: I1207 16:47:55.664482 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7c4bb2e7-29d0-4548-863b-ea89d11d68a7-config-data\") pod \"tempest-tests-tempest\" (UID: \"7c4bb2e7-29d0-4548-863b-ea89d11d68a7\") " pod="openstack/tempest-tests-tempest" Dec 07 16:47:55 crc kubenswrapper[4716]: I1207 16:47:55.664716 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"tempest-tests-tempest\" (UID: \"7c4bb2e7-29d0-4548-863b-ea89d11d68a7\") " pod="openstack/tempest-tests-tempest" Dec 07 16:47:55 crc kubenswrapper[4716]: I1207 16:47:55.664755 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7c4bb2e7-29d0-4548-863b-ea89d11d68a7-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"7c4bb2e7-29d0-4548-863b-ea89d11d68a7\") " pod="openstack/tempest-tests-tempest" Dec 07 16:47:55 crc kubenswrapper[4716]: I1207 16:47:55.665255 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/7c4bb2e7-29d0-4548-863b-ea89d11d68a7-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"7c4bb2e7-29d0-4548-863b-ea89d11d68a7\") " pod="openstack/tempest-tests-tempest" Dec 07 16:47:55 crc kubenswrapper[4716]: I1207 16:47:55.665431 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/7c4bb2e7-29d0-4548-863b-ea89d11d68a7-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"7c4bb2e7-29d0-4548-863b-ea89d11d68a7\") " pod="openstack/tempest-tests-tempest" Dec 07 16:47:55 crc kubenswrapper[4716]: I1207 16:47:55.767607 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"tempest-tests-tempest\" (UID: \"7c4bb2e7-29d0-4548-863b-ea89d11d68a7\") " pod="openstack/tempest-tests-tempest" Dec 07 16:47:55 crc kubenswrapper[4716]: I1207 16:47:55.767666 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7c4bb2e7-29d0-4548-863b-ea89d11d68a7-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"7c4bb2e7-29d0-4548-863b-ea89d11d68a7\") " pod="openstack/tempest-tests-tempest" Dec 07 16:47:55 crc kubenswrapper[4716]: I1207 16:47:55.767733 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/7c4bb2e7-29d0-4548-863b-ea89d11d68a7-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"7c4bb2e7-29d0-4548-863b-ea89d11d68a7\") " pod="openstack/tempest-tests-tempest" Dec 07 16:47:55 crc kubenswrapper[4716]: I1207 16:47:55.767775 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/7c4bb2e7-29d0-4548-863b-ea89d11d68a7-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"7c4bb2e7-29d0-4548-863b-ea89d11d68a7\") " pod="openstack/tempest-tests-tempest" Dec 07 16:47:55 crc kubenswrapper[4716]: I1207 16:47:55.767856 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/7c4bb2e7-29d0-4548-863b-ea89d11d68a7-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"7c4bb2e7-29d0-4548-863b-ea89d11d68a7\") " pod="openstack/tempest-tests-tempest" Dec 07 16:47:55 crc kubenswrapper[4716]: I1207 16:47:55.767894 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/7c4bb2e7-29d0-4548-863b-ea89d11d68a7-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"7c4bb2e7-29d0-4548-863b-ea89d11d68a7\") " pod="openstack/tempest-tests-tempest" Dec 07 16:47:55 crc kubenswrapper[4716]: I1207 16:47:55.767920 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/7c4bb2e7-29d0-4548-863b-ea89d11d68a7-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"7c4bb2e7-29d0-4548-863b-ea89d11d68a7\") " pod="openstack/tempest-tests-tempest" Dec 07 16:47:55 crc kubenswrapper[4716]: I1207 16:47:55.767974 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-476jr\" (UniqueName: \"kubernetes.io/projected/7c4bb2e7-29d0-4548-863b-ea89d11d68a7-kube-api-access-476jr\") pod \"tempest-tests-tempest\" (UID: \"7c4bb2e7-29d0-4548-863b-ea89d11d68a7\") " pod="openstack/tempest-tests-tempest" Dec 07 16:47:55 crc kubenswrapper[4716]: I1207 16:47:55.768015 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7c4bb2e7-29d0-4548-863b-ea89d11d68a7-config-data\") pod \"tempest-tests-tempest\" (UID: \"7c4bb2e7-29d0-4548-863b-ea89d11d68a7\") " pod="openstack/tempest-tests-tempest" Dec 07 16:47:55 crc kubenswrapper[4716]: I1207 16:47:55.768158 4716 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"tempest-tests-tempest\" (UID: \"7c4bb2e7-29d0-4548-863b-ea89d11d68a7\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/tempest-tests-tempest" Dec 07 16:47:55 crc kubenswrapper[4716]: I1207 16:47:55.768637 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/7c4bb2e7-29d0-4548-863b-ea89d11d68a7-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"7c4bb2e7-29d0-4548-863b-ea89d11d68a7\") " pod="openstack/tempest-tests-tempest" Dec 07 16:47:55 crc kubenswrapper[4716]: I1207 16:47:55.768708 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/7c4bb2e7-29d0-4548-863b-ea89d11d68a7-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"7c4bb2e7-29d0-4548-863b-ea89d11d68a7\") " pod="openstack/tempest-tests-tempest" Dec 07 16:47:55 crc kubenswrapper[4716]: I1207 16:47:55.769558 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7c4bb2e7-29d0-4548-863b-ea89d11d68a7-config-data\") pod \"tempest-tests-tempest\" (UID: \"7c4bb2e7-29d0-4548-863b-ea89d11d68a7\") " pod="openstack/tempest-tests-tempest" Dec 07 16:47:55 crc kubenswrapper[4716]: I1207 16:47:55.769899 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/7c4bb2e7-29d0-4548-863b-ea89d11d68a7-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"7c4bb2e7-29d0-4548-863b-ea89d11d68a7\") " pod="openstack/tempest-tests-tempest" Dec 07 16:47:55 crc kubenswrapper[4716]: I1207 16:47:55.775844 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7c4bb2e7-29d0-4548-863b-ea89d11d68a7-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"7c4bb2e7-29d0-4548-863b-ea89d11d68a7\") " pod="openstack/tempest-tests-tempest" Dec 07 16:47:55 crc kubenswrapper[4716]: I1207 16:47:55.780369 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/7c4bb2e7-29d0-4548-863b-ea89d11d68a7-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"7c4bb2e7-29d0-4548-863b-ea89d11d68a7\") " pod="openstack/tempest-tests-tempest" Dec 07 16:47:55 crc kubenswrapper[4716]: I1207 16:47:55.780592 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/7c4bb2e7-29d0-4548-863b-ea89d11d68a7-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"7c4bb2e7-29d0-4548-863b-ea89d11d68a7\") " pod="openstack/tempest-tests-tempest" Dec 07 16:47:55 crc kubenswrapper[4716]: I1207 16:47:55.799505 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-476jr\" (UniqueName: \"kubernetes.io/projected/7c4bb2e7-29d0-4548-863b-ea89d11d68a7-kube-api-access-476jr\") pod \"tempest-tests-tempest\" (UID: \"7c4bb2e7-29d0-4548-863b-ea89d11d68a7\") " pod="openstack/tempest-tests-tempest" Dec 07 16:47:55 crc kubenswrapper[4716]: I1207 16:47:55.799756 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"tempest-tests-tempest\" (UID: \"7c4bb2e7-29d0-4548-863b-ea89d11d68a7\") " pod="openstack/tempest-tests-tempest" Dec 07 16:47:55 crc kubenswrapper[4716]: I1207 16:47:55.850746 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 07 16:47:56 crc kubenswrapper[4716]: I1207 16:47:56.371480 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 07 16:47:56 crc kubenswrapper[4716]: I1207 16:47:56.381563 4716 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 07 16:47:57 crc kubenswrapper[4716]: I1207 16:47:57.207297 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"7c4bb2e7-29d0-4548-863b-ea89d11d68a7","Type":"ContainerStarted","Data":"9b4ab7ee63628f7976099cf1e42c2babc5b245cd708504e41763ba408bb35b43"} Dec 07 16:47:58 crc kubenswrapper[4716]: I1207 16:47:58.905687 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-qhf7r"] Dec 07 16:47:58 crc kubenswrapper[4716]: I1207 16:47:58.908589 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qhf7r" Dec 07 16:47:58 crc kubenswrapper[4716]: I1207 16:47:58.917017 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qhf7r"] Dec 07 16:47:59 crc kubenswrapper[4716]: I1207 16:47:59.031225 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de3cdf40-b257-4964-85a2-577c2e596eb6-catalog-content\") pod \"redhat-operators-qhf7r\" (UID: \"de3cdf40-b257-4964-85a2-577c2e596eb6\") " pod="openshift-marketplace/redhat-operators-qhf7r" Dec 07 16:47:59 crc kubenswrapper[4716]: I1207 16:47:59.031298 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de3cdf40-b257-4964-85a2-577c2e596eb6-utilities\") pod \"redhat-operators-qhf7r\" (UID: \"de3cdf40-b257-4964-85a2-577c2e596eb6\") " pod="openshift-marketplace/redhat-operators-qhf7r" Dec 07 16:47:59 crc kubenswrapper[4716]: I1207 16:47:59.031383 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-44qsk\" (UniqueName: \"kubernetes.io/projected/de3cdf40-b257-4964-85a2-577c2e596eb6-kube-api-access-44qsk\") pod \"redhat-operators-qhf7r\" (UID: \"de3cdf40-b257-4964-85a2-577c2e596eb6\") " pod="openshift-marketplace/redhat-operators-qhf7r" Dec 07 16:47:59 crc kubenswrapper[4716]: I1207 16:47:59.133596 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de3cdf40-b257-4964-85a2-577c2e596eb6-catalog-content\") pod \"redhat-operators-qhf7r\" (UID: \"de3cdf40-b257-4964-85a2-577c2e596eb6\") " pod="openshift-marketplace/redhat-operators-qhf7r" Dec 07 16:47:59 crc kubenswrapper[4716]: I1207 16:47:59.133658 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de3cdf40-b257-4964-85a2-577c2e596eb6-utilities\") pod \"redhat-operators-qhf7r\" (UID: \"de3cdf40-b257-4964-85a2-577c2e596eb6\") " pod="openshift-marketplace/redhat-operators-qhf7r" Dec 07 16:47:59 crc kubenswrapper[4716]: I1207 16:47:59.133733 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-44qsk\" (UniqueName: \"kubernetes.io/projected/de3cdf40-b257-4964-85a2-577c2e596eb6-kube-api-access-44qsk\") pod \"redhat-operators-qhf7r\" (UID: \"de3cdf40-b257-4964-85a2-577c2e596eb6\") " pod="openshift-marketplace/redhat-operators-qhf7r" Dec 07 16:47:59 crc kubenswrapper[4716]: I1207 16:47:59.134399 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de3cdf40-b257-4964-85a2-577c2e596eb6-catalog-content\") pod \"redhat-operators-qhf7r\" (UID: \"de3cdf40-b257-4964-85a2-577c2e596eb6\") " pod="openshift-marketplace/redhat-operators-qhf7r" Dec 07 16:47:59 crc kubenswrapper[4716]: I1207 16:47:59.134474 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de3cdf40-b257-4964-85a2-577c2e596eb6-utilities\") pod \"redhat-operators-qhf7r\" (UID: \"de3cdf40-b257-4964-85a2-577c2e596eb6\") " pod="openshift-marketplace/redhat-operators-qhf7r" Dec 07 16:47:59 crc kubenswrapper[4716]: I1207 16:47:59.157831 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-44qsk\" (UniqueName: \"kubernetes.io/projected/de3cdf40-b257-4964-85a2-577c2e596eb6-kube-api-access-44qsk\") pod \"redhat-operators-qhf7r\" (UID: \"de3cdf40-b257-4964-85a2-577c2e596eb6\") " pod="openshift-marketplace/redhat-operators-qhf7r" Dec 07 16:47:59 crc kubenswrapper[4716]: I1207 16:47:59.241483 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qhf7r" Dec 07 16:48:02 crc kubenswrapper[4716]: I1207 16:48:02.268812 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qhf7r"] Dec 07 16:48:02 crc kubenswrapper[4716]: W1207 16:48:02.279122 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podde3cdf40_b257_4964_85a2_577c2e596eb6.slice/crio-1e5fe7b9917bd1d102fc764ae05e78955f54259151224500c1a1eecb50ed95aa WatchSource:0}: Error finding container 1e5fe7b9917bd1d102fc764ae05e78955f54259151224500c1a1eecb50ed95aa: Status 404 returned error can't find the container with id 1e5fe7b9917bd1d102fc764ae05e78955f54259151224500c1a1eecb50ed95aa Dec 07 16:48:02 crc kubenswrapper[4716]: I1207 16:48:02.985448 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-vvmlg"] Dec 07 16:48:02 crc kubenswrapper[4716]: I1207 16:48:02.987824 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vvmlg" Dec 07 16:48:02 crc kubenswrapper[4716]: I1207 16:48:02.999184 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vvmlg"] Dec 07 16:48:03 crc kubenswrapper[4716]: I1207 16:48:03.109828 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0cf7c849-73ad-4681-9be9-6dc84ddc6c13-catalog-content\") pod \"redhat-marketplace-vvmlg\" (UID: \"0cf7c849-73ad-4681-9be9-6dc84ddc6c13\") " pod="openshift-marketplace/redhat-marketplace-vvmlg" Dec 07 16:48:03 crc kubenswrapper[4716]: I1207 16:48:03.110698 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0cf7c849-73ad-4681-9be9-6dc84ddc6c13-utilities\") pod \"redhat-marketplace-vvmlg\" (UID: \"0cf7c849-73ad-4681-9be9-6dc84ddc6c13\") " pod="openshift-marketplace/redhat-marketplace-vvmlg" Dec 07 16:48:03 crc kubenswrapper[4716]: I1207 16:48:03.110818 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8n8qc\" (UniqueName: \"kubernetes.io/projected/0cf7c849-73ad-4681-9be9-6dc84ddc6c13-kube-api-access-8n8qc\") pod \"redhat-marketplace-vvmlg\" (UID: \"0cf7c849-73ad-4681-9be9-6dc84ddc6c13\") " pod="openshift-marketplace/redhat-marketplace-vvmlg" Dec 07 16:48:03 crc kubenswrapper[4716]: I1207 16:48:03.212116 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0cf7c849-73ad-4681-9be9-6dc84ddc6c13-utilities\") pod \"redhat-marketplace-vvmlg\" (UID: \"0cf7c849-73ad-4681-9be9-6dc84ddc6c13\") " pod="openshift-marketplace/redhat-marketplace-vvmlg" Dec 07 16:48:03 crc kubenswrapper[4716]: I1207 16:48:03.212179 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8n8qc\" (UniqueName: \"kubernetes.io/projected/0cf7c849-73ad-4681-9be9-6dc84ddc6c13-kube-api-access-8n8qc\") pod \"redhat-marketplace-vvmlg\" (UID: \"0cf7c849-73ad-4681-9be9-6dc84ddc6c13\") " pod="openshift-marketplace/redhat-marketplace-vvmlg" Dec 07 16:48:03 crc kubenswrapper[4716]: I1207 16:48:03.212282 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0cf7c849-73ad-4681-9be9-6dc84ddc6c13-catalog-content\") pod \"redhat-marketplace-vvmlg\" (UID: \"0cf7c849-73ad-4681-9be9-6dc84ddc6c13\") " pod="openshift-marketplace/redhat-marketplace-vvmlg" Dec 07 16:48:03 crc kubenswrapper[4716]: I1207 16:48:03.212814 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0cf7c849-73ad-4681-9be9-6dc84ddc6c13-catalog-content\") pod \"redhat-marketplace-vvmlg\" (UID: \"0cf7c849-73ad-4681-9be9-6dc84ddc6c13\") " pod="openshift-marketplace/redhat-marketplace-vvmlg" Dec 07 16:48:03 crc kubenswrapper[4716]: I1207 16:48:03.213099 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0cf7c849-73ad-4681-9be9-6dc84ddc6c13-utilities\") pod \"redhat-marketplace-vvmlg\" (UID: \"0cf7c849-73ad-4681-9be9-6dc84ddc6c13\") " pod="openshift-marketplace/redhat-marketplace-vvmlg" Dec 07 16:48:03 crc kubenswrapper[4716]: I1207 16:48:03.232225 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8n8qc\" (UniqueName: \"kubernetes.io/projected/0cf7c849-73ad-4681-9be9-6dc84ddc6c13-kube-api-access-8n8qc\") pod \"redhat-marketplace-vvmlg\" (UID: \"0cf7c849-73ad-4681-9be9-6dc84ddc6c13\") " pod="openshift-marketplace/redhat-marketplace-vvmlg" Dec 07 16:48:03 crc kubenswrapper[4716]: I1207 16:48:03.289813 4716 generic.go:334] "Generic (PLEG): container finished" podID="de3cdf40-b257-4964-85a2-577c2e596eb6" containerID="3ffef06fe3c15c87cdad2c881969d0af9161e55d7acadc67ea6a788622a6e94f" exitCode=0 Dec 07 16:48:03 crc kubenswrapper[4716]: I1207 16:48:03.289860 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qhf7r" event={"ID":"de3cdf40-b257-4964-85a2-577c2e596eb6","Type":"ContainerDied","Data":"3ffef06fe3c15c87cdad2c881969d0af9161e55d7acadc67ea6a788622a6e94f"} Dec 07 16:48:03 crc kubenswrapper[4716]: I1207 16:48:03.289885 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qhf7r" event={"ID":"de3cdf40-b257-4964-85a2-577c2e596eb6","Type":"ContainerStarted","Data":"1e5fe7b9917bd1d102fc764ae05e78955f54259151224500c1a1eecb50ed95aa"} Dec 07 16:48:03 crc kubenswrapper[4716]: I1207 16:48:03.314624 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vvmlg" Dec 07 16:48:03 crc kubenswrapper[4716]: I1207 16:48:03.831757 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vvmlg"] Dec 07 16:48:04 crc kubenswrapper[4716]: I1207 16:48:04.299011 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qhf7r" event={"ID":"de3cdf40-b257-4964-85a2-577c2e596eb6","Type":"ContainerStarted","Data":"49da6350f43c73b630681f1af7e397e27f031059c78451caf69650ffeadc1ca5"} Dec 07 16:48:04 crc kubenswrapper[4716]: I1207 16:48:04.303277 4716 generic.go:334] "Generic (PLEG): container finished" podID="0cf7c849-73ad-4681-9be9-6dc84ddc6c13" containerID="185b58760bda736b02174067c6192a68676e573a752392ac42cea5db23e0e6bd" exitCode=0 Dec 07 16:48:04 crc kubenswrapper[4716]: I1207 16:48:04.303313 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vvmlg" event={"ID":"0cf7c849-73ad-4681-9be9-6dc84ddc6c13","Type":"ContainerDied","Data":"185b58760bda736b02174067c6192a68676e573a752392ac42cea5db23e0e6bd"} Dec 07 16:48:04 crc kubenswrapper[4716]: I1207 16:48:04.303332 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vvmlg" event={"ID":"0cf7c849-73ad-4681-9be9-6dc84ddc6c13","Type":"ContainerStarted","Data":"af467c333e1ecbaa0232713661a6ec474c23fe599f90ae53982319b1981c38f6"} Dec 07 16:48:07 crc kubenswrapper[4716]: I1207 16:48:07.359548 4716 generic.go:334] "Generic (PLEG): container finished" podID="de3cdf40-b257-4964-85a2-577c2e596eb6" containerID="49da6350f43c73b630681f1af7e397e27f031059c78451caf69650ffeadc1ca5" exitCode=0 Dec 07 16:48:07 crc kubenswrapper[4716]: I1207 16:48:07.359634 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qhf7r" event={"ID":"de3cdf40-b257-4964-85a2-577c2e596eb6","Type":"ContainerDied","Data":"49da6350f43c73b630681f1af7e397e27f031059c78451caf69650ffeadc1ca5"} Dec 07 16:48:12 crc kubenswrapper[4716]: I1207 16:48:12.421777 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vvmlg" event={"ID":"0cf7c849-73ad-4681-9be9-6dc84ddc6c13","Type":"ContainerStarted","Data":"32f8907f00dd09b1a34294d10efc6db91960dbf17d635caa222af51af8203347"} Dec 07 16:48:13 crc kubenswrapper[4716]: I1207 16:48:13.433499 4716 generic.go:334] "Generic (PLEG): container finished" podID="0cf7c849-73ad-4681-9be9-6dc84ddc6c13" containerID="32f8907f00dd09b1a34294d10efc6db91960dbf17d635caa222af51af8203347" exitCode=0 Dec 07 16:48:13 crc kubenswrapper[4716]: I1207 16:48:13.433545 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vvmlg" event={"ID":"0cf7c849-73ad-4681-9be9-6dc84ddc6c13","Type":"ContainerDied","Data":"32f8907f00dd09b1a34294d10efc6db91960dbf17d635caa222af51af8203347"} Dec 07 16:48:31 crc kubenswrapper[4716]: E1207 16:48:31.257970 4716 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Dec 07 16:48:31 crc kubenswrapper[4716]: E1207 16:48:31.258712 4716 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-476jr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(7c4bb2e7-29d0-4548-863b-ea89d11d68a7): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 07 16:48:31 crc kubenswrapper[4716]: E1207 16:48:31.259932 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="7c4bb2e7-29d0-4548-863b-ea89d11d68a7" Dec 07 16:48:31 crc kubenswrapper[4716]: I1207 16:48:31.612574 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qhf7r" event={"ID":"de3cdf40-b257-4964-85a2-577c2e596eb6","Type":"ContainerStarted","Data":"2d3233bbed0268d3fa8d20c61edd04043602a04bd39baa2bdfeb6a0e57fea12a"} Dec 07 16:48:31 crc kubenswrapper[4716]: E1207 16:48:31.613915 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="7c4bb2e7-29d0-4548-863b-ea89d11d68a7" Dec 07 16:48:31 crc kubenswrapper[4716]: I1207 16:48:31.663546 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-qhf7r" podStartSLOduration=5.730996448 podStartE2EDuration="33.663516391s" podCreationTimestamp="2025-12-07 16:47:58 +0000 UTC" firstStartedPulling="2025-12-07 16:48:03.294203581 +0000 UTC m=+2745.984488493" lastFinishedPulling="2025-12-07 16:48:31.226723524 +0000 UTC m=+2773.917008436" observedRunningTime="2025-12-07 16:48:31.647796576 +0000 UTC m=+2774.338081488" watchObservedRunningTime="2025-12-07 16:48:31.663516391 +0000 UTC m=+2774.353801303" Dec 07 16:48:32 crc kubenswrapper[4716]: I1207 16:48:32.626974 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vvmlg" event={"ID":"0cf7c849-73ad-4681-9be9-6dc84ddc6c13","Type":"ContainerStarted","Data":"0125d0434b91dc211b556f824b4d23c1f5dfa2b6ff6b15efc6d7498c9afbe87b"} Dec 07 16:48:32 crc kubenswrapper[4716]: I1207 16:48:32.643047 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-vvmlg" podStartSLOduration=3.337838489 podStartE2EDuration="30.643032523s" podCreationTimestamp="2025-12-07 16:48:02 +0000 UTC" firstStartedPulling="2025-12-07 16:48:04.305185693 +0000 UTC m=+2746.995470605" lastFinishedPulling="2025-12-07 16:48:31.610379727 +0000 UTC m=+2774.300664639" observedRunningTime="2025-12-07 16:48:32.641554814 +0000 UTC m=+2775.331839726" watchObservedRunningTime="2025-12-07 16:48:32.643032523 +0000 UTC m=+2775.333317435" Dec 07 16:48:33 crc kubenswrapper[4716]: I1207 16:48:33.315958 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-vvmlg" Dec 07 16:48:33 crc kubenswrapper[4716]: I1207 16:48:33.316368 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-vvmlg" Dec 07 16:48:34 crc kubenswrapper[4716]: I1207 16:48:34.383413 4716 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-vvmlg" podUID="0cf7c849-73ad-4681-9be9-6dc84ddc6c13" containerName="registry-server" probeResult="failure" output=< Dec 07 16:48:34 crc kubenswrapper[4716]: timeout: failed to connect service ":50051" within 1s Dec 07 16:48:34 crc kubenswrapper[4716]: > Dec 07 16:48:39 crc kubenswrapper[4716]: I1207 16:48:39.242481 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-qhf7r" Dec 07 16:48:39 crc kubenswrapper[4716]: I1207 16:48:39.243206 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-qhf7r" Dec 07 16:48:40 crc kubenswrapper[4716]: I1207 16:48:40.285627 4716 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-qhf7r" podUID="de3cdf40-b257-4964-85a2-577c2e596eb6" containerName="registry-server" probeResult="failure" output=< Dec 07 16:48:40 crc kubenswrapper[4716]: timeout: failed to connect service ":50051" within 1s Dec 07 16:48:40 crc kubenswrapper[4716]: > Dec 07 16:48:43 crc kubenswrapper[4716]: I1207 16:48:43.366652 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-vvmlg" Dec 07 16:48:43 crc kubenswrapper[4716]: I1207 16:48:43.414629 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-vvmlg" Dec 07 16:48:43 crc kubenswrapper[4716]: I1207 16:48:43.608676 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vvmlg"] Dec 07 16:48:44 crc kubenswrapper[4716]: I1207 16:48:44.745068 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-vvmlg" podUID="0cf7c849-73ad-4681-9be9-6dc84ddc6c13" containerName="registry-server" containerID="cri-o://0125d0434b91dc211b556f824b4d23c1f5dfa2b6ff6b15efc6d7498c9afbe87b" gracePeriod=2 Dec 07 16:48:45 crc kubenswrapper[4716]: I1207 16:48:45.138692 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 07 16:48:45 crc kubenswrapper[4716]: I1207 16:48:45.312273 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vvmlg" Dec 07 16:48:45 crc kubenswrapper[4716]: I1207 16:48:45.501327 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0cf7c849-73ad-4681-9be9-6dc84ddc6c13-catalog-content\") pod \"0cf7c849-73ad-4681-9be9-6dc84ddc6c13\" (UID: \"0cf7c849-73ad-4681-9be9-6dc84ddc6c13\") " Dec 07 16:48:45 crc kubenswrapper[4716]: I1207 16:48:45.501818 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0cf7c849-73ad-4681-9be9-6dc84ddc6c13-utilities\") pod \"0cf7c849-73ad-4681-9be9-6dc84ddc6c13\" (UID: \"0cf7c849-73ad-4681-9be9-6dc84ddc6c13\") " Dec 07 16:48:45 crc kubenswrapper[4716]: I1207 16:48:45.501896 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8n8qc\" (UniqueName: \"kubernetes.io/projected/0cf7c849-73ad-4681-9be9-6dc84ddc6c13-kube-api-access-8n8qc\") pod \"0cf7c849-73ad-4681-9be9-6dc84ddc6c13\" (UID: \"0cf7c849-73ad-4681-9be9-6dc84ddc6c13\") " Dec 07 16:48:45 crc kubenswrapper[4716]: I1207 16:48:45.502699 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0cf7c849-73ad-4681-9be9-6dc84ddc6c13-utilities" (OuterVolumeSpecName: "utilities") pod "0cf7c849-73ad-4681-9be9-6dc84ddc6c13" (UID: "0cf7c849-73ad-4681-9be9-6dc84ddc6c13"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:48:45 crc kubenswrapper[4716]: I1207 16:48:45.506476 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0cf7c849-73ad-4681-9be9-6dc84ddc6c13-kube-api-access-8n8qc" (OuterVolumeSpecName: "kube-api-access-8n8qc") pod "0cf7c849-73ad-4681-9be9-6dc84ddc6c13" (UID: "0cf7c849-73ad-4681-9be9-6dc84ddc6c13"). InnerVolumeSpecName "kube-api-access-8n8qc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:48:45 crc kubenswrapper[4716]: I1207 16:48:45.532462 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0cf7c849-73ad-4681-9be9-6dc84ddc6c13-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0cf7c849-73ad-4681-9be9-6dc84ddc6c13" (UID: "0cf7c849-73ad-4681-9be9-6dc84ddc6c13"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:48:45 crc kubenswrapper[4716]: I1207 16:48:45.604406 4716 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0cf7c849-73ad-4681-9be9-6dc84ddc6c13-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 07 16:48:45 crc kubenswrapper[4716]: I1207 16:48:45.604438 4716 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0cf7c849-73ad-4681-9be9-6dc84ddc6c13-utilities\") on node \"crc\" DevicePath \"\"" Dec 07 16:48:45 crc kubenswrapper[4716]: I1207 16:48:45.604448 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8n8qc\" (UniqueName: \"kubernetes.io/projected/0cf7c849-73ad-4681-9be9-6dc84ddc6c13-kube-api-access-8n8qc\") on node \"crc\" DevicePath \"\"" Dec 07 16:48:45 crc kubenswrapper[4716]: I1207 16:48:45.755905 4716 generic.go:334] "Generic (PLEG): container finished" podID="0cf7c849-73ad-4681-9be9-6dc84ddc6c13" containerID="0125d0434b91dc211b556f824b4d23c1f5dfa2b6ff6b15efc6d7498c9afbe87b" exitCode=0 Dec 07 16:48:45 crc kubenswrapper[4716]: I1207 16:48:45.755962 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vvmlg" event={"ID":"0cf7c849-73ad-4681-9be9-6dc84ddc6c13","Type":"ContainerDied","Data":"0125d0434b91dc211b556f824b4d23c1f5dfa2b6ff6b15efc6d7498c9afbe87b"} Dec 07 16:48:45 crc kubenswrapper[4716]: I1207 16:48:45.755998 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vvmlg" event={"ID":"0cf7c849-73ad-4681-9be9-6dc84ddc6c13","Type":"ContainerDied","Data":"af467c333e1ecbaa0232713661a6ec474c23fe599f90ae53982319b1981c38f6"} Dec 07 16:48:45 crc kubenswrapper[4716]: I1207 16:48:45.756022 4716 scope.go:117] "RemoveContainer" containerID="0125d0434b91dc211b556f824b4d23c1f5dfa2b6ff6b15efc6d7498c9afbe87b" Dec 07 16:48:45 crc kubenswrapper[4716]: I1207 16:48:45.755963 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vvmlg" Dec 07 16:48:45 crc kubenswrapper[4716]: I1207 16:48:45.782190 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vvmlg"] Dec 07 16:48:45 crc kubenswrapper[4716]: I1207 16:48:45.791116 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-vvmlg"] Dec 07 16:48:45 crc kubenswrapper[4716]: I1207 16:48:45.791985 4716 scope.go:117] "RemoveContainer" containerID="32f8907f00dd09b1a34294d10efc6db91960dbf17d635caa222af51af8203347" Dec 07 16:48:45 crc kubenswrapper[4716]: I1207 16:48:45.830164 4716 scope.go:117] "RemoveContainer" containerID="185b58760bda736b02174067c6192a68676e573a752392ac42cea5db23e0e6bd" Dec 07 16:48:45 crc kubenswrapper[4716]: I1207 16:48:45.850812 4716 scope.go:117] "RemoveContainer" containerID="0125d0434b91dc211b556f824b4d23c1f5dfa2b6ff6b15efc6d7498c9afbe87b" Dec 07 16:48:45 crc kubenswrapper[4716]: E1207 16:48:45.851375 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0125d0434b91dc211b556f824b4d23c1f5dfa2b6ff6b15efc6d7498c9afbe87b\": container with ID starting with 0125d0434b91dc211b556f824b4d23c1f5dfa2b6ff6b15efc6d7498c9afbe87b not found: ID does not exist" containerID="0125d0434b91dc211b556f824b4d23c1f5dfa2b6ff6b15efc6d7498c9afbe87b" Dec 07 16:48:45 crc kubenswrapper[4716]: I1207 16:48:45.851413 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0125d0434b91dc211b556f824b4d23c1f5dfa2b6ff6b15efc6d7498c9afbe87b"} err="failed to get container status \"0125d0434b91dc211b556f824b4d23c1f5dfa2b6ff6b15efc6d7498c9afbe87b\": rpc error: code = NotFound desc = could not find container \"0125d0434b91dc211b556f824b4d23c1f5dfa2b6ff6b15efc6d7498c9afbe87b\": container with ID starting with 0125d0434b91dc211b556f824b4d23c1f5dfa2b6ff6b15efc6d7498c9afbe87b not found: ID does not exist" Dec 07 16:48:45 crc kubenswrapper[4716]: I1207 16:48:45.851438 4716 scope.go:117] "RemoveContainer" containerID="32f8907f00dd09b1a34294d10efc6db91960dbf17d635caa222af51af8203347" Dec 07 16:48:45 crc kubenswrapper[4716]: E1207 16:48:45.851745 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"32f8907f00dd09b1a34294d10efc6db91960dbf17d635caa222af51af8203347\": container with ID starting with 32f8907f00dd09b1a34294d10efc6db91960dbf17d635caa222af51af8203347 not found: ID does not exist" containerID="32f8907f00dd09b1a34294d10efc6db91960dbf17d635caa222af51af8203347" Dec 07 16:48:45 crc kubenswrapper[4716]: I1207 16:48:45.851771 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32f8907f00dd09b1a34294d10efc6db91960dbf17d635caa222af51af8203347"} err="failed to get container status \"32f8907f00dd09b1a34294d10efc6db91960dbf17d635caa222af51af8203347\": rpc error: code = NotFound desc = could not find container \"32f8907f00dd09b1a34294d10efc6db91960dbf17d635caa222af51af8203347\": container with ID starting with 32f8907f00dd09b1a34294d10efc6db91960dbf17d635caa222af51af8203347 not found: ID does not exist" Dec 07 16:48:45 crc kubenswrapper[4716]: I1207 16:48:45.851786 4716 scope.go:117] "RemoveContainer" containerID="185b58760bda736b02174067c6192a68676e573a752392ac42cea5db23e0e6bd" Dec 07 16:48:45 crc kubenswrapper[4716]: E1207 16:48:45.852111 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"185b58760bda736b02174067c6192a68676e573a752392ac42cea5db23e0e6bd\": container with ID starting with 185b58760bda736b02174067c6192a68676e573a752392ac42cea5db23e0e6bd not found: ID does not exist" containerID="185b58760bda736b02174067c6192a68676e573a752392ac42cea5db23e0e6bd" Dec 07 16:48:45 crc kubenswrapper[4716]: I1207 16:48:45.852131 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"185b58760bda736b02174067c6192a68676e573a752392ac42cea5db23e0e6bd"} err="failed to get container status \"185b58760bda736b02174067c6192a68676e573a752392ac42cea5db23e0e6bd\": rpc error: code = NotFound desc = could not find container \"185b58760bda736b02174067c6192a68676e573a752392ac42cea5db23e0e6bd\": container with ID starting with 185b58760bda736b02174067c6192a68676e573a752392ac42cea5db23e0e6bd not found: ID does not exist" Dec 07 16:48:46 crc kubenswrapper[4716]: I1207 16:48:46.768737 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"7c4bb2e7-29d0-4548-863b-ea89d11d68a7","Type":"ContainerStarted","Data":"73fea1b696cd9f27104920314586e5afab71cee2c49a27ca498d06a78f57428d"} Dec 07 16:48:46 crc kubenswrapper[4716]: I1207 16:48:46.801674 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=4.049662973 podStartE2EDuration="52.801650004s" podCreationTimestamp="2025-12-07 16:47:54 +0000 UTC" firstStartedPulling="2025-12-07 16:47:56.381279829 +0000 UTC m=+2739.071564741" lastFinishedPulling="2025-12-07 16:48:45.13326686 +0000 UTC m=+2787.823551772" observedRunningTime="2025-12-07 16:48:46.790312836 +0000 UTC m=+2789.480597758" watchObservedRunningTime="2025-12-07 16:48:46.801650004 +0000 UTC m=+2789.491934946" Dec 07 16:48:47 crc kubenswrapper[4716]: I1207 16:48:47.678743 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0cf7c849-73ad-4681-9be9-6dc84ddc6c13" path="/var/lib/kubelet/pods/0cf7c849-73ad-4681-9be9-6dc84ddc6c13/volumes" Dec 07 16:48:49 crc kubenswrapper[4716]: I1207 16:48:49.294838 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-qhf7r" Dec 07 16:48:49 crc kubenswrapper[4716]: I1207 16:48:49.349012 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-qhf7r" Dec 07 16:48:49 crc kubenswrapper[4716]: I1207 16:48:49.527959 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qhf7r"] Dec 07 16:48:50 crc kubenswrapper[4716]: I1207 16:48:50.824108 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-qhf7r" podUID="de3cdf40-b257-4964-85a2-577c2e596eb6" containerName="registry-server" containerID="cri-o://2d3233bbed0268d3fa8d20c61edd04043602a04bd39baa2bdfeb6a0e57fea12a" gracePeriod=2 Dec 07 16:48:51 crc kubenswrapper[4716]: I1207 16:48:51.246581 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qhf7r" Dec 07 16:48:51 crc kubenswrapper[4716]: I1207 16:48:51.415122 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de3cdf40-b257-4964-85a2-577c2e596eb6-catalog-content\") pod \"de3cdf40-b257-4964-85a2-577c2e596eb6\" (UID: \"de3cdf40-b257-4964-85a2-577c2e596eb6\") " Dec 07 16:48:51 crc kubenswrapper[4716]: I1207 16:48:51.415324 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-44qsk\" (UniqueName: \"kubernetes.io/projected/de3cdf40-b257-4964-85a2-577c2e596eb6-kube-api-access-44qsk\") pod \"de3cdf40-b257-4964-85a2-577c2e596eb6\" (UID: \"de3cdf40-b257-4964-85a2-577c2e596eb6\") " Dec 07 16:48:51 crc kubenswrapper[4716]: I1207 16:48:51.415403 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de3cdf40-b257-4964-85a2-577c2e596eb6-utilities\") pod \"de3cdf40-b257-4964-85a2-577c2e596eb6\" (UID: \"de3cdf40-b257-4964-85a2-577c2e596eb6\") " Dec 07 16:48:51 crc kubenswrapper[4716]: I1207 16:48:51.416276 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de3cdf40-b257-4964-85a2-577c2e596eb6-utilities" (OuterVolumeSpecName: "utilities") pod "de3cdf40-b257-4964-85a2-577c2e596eb6" (UID: "de3cdf40-b257-4964-85a2-577c2e596eb6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:48:51 crc kubenswrapper[4716]: I1207 16:48:51.424289 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de3cdf40-b257-4964-85a2-577c2e596eb6-kube-api-access-44qsk" (OuterVolumeSpecName: "kube-api-access-44qsk") pod "de3cdf40-b257-4964-85a2-577c2e596eb6" (UID: "de3cdf40-b257-4964-85a2-577c2e596eb6"). InnerVolumeSpecName "kube-api-access-44qsk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:48:51 crc kubenswrapper[4716]: I1207 16:48:51.517684 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-44qsk\" (UniqueName: \"kubernetes.io/projected/de3cdf40-b257-4964-85a2-577c2e596eb6-kube-api-access-44qsk\") on node \"crc\" DevicePath \"\"" Dec 07 16:48:51 crc kubenswrapper[4716]: I1207 16:48:51.517717 4716 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de3cdf40-b257-4964-85a2-577c2e596eb6-utilities\") on node \"crc\" DevicePath \"\"" Dec 07 16:48:51 crc kubenswrapper[4716]: I1207 16:48:51.532124 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de3cdf40-b257-4964-85a2-577c2e596eb6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "de3cdf40-b257-4964-85a2-577c2e596eb6" (UID: "de3cdf40-b257-4964-85a2-577c2e596eb6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:48:51 crc kubenswrapper[4716]: I1207 16:48:51.619005 4716 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de3cdf40-b257-4964-85a2-577c2e596eb6-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 07 16:48:51 crc kubenswrapper[4716]: I1207 16:48:51.834366 4716 generic.go:334] "Generic (PLEG): container finished" podID="de3cdf40-b257-4964-85a2-577c2e596eb6" containerID="2d3233bbed0268d3fa8d20c61edd04043602a04bd39baa2bdfeb6a0e57fea12a" exitCode=0 Dec 07 16:48:51 crc kubenswrapper[4716]: I1207 16:48:51.834414 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qhf7r" event={"ID":"de3cdf40-b257-4964-85a2-577c2e596eb6","Type":"ContainerDied","Data":"2d3233bbed0268d3fa8d20c61edd04043602a04bd39baa2bdfeb6a0e57fea12a"} Dec 07 16:48:51 crc kubenswrapper[4716]: I1207 16:48:51.834441 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qhf7r" Dec 07 16:48:51 crc kubenswrapper[4716]: I1207 16:48:51.834462 4716 scope.go:117] "RemoveContainer" containerID="2d3233bbed0268d3fa8d20c61edd04043602a04bd39baa2bdfeb6a0e57fea12a" Dec 07 16:48:51 crc kubenswrapper[4716]: I1207 16:48:51.834449 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qhf7r" event={"ID":"de3cdf40-b257-4964-85a2-577c2e596eb6","Type":"ContainerDied","Data":"1e5fe7b9917bd1d102fc764ae05e78955f54259151224500c1a1eecb50ed95aa"} Dec 07 16:48:51 crc kubenswrapper[4716]: I1207 16:48:51.870250 4716 scope.go:117] "RemoveContainer" containerID="49da6350f43c73b630681f1af7e397e27f031059c78451caf69650ffeadc1ca5" Dec 07 16:48:51 crc kubenswrapper[4716]: I1207 16:48:51.871842 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qhf7r"] Dec 07 16:48:51 crc kubenswrapper[4716]: I1207 16:48:51.883222 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-qhf7r"] Dec 07 16:48:51 crc kubenswrapper[4716]: I1207 16:48:51.954651 4716 scope.go:117] "RemoveContainer" containerID="3ffef06fe3c15c87cdad2c881969d0af9161e55d7acadc67ea6a788622a6e94f" Dec 07 16:48:51 crc kubenswrapper[4716]: I1207 16:48:51.996378 4716 scope.go:117] "RemoveContainer" containerID="2d3233bbed0268d3fa8d20c61edd04043602a04bd39baa2bdfeb6a0e57fea12a" Dec 07 16:48:51 crc kubenswrapper[4716]: E1207 16:48:51.997226 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d3233bbed0268d3fa8d20c61edd04043602a04bd39baa2bdfeb6a0e57fea12a\": container with ID starting with 2d3233bbed0268d3fa8d20c61edd04043602a04bd39baa2bdfeb6a0e57fea12a not found: ID does not exist" containerID="2d3233bbed0268d3fa8d20c61edd04043602a04bd39baa2bdfeb6a0e57fea12a" Dec 07 16:48:51 crc kubenswrapper[4716]: I1207 16:48:51.997271 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d3233bbed0268d3fa8d20c61edd04043602a04bd39baa2bdfeb6a0e57fea12a"} err="failed to get container status \"2d3233bbed0268d3fa8d20c61edd04043602a04bd39baa2bdfeb6a0e57fea12a\": rpc error: code = NotFound desc = could not find container \"2d3233bbed0268d3fa8d20c61edd04043602a04bd39baa2bdfeb6a0e57fea12a\": container with ID starting with 2d3233bbed0268d3fa8d20c61edd04043602a04bd39baa2bdfeb6a0e57fea12a not found: ID does not exist" Dec 07 16:48:51 crc kubenswrapper[4716]: I1207 16:48:51.997296 4716 scope.go:117] "RemoveContainer" containerID="49da6350f43c73b630681f1af7e397e27f031059c78451caf69650ffeadc1ca5" Dec 07 16:48:51 crc kubenswrapper[4716]: E1207 16:48:51.997874 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"49da6350f43c73b630681f1af7e397e27f031059c78451caf69650ffeadc1ca5\": container with ID starting with 49da6350f43c73b630681f1af7e397e27f031059c78451caf69650ffeadc1ca5 not found: ID does not exist" containerID="49da6350f43c73b630681f1af7e397e27f031059c78451caf69650ffeadc1ca5" Dec 07 16:48:51 crc kubenswrapper[4716]: I1207 16:48:51.997902 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49da6350f43c73b630681f1af7e397e27f031059c78451caf69650ffeadc1ca5"} err="failed to get container status \"49da6350f43c73b630681f1af7e397e27f031059c78451caf69650ffeadc1ca5\": rpc error: code = NotFound desc = could not find container \"49da6350f43c73b630681f1af7e397e27f031059c78451caf69650ffeadc1ca5\": container with ID starting with 49da6350f43c73b630681f1af7e397e27f031059c78451caf69650ffeadc1ca5 not found: ID does not exist" Dec 07 16:48:51 crc kubenswrapper[4716]: I1207 16:48:51.997919 4716 scope.go:117] "RemoveContainer" containerID="3ffef06fe3c15c87cdad2c881969d0af9161e55d7acadc67ea6a788622a6e94f" Dec 07 16:48:51 crc kubenswrapper[4716]: E1207 16:48:51.998113 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3ffef06fe3c15c87cdad2c881969d0af9161e55d7acadc67ea6a788622a6e94f\": container with ID starting with 3ffef06fe3c15c87cdad2c881969d0af9161e55d7acadc67ea6a788622a6e94f not found: ID does not exist" containerID="3ffef06fe3c15c87cdad2c881969d0af9161e55d7acadc67ea6a788622a6e94f" Dec 07 16:48:51 crc kubenswrapper[4716]: I1207 16:48:51.998134 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ffef06fe3c15c87cdad2c881969d0af9161e55d7acadc67ea6a788622a6e94f"} err="failed to get container status \"3ffef06fe3c15c87cdad2c881969d0af9161e55d7acadc67ea6a788622a6e94f\": rpc error: code = NotFound desc = could not find container \"3ffef06fe3c15c87cdad2c881969d0af9161e55d7acadc67ea6a788622a6e94f\": container with ID starting with 3ffef06fe3c15c87cdad2c881969d0af9161e55d7acadc67ea6a788622a6e94f not found: ID does not exist" Dec 07 16:48:53 crc kubenswrapper[4716]: I1207 16:48:53.667791 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de3cdf40-b257-4964-85a2-577c2e596eb6" path="/var/lib/kubelet/pods/de3cdf40-b257-4964-85a2-577c2e596eb6/volumes" Dec 07 16:48:54 crc kubenswrapper[4716]: I1207 16:48:54.936400 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-td7jx"] Dec 07 16:48:54 crc kubenswrapper[4716]: E1207 16:48:54.937125 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de3cdf40-b257-4964-85a2-577c2e596eb6" containerName="extract-content" Dec 07 16:48:54 crc kubenswrapper[4716]: I1207 16:48:54.937137 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="de3cdf40-b257-4964-85a2-577c2e596eb6" containerName="extract-content" Dec 07 16:48:54 crc kubenswrapper[4716]: E1207 16:48:54.937150 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cf7c849-73ad-4681-9be9-6dc84ddc6c13" containerName="extract-content" Dec 07 16:48:54 crc kubenswrapper[4716]: I1207 16:48:54.937155 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cf7c849-73ad-4681-9be9-6dc84ddc6c13" containerName="extract-content" Dec 07 16:48:54 crc kubenswrapper[4716]: E1207 16:48:54.937172 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cf7c849-73ad-4681-9be9-6dc84ddc6c13" containerName="extract-utilities" Dec 07 16:48:54 crc kubenswrapper[4716]: I1207 16:48:54.937180 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cf7c849-73ad-4681-9be9-6dc84ddc6c13" containerName="extract-utilities" Dec 07 16:48:54 crc kubenswrapper[4716]: E1207 16:48:54.937192 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de3cdf40-b257-4964-85a2-577c2e596eb6" containerName="extract-utilities" Dec 07 16:48:54 crc kubenswrapper[4716]: I1207 16:48:54.937198 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="de3cdf40-b257-4964-85a2-577c2e596eb6" containerName="extract-utilities" Dec 07 16:48:54 crc kubenswrapper[4716]: E1207 16:48:54.937216 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de3cdf40-b257-4964-85a2-577c2e596eb6" containerName="registry-server" Dec 07 16:48:54 crc kubenswrapper[4716]: I1207 16:48:54.937221 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="de3cdf40-b257-4964-85a2-577c2e596eb6" containerName="registry-server" Dec 07 16:48:54 crc kubenswrapper[4716]: E1207 16:48:54.937237 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cf7c849-73ad-4681-9be9-6dc84ddc6c13" containerName="registry-server" Dec 07 16:48:54 crc kubenswrapper[4716]: I1207 16:48:54.937243 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cf7c849-73ad-4681-9be9-6dc84ddc6c13" containerName="registry-server" Dec 07 16:48:54 crc kubenswrapper[4716]: I1207 16:48:54.937433 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cf7c849-73ad-4681-9be9-6dc84ddc6c13" containerName="registry-server" Dec 07 16:48:54 crc kubenswrapper[4716]: I1207 16:48:54.937451 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="de3cdf40-b257-4964-85a2-577c2e596eb6" containerName="registry-server" Dec 07 16:48:54 crc kubenswrapper[4716]: I1207 16:48:54.938742 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-td7jx" Dec 07 16:48:54 crc kubenswrapper[4716]: I1207 16:48:54.955972 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-td7jx"] Dec 07 16:48:55 crc kubenswrapper[4716]: I1207 16:48:55.098703 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6qhgq\" (UniqueName: \"kubernetes.io/projected/d9b5f365-75ff-4ee6-904f-fc2996731f37-kube-api-access-6qhgq\") pod \"certified-operators-td7jx\" (UID: \"d9b5f365-75ff-4ee6-904f-fc2996731f37\") " pod="openshift-marketplace/certified-operators-td7jx" Dec 07 16:48:55 crc kubenswrapper[4716]: I1207 16:48:55.099174 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9b5f365-75ff-4ee6-904f-fc2996731f37-catalog-content\") pod \"certified-operators-td7jx\" (UID: \"d9b5f365-75ff-4ee6-904f-fc2996731f37\") " pod="openshift-marketplace/certified-operators-td7jx" Dec 07 16:48:55 crc kubenswrapper[4716]: I1207 16:48:55.099364 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9b5f365-75ff-4ee6-904f-fc2996731f37-utilities\") pod \"certified-operators-td7jx\" (UID: \"d9b5f365-75ff-4ee6-904f-fc2996731f37\") " pod="openshift-marketplace/certified-operators-td7jx" Dec 07 16:48:55 crc kubenswrapper[4716]: I1207 16:48:55.201042 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9b5f365-75ff-4ee6-904f-fc2996731f37-utilities\") pod \"certified-operators-td7jx\" (UID: \"d9b5f365-75ff-4ee6-904f-fc2996731f37\") " pod="openshift-marketplace/certified-operators-td7jx" Dec 07 16:48:55 crc kubenswrapper[4716]: I1207 16:48:55.201553 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6qhgq\" (UniqueName: \"kubernetes.io/projected/d9b5f365-75ff-4ee6-904f-fc2996731f37-kube-api-access-6qhgq\") pod \"certified-operators-td7jx\" (UID: \"d9b5f365-75ff-4ee6-904f-fc2996731f37\") " pod="openshift-marketplace/certified-operators-td7jx" Dec 07 16:48:55 crc kubenswrapper[4716]: I1207 16:48:55.201642 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9b5f365-75ff-4ee6-904f-fc2996731f37-utilities\") pod \"certified-operators-td7jx\" (UID: \"d9b5f365-75ff-4ee6-904f-fc2996731f37\") " pod="openshift-marketplace/certified-operators-td7jx" Dec 07 16:48:55 crc kubenswrapper[4716]: I1207 16:48:55.201958 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9b5f365-75ff-4ee6-904f-fc2996731f37-catalog-content\") pod \"certified-operators-td7jx\" (UID: \"d9b5f365-75ff-4ee6-904f-fc2996731f37\") " pod="openshift-marketplace/certified-operators-td7jx" Dec 07 16:48:55 crc kubenswrapper[4716]: I1207 16:48:55.202302 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9b5f365-75ff-4ee6-904f-fc2996731f37-catalog-content\") pod \"certified-operators-td7jx\" (UID: \"d9b5f365-75ff-4ee6-904f-fc2996731f37\") " pod="openshift-marketplace/certified-operators-td7jx" Dec 07 16:48:55 crc kubenswrapper[4716]: I1207 16:48:55.233806 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6qhgq\" (UniqueName: \"kubernetes.io/projected/d9b5f365-75ff-4ee6-904f-fc2996731f37-kube-api-access-6qhgq\") pod \"certified-operators-td7jx\" (UID: \"d9b5f365-75ff-4ee6-904f-fc2996731f37\") " pod="openshift-marketplace/certified-operators-td7jx" Dec 07 16:48:55 crc kubenswrapper[4716]: I1207 16:48:55.268224 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-td7jx" Dec 07 16:48:55 crc kubenswrapper[4716]: I1207 16:48:55.781198 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-td7jx"] Dec 07 16:48:55 crc kubenswrapper[4716]: I1207 16:48:55.866387 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-td7jx" event={"ID":"d9b5f365-75ff-4ee6-904f-fc2996731f37","Type":"ContainerStarted","Data":"1630fe83a86f0e01ac55294b6ba87bcfa7a48a2863fedc5f41d29f2bb3659d43"} Dec 07 16:48:56 crc kubenswrapper[4716]: I1207 16:48:56.875188 4716 generic.go:334] "Generic (PLEG): container finished" podID="d9b5f365-75ff-4ee6-904f-fc2996731f37" containerID="79443cae1a8973c53920773a94ed28197e3e01580da2de00c6417870aa68d0f7" exitCode=0 Dec 07 16:48:56 crc kubenswrapper[4716]: I1207 16:48:56.875345 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-td7jx" event={"ID":"d9b5f365-75ff-4ee6-904f-fc2996731f37","Type":"ContainerDied","Data":"79443cae1a8973c53920773a94ed28197e3e01580da2de00c6417870aa68d0f7"} Dec 07 16:48:57 crc kubenswrapper[4716]: I1207 16:48:57.885935 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-td7jx" event={"ID":"d9b5f365-75ff-4ee6-904f-fc2996731f37","Type":"ContainerStarted","Data":"c6d8216419e86f6b8a1d8d42e4198e2ce1689f813679d2df60e9520c946e3538"} Dec 07 16:48:58 crc kubenswrapper[4716]: I1207 16:48:58.918497 4716 generic.go:334] "Generic (PLEG): container finished" podID="d9b5f365-75ff-4ee6-904f-fc2996731f37" containerID="c6d8216419e86f6b8a1d8d42e4198e2ce1689f813679d2df60e9520c946e3538" exitCode=0 Dec 07 16:48:58 crc kubenswrapper[4716]: I1207 16:48:58.918581 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-td7jx" event={"ID":"d9b5f365-75ff-4ee6-904f-fc2996731f37","Type":"ContainerDied","Data":"c6d8216419e86f6b8a1d8d42e4198e2ce1689f813679d2df60e9520c946e3538"} Dec 07 16:48:59 crc kubenswrapper[4716]: I1207 16:48:59.931269 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-td7jx" event={"ID":"d9b5f365-75ff-4ee6-904f-fc2996731f37","Type":"ContainerStarted","Data":"32dd80cc100777d51628d4d649bfe1590be15328784cd2cfc26baa694d43e66e"} Dec 07 16:48:59 crc kubenswrapper[4716]: I1207 16:48:59.945535 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-td7jx" podStartSLOduration=3.507764263 podStartE2EDuration="5.945520508s" podCreationTimestamp="2025-12-07 16:48:54 +0000 UTC" firstStartedPulling="2025-12-07 16:48:56.877683516 +0000 UTC m=+2799.567968428" lastFinishedPulling="2025-12-07 16:48:59.315439761 +0000 UTC m=+2802.005724673" observedRunningTime="2025-12-07 16:48:59.945333113 +0000 UTC m=+2802.635618035" watchObservedRunningTime="2025-12-07 16:48:59.945520508 +0000 UTC m=+2802.635805420" Dec 07 16:49:05 crc kubenswrapper[4716]: I1207 16:49:05.269246 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-td7jx" Dec 07 16:49:05 crc kubenswrapper[4716]: I1207 16:49:05.270029 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-td7jx" Dec 07 16:49:05 crc kubenswrapper[4716]: I1207 16:49:05.313499 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-td7jx" Dec 07 16:49:06 crc kubenswrapper[4716]: I1207 16:49:06.025575 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-td7jx" Dec 07 16:49:06 crc kubenswrapper[4716]: I1207 16:49:06.074914 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-td7jx"] Dec 07 16:49:07 crc kubenswrapper[4716]: I1207 16:49:07.996456 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-td7jx" podUID="d9b5f365-75ff-4ee6-904f-fc2996731f37" containerName="registry-server" containerID="cri-o://32dd80cc100777d51628d4d649bfe1590be15328784cd2cfc26baa694d43e66e" gracePeriod=2 Dec 07 16:49:08 crc kubenswrapper[4716]: I1207 16:49:08.489484 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-td7jx" Dec 07 16:49:08 crc kubenswrapper[4716]: I1207 16:49:08.586621 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9b5f365-75ff-4ee6-904f-fc2996731f37-utilities\") pod \"d9b5f365-75ff-4ee6-904f-fc2996731f37\" (UID: \"d9b5f365-75ff-4ee6-904f-fc2996731f37\") " Dec 07 16:49:08 crc kubenswrapper[4716]: I1207 16:49:08.586770 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9b5f365-75ff-4ee6-904f-fc2996731f37-catalog-content\") pod \"d9b5f365-75ff-4ee6-904f-fc2996731f37\" (UID: \"d9b5f365-75ff-4ee6-904f-fc2996731f37\") " Dec 07 16:49:08 crc kubenswrapper[4716]: I1207 16:49:08.586999 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6qhgq\" (UniqueName: \"kubernetes.io/projected/d9b5f365-75ff-4ee6-904f-fc2996731f37-kube-api-access-6qhgq\") pod \"d9b5f365-75ff-4ee6-904f-fc2996731f37\" (UID: \"d9b5f365-75ff-4ee6-904f-fc2996731f37\") " Dec 07 16:49:08 crc kubenswrapper[4716]: I1207 16:49:08.591554 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d9b5f365-75ff-4ee6-904f-fc2996731f37-utilities" (OuterVolumeSpecName: "utilities") pod "d9b5f365-75ff-4ee6-904f-fc2996731f37" (UID: "d9b5f365-75ff-4ee6-904f-fc2996731f37"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:49:08 crc kubenswrapper[4716]: I1207 16:49:08.593629 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9b5f365-75ff-4ee6-904f-fc2996731f37-kube-api-access-6qhgq" (OuterVolumeSpecName: "kube-api-access-6qhgq") pod "d9b5f365-75ff-4ee6-904f-fc2996731f37" (UID: "d9b5f365-75ff-4ee6-904f-fc2996731f37"). InnerVolumeSpecName "kube-api-access-6qhgq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:49:08 crc kubenswrapper[4716]: I1207 16:49:08.653158 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d9b5f365-75ff-4ee6-904f-fc2996731f37-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d9b5f365-75ff-4ee6-904f-fc2996731f37" (UID: "d9b5f365-75ff-4ee6-904f-fc2996731f37"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:49:08 crc kubenswrapper[4716]: I1207 16:49:08.689149 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6qhgq\" (UniqueName: \"kubernetes.io/projected/d9b5f365-75ff-4ee6-904f-fc2996731f37-kube-api-access-6qhgq\") on node \"crc\" DevicePath \"\"" Dec 07 16:49:08 crc kubenswrapper[4716]: I1207 16:49:08.689183 4716 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9b5f365-75ff-4ee6-904f-fc2996731f37-utilities\") on node \"crc\" DevicePath \"\"" Dec 07 16:49:08 crc kubenswrapper[4716]: I1207 16:49:08.689196 4716 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9b5f365-75ff-4ee6-904f-fc2996731f37-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 07 16:49:09 crc kubenswrapper[4716]: I1207 16:49:09.008984 4716 generic.go:334] "Generic (PLEG): container finished" podID="d9b5f365-75ff-4ee6-904f-fc2996731f37" containerID="32dd80cc100777d51628d4d649bfe1590be15328784cd2cfc26baa694d43e66e" exitCode=0 Dec 07 16:49:09 crc kubenswrapper[4716]: I1207 16:49:09.009045 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-td7jx" event={"ID":"d9b5f365-75ff-4ee6-904f-fc2996731f37","Type":"ContainerDied","Data":"32dd80cc100777d51628d4d649bfe1590be15328784cd2cfc26baa694d43e66e"} Dec 07 16:49:09 crc kubenswrapper[4716]: I1207 16:49:09.009118 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-td7jx" event={"ID":"d9b5f365-75ff-4ee6-904f-fc2996731f37","Type":"ContainerDied","Data":"1630fe83a86f0e01ac55294b6ba87bcfa7a48a2863fedc5f41d29f2bb3659d43"} Dec 07 16:49:09 crc kubenswrapper[4716]: I1207 16:49:09.009148 4716 scope.go:117] "RemoveContainer" containerID="32dd80cc100777d51628d4d649bfe1590be15328784cd2cfc26baa694d43e66e" Dec 07 16:49:09 crc kubenswrapper[4716]: I1207 16:49:09.009390 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-td7jx" Dec 07 16:49:09 crc kubenswrapper[4716]: I1207 16:49:09.032407 4716 scope.go:117] "RemoveContainer" containerID="c6d8216419e86f6b8a1d8d42e4198e2ce1689f813679d2df60e9520c946e3538" Dec 07 16:49:09 crc kubenswrapper[4716]: I1207 16:49:09.077069 4716 scope.go:117] "RemoveContainer" containerID="79443cae1a8973c53920773a94ed28197e3e01580da2de00c6417870aa68d0f7" Dec 07 16:49:09 crc kubenswrapper[4716]: I1207 16:49:09.080767 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-td7jx"] Dec 07 16:49:09 crc kubenswrapper[4716]: I1207 16:49:09.091009 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-td7jx"] Dec 07 16:49:09 crc kubenswrapper[4716]: I1207 16:49:09.099815 4716 scope.go:117] "RemoveContainer" containerID="32dd80cc100777d51628d4d649bfe1590be15328784cd2cfc26baa694d43e66e" Dec 07 16:49:09 crc kubenswrapper[4716]: E1207 16:49:09.100349 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"32dd80cc100777d51628d4d649bfe1590be15328784cd2cfc26baa694d43e66e\": container with ID starting with 32dd80cc100777d51628d4d649bfe1590be15328784cd2cfc26baa694d43e66e not found: ID does not exist" containerID="32dd80cc100777d51628d4d649bfe1590be15328784cd2cfc26baa694d43e66e" Dec 07 16:49:09 crc kubenswrapper[4716]: I1207 16:49:09.100400 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32dd80cc100777d51628d4d649bfe1590be15328784cd2cfc26baa694d43e66e"} err="failed to get container status \"32dd80cc100777d51628d4d649bfe1590be15328784cd2cfc26baa694d43e66e\": rpc error: code = NotFound desc = could not find container \"32dd80cc100777d51628d4d649bfe1590be15328784cd2cfc26baa694d43e66e\": container with ID starting with 32dd80cc100777d51628d4d649bfe1590be15328784cd2cfc26baa694d43e66e not found: ID does not exist" Dec 07 16:49:09 crc kubenswrapper[4716]: I1207 16:49:09.100432 4716 scope.go:117] "RemoveContainer" containerID="c6d8216419e86f6b8a1d8d42e4198e2ce1689f813679d2df60e9520c946e3538" Dec 07 16:49:09 crc kubenswrapper[4716]: E1207 16:49:09.101041 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c6d8216419e86f6b8a1d8d42e4198e2ce1689f813679d2df60e9520c946e3538\": container with ID starting with c6d8216419e86f6b8a1d8d42e4198e2ce1689f813679d2df60e9520c946e3538 not found: ID does not exist" containerID="c6d8216419e86f6b8a1d8d42e4198e2ce1689f813679d2df60e9520c946e3538" Dec 07 16:49:09 crc kubenswrapper[4716]: I1207 16:49:09.101103 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6d8216419e86f6b8a1d8d42e4198e2ce1689f813679d2df60e9520c946e3538"} err="failed to get container status \"c6d8216419e86f6b8a1d8d42e4198e2ce1689f813679d2df60e9520c946e3538\": rpc error: code = NotFound desc = could not find container \"c6d8216419e86f6b8a1d8d42e4198e2ce1689f813679d2df60e9520c946e3538\": container with ID starting with c6d8216419e86f6b8a1d8d42e4198e2ce1689f813679d2df60e9520c946e3538 not found: ID does not exist" Dec 07 16:49:09 crc kubenswrapper[4716]: I1207 16:49:09.101133 4716 scope.go:117] "RemoveContainer" containerID="79443cae1a8973c53920773a94ed28197e3e01580da2de00c6417870aa68d0f7" Dec 07 16:49:09 crc kubenswrapper[4716]: E1207 16:49:09.101512 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"79443cae1a8973c53920773a94ed28197e3e01580da2de00c6417870aa68d0f7\": container with ID starting with 79443cae1a8973c53920773a94ed28197e3e01580da2de00c6417870aa68d0f7 not found: ID does not exist" containerID="79443cae1a8973c53920773a94ed28197e3e01580da2de00c6417870aa68d0f7" Dec 07 16:49:09 crc kubenswrapper[4716]: I1207 16:49:09.101544 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79443cae1a8973c53920773a94ed28197e3e01580da2de00c6417870aa68d0f7"} err="failed to get container status \"79443cae1a8973c53920773a94ed28197e3e01580da2de00c6417870aa68d0f7\": rpc error: code = NotFound desc = could not find container \"79443cae1a8973c53920773a94ed28197e3e01580da2de00c6417870aa68d0f7\": container with ID starting with 79443cae1a8973c53920773a94ed28197e3e01580da2de00c6417870aa68d0f7 not found: ID does not exist" Dec 07 16:49:09 crc kubenswrapper[4716]: I1207 16:49:09.677148 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d9b5f365-75ff-4ee6-904f-fc2996731f37" path="/var/lib/kubelet/pods/d9b5f365-75ff-4ee6-904f-fc2996731f37/volumes" Dec 07 16:49:52 crc kubenswrapper[4716]: I1207 16:49:52.761580 4716 patch_prober.go:28] interesting pod/machine-config-daemon-zcxxp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 16:49:52 crc kubenswrapper[4716]: I1207 16:49:52.762304 4716 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 16:50:22 crc kubenswrapper[4716]: I1207 16:50:22.761203 4716 patch_prober.go:28] interesting pod/machine-config-daemon-zcxxp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 16:50:22 crc kubenswrapper[4716]: I1207 16:50:22.761917 4716 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 16:50:52 crc kubenswrapper[4716]: I1207 16:50:52.761411 4716 patch_prober.go:28] interesting pod/machine-config-daemon-zcxxp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 16:50:52 crc kubenswrapper[4716]: I1207 16:50:52.762246 4716 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 16:50:52 crc kubenswrapper[4716]: I1207 16:50:52.762299 4716 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" Dec 07 16:50:52 crc kubenswrapper[4716]: I1207 16:50:52.763058 4716 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"be6a29d905d404203a7ce5b0088583826cc3981f161994b9094c7227407ce5aa"} pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 07 16:50:52 crc kubenswrapper[4716]: I1207 16:50:52.763149 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" containerName="machine-config-daemon" containerID="cri-o://be6a29d905d404203a7ce5b0088583826cc3981f161994b9094c7227407ce5aa" gracePeriod=600 Dec 07 16:50:52 crc kubenswrapper[4716]: I1207 16:50:52.985838 4716 generic.go:334] "Generic (PLEG): container finished" podID="c15b59eb-565d-4556-a4ce-75afdf159dc8" containerID="be6a29d905d404203a7ce5b0088583826cc3981f161994b9094c7227407ce5aa" exitCode=0 Dec 07 16:50:52 crc kubenswrapper[4716]: I1207 16:50:52.985891 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" event={"ID":"c15b59eb-565d-4556-a4ce-75afdf159dc8","Type":"ContainerDied","Data":"be6a29d905d404203a7ce5b0088583826cc3981f161994b9094c7227407ce5aa"} Dec 07 16:50:52 crc kubenswrapper[4716]: I1207 16:50:52.985934 4716 scope.go:117] "RemoveContainer" containerID="3eb63b7962def203a31ba32ba16ac93470e254c2a4edb4a4dd2033b290e2f217" Dec 07 16:50:53 crc kubenswrapper[4716]: I1207 16:50:53.998167 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" event={"ID":"c15b59eb-565d-4556-a4ce-75afdf159dc8","Type":"ContainerStarted","Data":"320114e03a98e202a688ac2d6205631a9f8d9952773c6722837647cd3ea2b74d"} Dec 07 16:52:58 crc kubenswrapper[4716]: I1207 16:52:58.966363 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-pmqgv"] Dec 07 16:52:58 crc kubenswrapper[4716]: E1207 16:52:58.968032 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9b5f365-75ff-4ee6-904f-fc2996731f37" containerName="extract-utilities" Dec 07 16:52:58 crc kubenswrapper[4716]: I1207 16:52:58.968068 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9b5f365-75ff-4ee6-904f-fc2996731f37" containerName="extract-utilities" Dec 07 16:52:58 crc kubenswrapper[4716]: E1207 16:52:58.968658 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9b5f365-75ff-4ee6-904f-fc2996731f37" containerName="extract-content" Dec 07 16:52:58 crc kubenswrapper[4716]: I1207 16:52:58.968719 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9b5f365-75ff-4ee6-904f-fc2996731f37" containerName="extract-content" Dec 07 16:52:58 crc kubenswrapper[4716]: E1207 16:52:58.968753 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9b5f365-75ff-4ee6-904f-fc2996731f37" containerName="registry-server" Dec 07 16:52:58 crc kubenswrapper[4716]: I1207 16:52:58.968773 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9b5f365-75ff-4ee6-904f-fc2996731f37" containerName="registry-server" Dec 07 16:52:58 crc kubenswrapper[4716]: I1207 16:52:58.969191 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9b5f365-75ff-4ee6-904f-fc2996731f37" containerName="registry-server" Dec 07 16:52:58 crc kubenswrapper[4716]: I1207 16:52:58.971849 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pmqgv" Dec 07 16:52:58 crc kubenswrapper[4716]: I1207 16:52:58.994334 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f12a967f-1393-4f44-a3ec-36966a4c3445-utilities\") pod \"community-operators-pmqgv\" (UID: \"f12a967f-1393-4f44-a3ec-36966a4c3445\") " pod="openshift-marketplace/community-operators-pmqgv" Dec 07 16:52:58 crc kubenswrapper[4716]: I1207 16:52:58.994456 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f12a967f-1393-4f44-a3ec-36966a4c3445-catalog-content\") pod \"community-operators-pmqgv\" (UID: \"f12a967f-1393-4f44-a3ec-36966a4c3445\") " pod="openshift-marketplace/community-operators-pmqgv" Dec 07 16:52:58 crc kubenswrapper[4716]: I1207 16:52:58.994535 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tz5qp\" (UniqueName: \"kubernetes.io/projected/f12a967f-1393-4f44-a3ec-36966a4c3445-kube-api-access-tz5qp\") pod \"community-operators-pmqgv\" (UID: \"f12a967f-1393-4f44-a3ec-36966a4c3445\") " pod="openshift-marketplace/community-operators-pmqgv" Dec 07 16:52:59 crc kubenswrapper[4716]: I1207 16:52:59.022765 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pmqgv"] Dec 07 16:52:59 crc kubenswrapper[4716]: I1207 16:52:59.096506 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f12a967f-1393-4f44-a3ec-36966a4c3445-catalog-content\") pod \"community-operators-pmqgv\" (UID: \"f12a967f-1393-4f44-a3ec-36966a4c3445\") " pod="openshift-marketplace/community-operators-pmqgv" Dec 07 16:52:59 crc kubenswrapper[4716]: I1207 16:52:59.096615 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tz5qp\" (UniqueName: \"kubernetes.io/projected/f12a967f-1393-4f44-a3ec-36966a4c3445-kube-api-access-tz5qp\") pod \"community-operators-pmqgv\" (UID: \"f12a967f-1393-4f44-a3ec-36966a4c3445\") " pod="openshift-marketplace/community-operators-pmqgv" Dec 07 16:52:59 crc kubenswrapper[4716]: I1207 16:52:59.097183 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f12a967f-1393-4f44-a3ec-36966a4c3445-catalog-content\") pod \"community-operators-pmqgv\" (UID: \"f12a967f-1393-4f44-a3ec-36966a4c3445\") " pod="openshift-marketplace/community-operators-pmqgv" Dec 07 16:52:59 crc kubenswrapper[4716]: I1207 16:52:59.097192 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f12a967f-1393-4f44-a3ec-36966a4c3445-utilities\") pod \"community-operators-pmqgv\" (UID: \"f12a967f-1393-4f44-a3ec-36966a4c3445\") " pod="openshift-marketplace/community-operators-pmqgv" Dec 07 16:52:59 crc kubenswrapper[4716]: I1207 16:52:59.097510 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f12a967f-1393-4f44-a3ec-36966a4c3445-utilities\") pod \"community-operators-pmqgv\" (UID: \"f12a967f-1393-4f44-a3ec-36966a4c3445\") " pod="openshift-marketplace/community-operators-pmqgv" Dec 07 16:52:59 crc kubenswrapper[4716]: I1207 16:52:59.114942 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tz5qp\" (UniqueName: \"kubernetes.io/projected/f12a967f-1393-4f44-a3ec-36966a4c3445-kube-api-access-tz5qp\") pod \"community-operators-pmqgv\" (UID: \"f12a967f-1393-4f44-a3ec-36966a4c3445\") " pod="openshift-marketplace/community-operators-pmqgv" Dec 07 16:52:59 crc kubenswrapper[4716]: I1207 16:52:59.311822 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pmqgv" Dec 07 16:52:59 crc kubenswrapper[4716]: I1207 16:52:59.907043 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pmqgv"] Dec 07 16:53:00 crc kubenswrapper[4716]: I1207 16:53:00.192664 4716 generic.go:334] "Generic (PLEG): container finished" podID="f12a967f-1393-4f44-a3ec-36966a4c3445" containerID="24ede22a27147b6e7daf00decbe0fe4e897495d228bee579877831d54416a17a" exitCode=0 Dec 07 16:53:00 crc kubenswrapper[4716]: I1207 16:53:00.192892 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pmqgv" event={"ID":"f12a967f-1393-4f44-a3ec-36966a4c3445","Type":"ContainerDied","Data":"24ede22a27147b6e7daf00decbe0fe4e897495d228bee579877831d54416a17a"} Dec 07 16:53:00 crc kubenswrapper[4716]: I1207 16:53:00.193051 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pmqgv" event={"ID":"f12a967f-1393-4f44-a3ec-36966a4c3445","Type":"ContainerStarted","Data":"b14ef2a00f8d3f3d11efba66cf244d4d5c2ed63fca5813a87cdcfa779864b749"} Dec 07 16:53:00 crc kubenswrapper[4716]: I1207 16:53:00.195794 4716 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 07 16:53:01 crc kubenswrapper[4716]: I1207 16:53:01.213668 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pmqgv" event={"ID":"f12a967f-1393-4f44-a3ec-36966a4c3445","Type":"ContainerStarted","Data":"5f63f946f46892c7a6284d8b21a41950d05ce353632163ca69a7c371ecc4969d"} Dec 07 16:53:02 crc kubenswrapper[4716]: I1207 16:53:02.223841 4716 generic.go:334] "Generic (PLEG): container finished" podID="f12a967f-1393-4f44-a3ec-36966a4c3445" containerID="5f63f946f46892c7a6284d8b21a41950d05ce353632163ca69a7c371ecc4969d" exitCode=0 Dec 07 16:53:02 crc kubenswrapper[4716]: I1207 16:53:02.224388 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pmqgv" event={"ID":"f12a967f-1393-4f44-a3ec-36966a4c3445","Type":"ContainerDied","Data":"5f63f946f46892c7a6284d8b21a41950d05ce353632163ca69a7c371ecc4969d"} Dec 07 16:53:03 crc kubenswrapper[4716]: I1207 16:53:03.234447 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pmqgv" event={"ID":"f12a967f-1393-4f44-a3ec-36966a4c3445","Type":"ContainerStarted","Data":"4b15d0597881154ba897adc3a1cdd24138ef2c72b6efb80a638cf86c6c68d955"} Dec 07 16:53:03 crc kubenswrapper[4716]: I1207 16:53:03.258582 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-pmqgv" podStartSLOduration=2.837585367 podStartE2EDuration="5.258558747s" podCreationTimestamp="2025-12-07 16:52:58 +0000 UTC" firstStartedPulling="2025-12-07 16:53:00.195533445 +0000 UTC m=+3042.885818357" lastFinishedPulling="2025-12-07 16:53:02.616506805 +0000 UTC m=+3045.306791737" observedRunningTime="2025-12-07 16:53:03.253239062 +0000 UTC m=+3045.943524014" watchObservedRunningTime="2025-12-07 16:53:03.258558747 +0000 UTC m=+3045.948843669" Dec 07 16:53:09 crc kubenswrapper[4716]: I1207 16:53:09.312751 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-pmqgv" Dec 07 16:53:09 crc kubenswrapper[4716]: I1207 16:53:09.313422 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-pmqgv" Dec 07 16:53:09 crc kubenswrapper[4716]: I1207 16:53:09.373850 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-pmqgv" Dec 07 16:53:10 crc kubenswrapper[4716]: I1207 16:53:10.397583 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-pmqgv" Dec 07 16:53:10 crc kubenswrapper[4716]: I1207 16:53:10.476944 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pmqgv"] Dec 07 16:53:12 crc kubenswrapper[4716]: I1207 16:53:12.348222 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-pmqgv" podUID="f12a967f-1393-4f44-a3ec-36966a4c3445" containerName="registry-server" containerID="cri-o://4b15d0597881154ba897adc3a1cdd24138ef2c72b6efb80a638cf86c6c68d955" gracePeriod=2 Dec 07 16:53:12 crc kubenswrapper[4716]: I1207 16:53:12.969544 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pmqgv" Dec 07 16:53:13 crc kubenswrapper[4716]: I1207 16:53:13.076628 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f12a967f-1393-4f44-a3ec-36966a4c3445-catalog-content\") pod \"f12a967f-1393-4f44-a3ec-36966a4c3445\" (UID: \"f12a967f-1393-4f44-a3ec-36966a4c3445\") " Dec 07 16:53:13 crc kubenswrapper[4716]: I1207 16:53:13.076774 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f12a967f-1393-4f44-a3ec-36966a4c3445-utilities\") pod \"f12a967f-1393-4f44-a3ec-36966a4c3445\" (UID: \"f12a967f-1393-4f44-a3ec-36966a4c3445\") " Dec 07 16:53:13 crc kubenswrapper[4716]: I1207 16:53:13.076923 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tz5qp\" (UniqueName: \"kubernetes.io/projected/f12a967f-1393-4f44-a3ec-36966a4c3445-kube-api-access-tz5qp\") pod \"f12a967f-1393-4f44-a3ec-36966a4c3445\" (UID: \"f12a967f-1393-4f44-a3ec-36966a4c3445\") " Dec 07 16:53:13 crc kubenswrapper[4716]: I1207 16:53:13.077651 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f12a967f-1393-4f44-a3ec-36966a4c3445-utilities" (OuterVolumeSpecName: "utilities") pod "f12a967f-1393-4f44-a3ec-36966a4c3445" (UID: "f12a967f-1393-4f44-a3ec-36966a4c3445"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:53:13 crc kubenswrapper[4716]: I1207 16:53:13.082244 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f12a967f-1393-4f44-a3ec-36966a4c3445-kube-api-access-tz5qp" (OuterVolumeSpecName: "kube-api-access-tz5qp") pod "f12a967f-1393-4f44-a3ec-36966a4c3445" (UID: "f12a967f-1393-4f44-a3ec-36966a4c3445"). InnerVolumeSpecName "kube-api-access-tz5qp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:53:13 crc kubenswrapper[4716]: I1207 16:53:13.148351 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f12a967f-1393-4f44-a3ec-36966a4c3445-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f12a967f-1393-4f44-a3ec-36966a4c3445" (UID: "f12a967f-1393-4f44-a3ec-36966a4c3445"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:53:13 crc kubenswrapper[4716]: I1207 16:53:13.179442 4716 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f12a967f-1393-4f44-a3ec-36966a4c3445-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 07 16:53:13 crc kubenswrapper[4716]: I1207 16:53:13.179472 4716 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f12a967f-1393-4f44-a3ec-36966a4c3445-utilities\") on node \"crc\" DevicePath \"\"" Dec 07 16:53:13 crc kubenswrapper[4716]: I1207 16:53:13.179481 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tz5qp\" (UniqueName: \"kubernetes.io/projected/f12a967f-1393-4f44-a3ec-36966a4c3445-kube-api-access-tz5qp\") on node \"crc\" DevicePath \"\"" Dec 07 16:53:13 crc kubenswrapper[4716]: I1207 16:53:13.365820 4716 generic.go:334] "Generic (PLEG): container finished" podID="f12a967f-1393-4f44-a3ec-36966a4c3445" containerID="4b15d0597881154ba897adc3a1cdd24138ef2c72b6efb80a638cf86c6c68d955" exitCode=0 Dec 07 16:53:13 crc kubenswrapper[4716]: I1207 16:53:13.365894 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pmqgv" event={"ID":"f12a967f-1393-4f44-a3ec-36966a4c3445","Type":"ContainerDied","Data":"4b15d0597881154ba897adc3a1cdd24138ef2c72b6efb80a638cf86c6c68d955"} Dec 07 16:53:13 crc kubenswrapper[4716]: I1207 16:53:13.365928 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pmqgv" Dec 07 16:53:13 crc kubenswrapper[4716]: I1207 16:53:13.365963 4716 scope.go:117] "RemoveContainer" containerID="4b15d0597881154ba897adc3a1cdd24138ef2c72b6efb80a638cf86c6c68d955" Dec 07 16:53:13 crc kubenswrapper[4716]: I1207 16:53:13.365940 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pmqgv" event={"ID":"f12a967f-1393-4f44-a3ec-36966a4c3445","Type":"ContainerDied","Data":"b14ef2a00f8d3f3d11efba66cf244d4d5c2ed63fca5813a87cdcfa779864b749"} Dec 07 16:53:13 crc kubenswrapper[4716]: I1207 16:53:13.430212 4716 scope.go:117] "RemoveContainer" containerID="5f63f946f46892c7a6284d8b21a41950d05ce353632163ca69a7c371ecc4969d" Dec 07 16:53:13 crc kubenswrapper[4716]: I1207 16:53:13.443529 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pmqgv"] Dec 07 16:53:13 crc kubenswrapper[4716]: I1207 16:53:13.454267 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-pmqgv"] Dec 07 16:53:13 crc kubenswrapper[4716]: I1207 16:53:13.459683 4716 scope.go:117] "RemoveContainer" containerID="24ede22a27147b6e7daf00decbe0fe4e897495d228bee579877831d54416a17a" Dec 07 16:53:13 crc kubenswrapper[4716]: I1207 16:53:13.515651 4716 scope.go:117] "RemoveContainer" containerID="4b15d0597881154ba897adc3a1cdd24138ef2c72b6efb80a638cf86c6c68d955" Dec 07 16:53:13 crc kubenswrapper[4716]: E1207 16:53:13.516338 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4b15d0597881154ba897adc3a1cdd24138ef2c72b6efb80a638cf86c6c68d955\": container with ID starting with 4b15d0597881154ba897adc3a1cdd24138ef2c72b6efb80a638cf86c6c68d955 not found: ID does not exist" containerID="4b15d0597881154ba897adc3a1cdd24138ef2c72b6efb80a638cf86c6c68d955" Dec 07 16:53:13 crc kubenswrapper[4716]: I1207 16:53:13.516388 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b15d0597881154ba897adc3a1cdd24138ef2c72b6efb80a638cf86c6c68d955"} err="failed to get container status \"4b15d0597881154ba897adc3a1cdd24138ef2c72b6efb80a638cf86c6c68d955\": rpc error: code = NotFound desc = could not find container \"4b15d0597881154ba897adc3a1cdd24138ef2c72b6efb80a638cf86c6c68d955\": container with ID starting with 4b15d0597881154ba897adc3a1cdd24138ef2c72b6efb80a638cf86c6c68d955 not found: ID does not exist" Dec 07 16:53:13 crc kubenswrapper[4716]: I1207 16:53:13.516425 4716 scope.go:117] "RemoveContainer" containerID="5f63f946f46892c7a6284d8b21a41950d05ce353632163ca69a7c371ecc4969d" Dec 07 16:53:13 crc kubenswrapper[4716]: E1207 16:53:13.517111 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f63f946f46892c7a6284d8b21a41950d05ce353632163ca69a7c371ecc4969d\": container with ID starting with 5f63f946f46892c7a6284d8b21a41950d05ce353632163ca69a7c371ecc4969d not found: ID does not exist" containerID="5f63f946f46892c7a6284d8b21a41950d05ce353632163ca69a7c371ecc4969d" Dec 07 16:53:13 crc kubenswrapper[4716]: I1207 16:53:13.517135 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f63f946f46892c7a6284d8b21a41950d05ce353632163ca69a7c371ecc4969d"} err="failed to get container status \"5f63f946f46892c7a6284d8b21a41950d05ce353632163ca69a7c371ecc4969d\": rpc error: code = NotFound desc = could not find container \"5f63f946f46892c7a6284d8b21a41950d05ce353632163ca69a7c371ecc4969d\": container with ID starting with 5f63f946f46892c7a6284d8b21a41950d05ce353632163ca69a7c371ecc4969d not found: ID does not exist" Dec 07 16:53:13 crc kubenswrapper[4716]: I1207 16:53:13.517152 4716 scope.go:117] "RemoveContainer" containerID="24ede22a27147b6e7daf00decbe0fe4e897495d228bee579877831d54416a17a" Dec 07 16:53:13 crc kubenswrapper[4716]: E1207 16:53:13.517472 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"24ede22a27147b6e7daf00decbe0fe4e897495d228bee579877831d54416a17a\": container with ID starting with 24ede22a27147b6e7daf00decbe0fe4e897495d228bee579877831d54416a17a not found: ID does not exist" containerID="24ede22a27147b6e7daf00decbe0fe4e897495d228bee579877831d54416a17a" Dec 07 16:53:13 crc kubenswrapper[4716]: I1207 16:53:13.517521 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24ede22a27147b6e7daf00decbe0fe4e897495d228bee579877831d54416a17a"} err="failed to get container status \"24ede22a27147b6e7daf00decbe0fe4e897495d228bee579877831d54416a17a\": rpc error: code = NotFound desc = could not find container \"24ede22a27147b6e7daf00decbe0fe4e897495d228bee579877831d54416a17a\": container with ID starting with 24ede22a27147b6e7daf00decbe0fe4e897495d228bee579877831d54416a17a not found: ID does not exist" Dec 07 16:53:13 crc kubenswrapper[4716]: I1207 16:53:13.669931 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f12a967f-1393-4f44-a3ec-36966a4c3445" path="/var/lib/kubelet/pods/f12a967f-1393-4f44-a3ec-36966a4c3445/volumes" Dec 07 16:53:22 crc kubenswrapper[4716]: I1207 16:53:22.761778 4716 patch_prober.go:28] interesting pod/machine-config-daemon-zcxxp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 16:53:22 crc kubenswrapper[4716]: I1207 16:53:22.762580 4716 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 16:53:52 crc kubenswrapper[4716]: I1207 16:53:52.761848 4716 patch_prober.go:28] interesting pod/machine-config-daemon-zcxxp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 16:53:52 crc kubenswrapper[4716]: I1207 16:53:52.762467 4716 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 16:54:22 crc kubenswrapper[4716]: I1207 16:54:22.760997 4716 patch_prober.go:28] interesting pod/machine-config-daemon-zcxxp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 16:54:22 crc kubenswrapper[4716]: I1207 16:54:22.761505 4716 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 16:54:22 crc kubenswrapper[4716]: I1207 16:54:22.761577 4716 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" Dec 07 16:54:22 crc kubenswrapper[4716]: I1207 16:54:22.762220 4716 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"320114e03a98e202a688ac2d6205631a9f8d9952773c6722837647cd3ea2b74d"} pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 07 16:54:22 crc kubenswrapper[4716]: I1207 16:54:22.762275 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" containerName="machine-config-daemon" containerID="cri-o://320114e03a98e202a688ac2d6205631a9f8d9952773c6722837647cd3ea2b74d" gracePeriod=600 Dec 07 16:54:22 crc kubenswrapper[4716]: E1207 16:54:22.881024 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 16:54:23 crc kubenswrapper[4716]: I1207 16:54:23.068531 4716 generic.go:334] "Generic (PLEG): container finished" podID="c15b59eb-565d-4556-a4ce-75afdf159dc8" containerID="320114e03a98e202a688ac2d6205631a9f8d9952773c6722837647cd3ea2b74d" exitCode=0 Dec 07 16:54:23 crc kubenswrapper[4716]: I1207 16:54:23.068592 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" event={"ID":"c15b59eb-565d-4556-a4ce-75afdf159dc8","Type":"ContainerDied","Data":"320114e03a98e202a688ac2d6205631a9f8d9952773c6722837647cd3ea2b74d"} Dec 07 16:54:23 crc kubenswrapper[4716]: I1207 16:54:23.068677 4716 scope.go:117] "RemoveContainer" containerID="be6a29d905d404203a7ce5b0088583826cc3981f161994b9094c7227407ce5aa" Dec 07 16:54:23 crc kubenswrapper[4716]: I1207 16:54:23.069484 4716 scope.go:117] "RemoveContainer" containerID="320114e03a98e202a688ac2d6205631a9f8d9952773c6722837647cd3ea2b74d" Dec 07 16:54:23 crc kubenswrapper[4716]: E1207 16:54:23.069991 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 16:54:35 crc kubenswrapper[4716]: I1207 16:54:35.657877 4716 scope.go:117] "RemoveContainer" containerID="320114e03a98e202a688ac2d6205631a9f8d9952773c6722837647cd3ea2b74d" Dec 07 16:54:35 crc kubenswrapper[4716]: E1207 16:54:35.658940 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 16:54:49 crc kubenswrapper[4716]: I1207 16:54:49.659432 4716 scope.go:117] "RemoveContainer" containerID="320114e03a98e202a688ac2d6205631a9f8d9952773c6722837647cd3ea2b74d" Dec 07 16:54:49 crc kubenswrapper[4716]: E1207 16:54:49.660725 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 16:55:02 crc kubenswrapper[4716]: I1207 16:55:02.657148 4716 scope.go:117] "RemoveContainer" containerID="320114e03a98e202a688ac2d6205631a9f8d9952773c6722837647cd3ea2b74d" Dec 07 16:55:02 crc kubenswrapper[4716]: E1207 16:55:02.657958 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 16:55:16 crc kubenswrapper[4716]: I1207 16:55:16.657529 4716 scope.go:117] "RemoveContainer" containerID="320114e03a98e202a688ac2d6205631a9f8d9952773c6722837647cd3ea2b74d" Dec 07 16:55:16 crc kubenswrapper[4716]: E1207 16:55:16.658142 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 16:55:28 crc kubenswrapper[4716]: I1207 16:55:28.658758 4716 scope.go:117] "RemoveContainer" containerID="320114e03a98e202a688ac2d6205631a9f8d9952773c6722837647cd3ea2b74d" Dec 07 16:55:28 crc kubenswrapper[4716]: E1207 16:55:28.665858 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 16:55:43 crc kubenswrapper[4716]: I1207 16:55:43.672652 4716 scope.go:117] "RemoveContainer" containerID="320114e03a98e202a688ac2d6205631a9f8d9952773c6722837647cd3ea2b74d" Dec 07 16:55:43 crc kubenswrapper[4716]: E1207 16:55:43.677538 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 16:55:58 crc kubenswrapper[4716]: I1207 16:55:58.658119 4716 scope.go:117] "RemoveContainer" containerID="320114e03a98e202a688ac2d6205631a9f8d9952773c6722837647cd3ea2b74d" Dec 07 16:55:58 crc kubenswrapper[4716]: E1207 16:55:58.659186 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 16:56:09 crc kubenswrapper[4716]: I1207 16:56:09.657959 4716 scope.go:117] "RemoveContainer" containerID="320114e03a98e202a688ac2d6205631a9f8d9952773c6722837647cd3ea2b74d" Dec 07 16:56:09 crc kubenswrapper[4716]: E1207 16:56:09.658817 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 16:56:21 crc kubenswrapper[4716]: I1207 16:56:21.658649 4716 scope.go:117] "RemoveContainer" containerID="320114e03a98e202a688ac2d6205631a9f8d9952773c6722837647cd3ea2b74d" Dec 07 16:56:21 crc kubenswrapper[4716]: E1207 16:56:21.659613 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 16:56:35 crc kubenswrapper[4716]: I1207 16:56:35.658339 4716 scope.go:117] "RemoveContainer" containerID="320114e03a98e202a688ac2d6205631a9f8d9952773c6722837647cd3ea2b74d" Dec 07 16:56:35 crc kubenswrapper[4716]: E1207 16:56:35.660780 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 16:56:49 crc kubenswrapper[4716]: I1207 16:56:49.877172 4716 scope.go:117] "RemoveContainer" containerID="320114e03a98e202a688ac2d6205631a9f8d9952773c6722837647cd3ea2b74d" Dec 07 16:56:49 crc kubenswrapper[4716]: E1207 16:56:49.878372 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 16:57:01 crc kubenswrapper[4716]: I1207 16:57:01.658277 4716 scope.go:117] "RemoveContainer" containerID="320114e03a98e202a688ac2d6205631a9f8d9952773c6722837647cd3ea2b74d" Dec 07 16:57:01 crc kubenswrapper[4716]: E1207 16:57:01.658941 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 16:57:13 crc kubenswrapper[4716]: I1207 16:57:13.658186 4716 scope.go:117] "RemoveContainer" containerID="320114e03a98e202a688ac2d6205631a9f8d9952773c6722837647cd3ea2b74d" Dec 07 16:57:13 crc kubenswrapper[4716]: E1207 16:57:13.659106 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 16:57:27 crc kubenswrapper[4716]: I1207 16:57:27.665632 4716 scope.go:117] "RemoveContainer" containerID="320114e03a98e202a688ac2d6205631a9f8d9952773c6722837647cd3ea2b74d" Dec 07 16:57:27 crc kubenswrapper[4716]: E1207 16:57:27.666453 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 16:57:41 crc kubenswrapper[4716]: I1207 16:57:41.657611 4716 scope.go:117] "RemoveContainer" containerID="320114e03a98e202a688ac2d6205631a9f8d9952773c6722837647cd3ea2b74d" Dec 07 16:57:41 crc kubenswrapper[4716]: E1207 16:57:41.658448 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 16:57:54 crc kubenswrapper[4716]: I1207 16:57:54.658657 4716 scope.go:117] "RemoveContainer" containerID="320114e03a98e202a688ac2d6205631a9f8d9952773c6722837647cd3ea2b74d" Dec 07 16:57:54 crc kubenswrapper[4716]: E1207 16:57:54.659811 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 16:58:02 crc kubenswrapper[4716]: I1207 16:58:02.709725 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-lxjnx"] Dec 07 16:58:02 crc kubenswrapper[4716]: E1207 16:58:02.710985 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f12a967f-1393-4f44-a3ec-36966a4c3445" containerName="registry-server" Dec 07 16:58:02 crc kubenswrapper[4716]: I1207 16:58:02.711006 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="f12a967f-1393-4f44-a3ec-36966a4c3445" containerName="registry-server" Dec 07 16:58:02 crc kubenswrapper[4716]: E1207 16:58:02.711023 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f12a967f-1393-4f44-a3ec-36966a4c3445" containerName="extract-content" Dec 07 16:58:02 crc kubenswrapper[4716]: I1207 16:58:02.711034 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="f12a967f-1393-4f44-a3ec-36966a4c3445" containerName="extract-content" Dec 07 16:58:02 crc kubenswrapper[4716]: E1207 16:58:02.711066 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f12a967f-1393-4f44-a3ec-36966a4c3445" containerName="extract-utilities" Dec 07 16:58:02 crc kubenswrapper[4716]: I1207 16:58:02.711102 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="f12a967f-1393-4f44-a3ec-36966a4c3445" containerName="extract-utilities" Dec 07 16:58:02 crc kubenswrapper[4716]: I1207 16:58:02.711443 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="f12a967f-1393-4f44-a3ec-36966a4c3445" containerName="registry-server" Dec 07 16:58:02 crc kubenswrapper[4716]: I1207 16:58:02.713549 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lxjnx" Dec 07 16:58:02 crc kubenswrapper[4716]: I1207 16:58:02.736902 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lxjnx"] Dec 07 16:58:02 crc kubenswrapper[4716]: I1207 16:58:02.882449 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89b3c2a2-81ac-4f41-8c50-7d9a8a4d2041-utilities\") pod \"redhat-operators-lxjnx\" (UID: \"89b3c2a2-81ac-4f41-8c50-7d9a8a4d2041\") " pod="openshift-marketplace/redhat-operators-lxjnx" Dec 07 16:58:02 crc kubenswrapper[4716]: I1207 16:58:02.882789 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89b3c2a2-81ac-4f41-8c50-7d9a8a4d2041-catalog-content\") pod \"redhat-operators-lxjnx\" (UID: \"89b3c2a2-81ac-4f41-8c50-7d9a8a4d2041\") " pod="openshift-marketplace/redhat-operators-lxjnx" Dec 07 16:58:02 crc kubenswrapper[4716]: I1207 16:58:02.883330 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghkzh\" (UniqueName: \"kubernetes.io/projected/89b3c2a2-81ac-4f41-8c50-7d9a8a4d2041-kube-api-access-ghkzh\") pod \"redhat-operators-lxjnx\" (UID: \"89b3c2a2-81ac-4f41-8c50-7d9a8a4d2041\") " pod="openshift-marketplace/redhat-operators-lxjnx" Dec 07 16:58:02 crc kubenswrapper[4716]: I1207 16:58:02.984846 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89b3c2a2-81ac-4f41-8c50-7d9a8a4d2041-catalog-content\") pod \"redhat-operators-lxjnx\" (UID: \"89b3c2a2-81ac-4f41-8c50-7d9a8a4d2041\") " pod="openshift-marketplace/redhat-operators-lxjnx" Dec 07 16:58:02 crc kubenswrapper[4716]: I1207 16:58:02.984989 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ghkzh\" (UniqueName: \"kubernetes.io/projected/89b3c2a2-81ac-4f41-8c50-7d9a8a4d2041-kube-api-access-ghkzh\") pod \"redhat-operators-lxjnx\" (UID: \"89b3c2a2-81ac-4f41-8c50-7d9a8a4d2041\") " pod="openshift-marketplace/redhat-operators-lxjnx" Dec 07 16:58:02 crc kubenswrapper[4716]: I1207 16:58:02.985019 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89b3c2a2-81ac-4f41-8c50-7d9a8a4d2041-utilities\") pod \"redhat-operators-lxjnx\" (UID: \"89b3c2a2-81ac-4f41-8c50-7d9a8a4d2041\") " pod="openshift-marketplace/redhat-operators-lxjnx" Dec 07 16:58:02 crc kubenswrapper[4716]: I1207 16:58:02.985321 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89b3c2a2-81ac-4f41-8c50-7d9a8a4d2041-catalog-content\") pod \"redhat-operators-lxjnx\" (UID: \"89b3c2a2-81ac-4f41-8c50-7d9a8a4d2041\") " pod="openshift-marketplace/redhat-operators-lxjnx" Dec 07 16:58:02 crc kubenswrapper[4716]: I1207 16:58:02.985408 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89b3c2a2-81ac-4f41-8c50-7d9a8a4d2041-utilities\") pod \"redhat-operators-lxjnx\" (UID: \"89b3c2a2-81ac-4f41-8c50-7d9a8a4d2041\") " pod="openshift-marketplace/redhat-operators-lxjnx" Dec 07 16:58:03 crc kubenswrapper[4716]: I1207 16:58:03.003796 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ghkzh\" (UniqueName: \"kubernetes.io/projected/89b3c2a2-81ac-4f41-8c50-7d9a8a4d2041-kube-api-access-ghkzh\") pod \"redhat-operators-lxjnx\" (UID: \"89b3c2a2-81ac-4f41-8c50-7d9a8a4d2041\") " pod="openshift-marketplace/redhat-operators-lxjnx" Dec 07 16:58:03 crc kubenswrapper[4716]: I1207 16:58:03.033966 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lxjnx" Dec 07 16:58:03 crc kubenswrapper[4716]: I1207 16:58:03.513728 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lxjnx"] Dec 07 16:58:04 crc kubenswrapper[4716]: I1207 16:58:04.531958 4716 generic.go:334] "Generic (PLEG): container finished" podID="89b3c2a2-81ac-4f41-8c50-7d9a8a4d2041" containerID="69a4c38acf2d20e75f8b080764c05560b6fcf6ce679ecf199d3217723d313ed8" exitCode=0 Dec 07 16:58:04 crc kubenswrapper[4716]: I1207 16:58:04.532005 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lxjnx" event={"ID":"89b3c2a2-81ac-4f41-8c50-7d9a8a4d2041","Type":"ContainerDied","Data":"69a4c38acf2d20e75f8b080764c05560b6fcf6ce679ecf199d3217723d313ed8"} Dec 07 16:58:04 crc kubenswrapper[4716]: I1207 16:58:04.532258 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lxjnx" event={"ID":"89b3c2a2-81ac-4f41-8c50-7d9a8a4d2041","Type":"ContainerStarted","Data":"03082e01c7715c0744f5883fcd1c1a3c4870f48e5c61840a120fcfa975f3faed"} Dec 07 16:58:04 crc kubenswrapper[4716]: I1207 16:58:04.533992 4716 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 07 16:58:05 crc kubenswrapper[4716]: I1207 16:58:05.545788 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lxjnx" event={"ID":"89b3c2a2-81ac-4f41-8c50-7d9a8a4d2041","Type":"ContainerStarted","Data":"ff1d1f2b0dd0b1754b302dcea15629caff396cda8a06d4d2c1d082dbc1cfd9ad"} Dec 07 16:58:06 crc kubenswrapper[4716]: I1207 16:58:06.556632 4716 generic.go:334] "Generic (PLEG): container finished" podID="89b3c2a2-81ac-4f41-8c50-7d9a8a4d2041" containerID="ff1d1f2b0dd0b1754b302dcea15629caff396cda8a06d4d2c1d082dbc1cfd9ad" exitCode=0 Dec 07 16:58:06 crc kubenswrapper[4716]: I1207 16:58:06.556733 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lxjnx" event={"ID":"89b3c2a2-81ac-4f41-8c50-7d9a8a4d2041","Type":"ContainerDied","Data":"ff1d1f2b0dd0b1754b302dcea15629caff396cda8a06d4d2c1d082dbc1cfd9ad"} Dec 07 16:58:06 crc kubenswrapper[4716]: I1207 16:58:06.658102 4716 scope.go:117] "RemoveContainer" containerID="320114e03a98e202a688ac2d6205631a9f8d9952773c6722837647cd3ea2b74d" Dec 07 16:58:06 crc kubenswrapper[4716]: E1207 16:58:06.658351 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 16:58:07 crc kubenswrapper[4716]: I1207 16:58:07.574710 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lxjnx" event={"ID":"89b3c2a2-81ac-4f41-8c50-7d9a8a4d2041","Type":"ContainerStarted","Data":"d7a25105a96b38bddb83a1d2e6e2e9aa49701e1ba7c80838e215b31171162100"} Dec 07 16:58:07 crc kubenswrapper[4716]: I1207 16:58:07.594922 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-lxjnx" podStartSLOduration=3.048007668 podStartE2EDuration="5.594907433s" podCreationTimestamp="2025-12-07 16:58:02 +0000 UTC" firstStartedPulling="2025-12-07 16:58:04.533790772 +0000 UTC m=+3347.224075684" lastFinishedPulling="2025-12-07 16:58:07.080690537 +0000 UTC m=+3349.770975449" observedRunningTime="2025-12-07 16:58:07.593293889 +0000 UTC m=+3350.283578811" watchObservedRunningTime="2025-12-07 16:58:07.594907433 +0000 UTC m=+3350.285192345" Dec 07 16:58:13 crc kubenswrapper[4716]: I1207 16:58:13.035068 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-lxjnx" Dec 07 16:58:13 crc kubenswrapper[4716]: I1207 16:58:13.035622 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-lxjnx" Dec 07 16:58:13 crc kubenswrapper[4716]: I1207 16:58:13.078262 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-lxjnx" Dec 07 16:58:13 crc kubenswrapper[4716]: I1207 16:58:13.730476 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-lxjnx" Dec 07 16:58:13 crc kubenswrapper[4716]: I1207 16:58:13.794613 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lxjnx"] Dec 07 16:58:15 crc kubenswrapper[4716]: I1207 16:58:15.689130 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-lxjnx" podUID="89b3c2a2-81ac-4f41-8c50-7d9a8a4d2041" containerName="registry-server" containerID="cri-o://d7a25105a96b38bddb83a1d2e6e2e9aa49701e1ba7c80838e215b31171162100" gracePeriod=2 Dec 07 16:58:16 crc kubenswrapper[4716]: I1207 16:58:16.260660 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lxjnx" Dec 07 16:58:16 crc kubenswrapper[4716]: I1207 16:58:16.379377 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89b3c2a2-81ac-4f41-8c50-7d9a8a4d2041-utilities\") pod \"89b3c2a2-81ac-4f41-8c50-7d9a8a4d2041\" (UID: \"89b3c2a2-81ac-4f41-8c50-7d9a8a4d2041\") " Dec 07 16:58:16 crc kubenswrapper[4716]: I1207 16:58:16.379443 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89b3c2a2-81ac-4f41-8c50-7d9a8a4d2041-catalog-content\") pod \"89b3c2a2-81ac-4f41-8c50-7d9a8a4d2041\" (UID: \"89b3c2a2-81ac-4f41-8c50-7d9a8a4d2041\") " Dec 07 16:58:16 crc kubenswrapper[4716]: I1207 16:58:16.379601 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ghkzh\" (UniqueName: \"kubernetes.io/projected/89b3c2a2-81ac-4f41-8c50-7d9a8a4d2041-kube-api-access-ghkzh\") pod \"89b3c2a2-81ac-4f41-8c50-7d9a8a4d2041\" (UID: \"89b3c2a2-81ac-4f41-8c50-7d9a8a4d2041\") " Dec 07 16:58:16 crc kubenswrapper[4716]: I1207 16:58:16.381188 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/89b3c2a2-81ac-4f41-8c50-7d9a8a4d2041-utilities" (OuterVolumeSpecName: "utilities") pod "89b3c2a2-81ac-4f41-8c50-7d9a8a4d2041" (UID: "89b3c2a2-81ac-4f41-8c50-7d9a8a4d2041"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:58:16 crc kubenswrapper[4716]: I1207 16:58:16.387456 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89b3c2a2-81ac-4f41-8c50-7d9a8a4d2041-kube-api-access-ghkzh" (OuterVolumeSpecName: "kube-api-access-ghkzh") pod "89b3c2a2-81ac-4f41-8c50-7d9a8a4d2041" (UID: "89b3c2a2-81ac-4f41-8c50-7d9a8a4d2041"). InnerVolumeSpecName "kube-api-access-ghkzh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:58:16 crc kubenswrapper[4716]: I1207 16:58:16.481694 4716 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89b3c2a2-81ac-4f41-8c50-7d9a8a4d2041-utilities\") on node \"crc\" DevicePath \"\"" Dec 07 16:58:16 crc kubenswrapper[4716]: I1207 16:58:16.481724 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ghkzh\" (UniqueName: \"kubernetes.io/projected/89b3c2a2-81ac-4f41-8c50-7d9a8a4d2041-kube-api-access-ghkzh\") on node \"crc\" DevicePath \"\"" Dec 07 16:58:16 crc kubenswrapper[4716]: I1207 16:58:16.492685 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/89b3c2a2-81ac-4f41-8c50-7d9a8a4d2041-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "89b3c2a2-81ac-4f41-8c50-7d9a8a4d2041" (UID: "89b3c2a2-81ac-4f41-8c50-7d9a8a4d2041"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:58:16 crc kubenswrapper[4716]: I1207 16:58:16.582937 4716 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89b3c2a2-81ac-4f41-8c50-7d9a8a4d2041-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 07 16:58:16 crc kubenswrapper[4716]: I1207 16:58:16.698883 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lxjnx" Dec 07 16:58:16 crc kubenswrapper[4716]: I1207 16:58:16.698915 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lxjnx" event={"ID":"89b3c2a2-81ac-4f41-8c50-7d9a8a4d2041","Type":"ContainerDied","Data":"d7a25105a96b38bddb83a1d2e6e2e9aa49701e1ba7c80838e215b31171162100"} Dec 07 16:58:16 crc kubenswrapper[4716]: I1207 16:58:16.698867 4716 generic.go:334] "Generic (PLEG): container finished" podID="89b3c2a2-81ac-4f41-8c50-7d9a8a4d2041" containerID="d7a25105a96b38bddb83a1d2e6e2e9aa49701e1ba7c80838e215b31171162100" exitCode=0 Dec 07 16:58:16 crc kubenswrapper[4716]: I1207 16:58:16.698983 4716 scope.go:117] "RemoveContainer" containerID="d7a25105a96b38bddb83a1d2e6e2e9aa49701e1ba7c80838e215b31171162100" Dec 07 16:58:16 crc kubenswrapper[4716]: I1207 16:58:16.698997 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lxjnx" event={"ID":"89b3c2a2-81ac-4f41-8c50-7d9a8a4d2041","Type":"ContainerDied","Data":"03082e01c7715c0744f5883fcd1c1a3c4870f48e5c61840a120fcfa975f3faed"} Dec 07 16:58:16 crc kubenswrapper[4716]: I1207 16:58:16.720508 4716 scope.go:117] "RemoveContainer" containerID="ff1d1f2b0dd0b1754b302dcea15629caff396cda8a06d4d2c1d082dbc1cfd9ad" Dec 07 16:58:16 crc kubenswrapper[4716]: I1207 16:58:16.738790 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lxjnx"] Dec 07 16:58:16 crc kubenswrapper[4716]: I1207 16:58:16.753413 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-lxjnx"] Dec 07 16:58:16 crc kubenswrapper[4716]: I1207 16:58:16.760421 4716 scope.go:117] "RemoveContainer" containerID="69a4c38acf2d20e75f8b080764c05560b6fcf6ce679ecf199d3217723d313ed8" Dec 07 16:58:16 crc kubenswrapper[4716]: I1207 16:58:16.795724 4716 scope.go:117] "RemoveContainer" containerID="d7a25105a96b38bddb83a1d2e6e2e9aa49701e1ba7c80838e215b31171162100" Dec 07 16:58:16 crc kubenswrapper[4716]: E1207 16:58:16.796816 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7a25105a96b38bddb83a1d2e6e2e9aa49701e1ba7c80838e215b31171162100\": container with ID starting with d7a25105a96b38bddb83a1d2e6e2e9aa49701e1ba7c80838e215b31171162100 not found: ID does not exist" containerID="d7a25105a96b38bddb83a1d2e6e2e9aa49701e1ba7c80838e215b31171162100" Dec 07 16:58:16 crc kubenswrapper[4716]: I1207 16:58:16.796866 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7a25105a96b38bddb83a1d2e6e2e9aa49701e1ba7c80838e215b31171162100"} err="failed to get container status \"d7a25105a96b38bddb83a1d2e6e2e9aa49701e1ba7c80838e215b31171162100\": rpc error: code = NotFound desc = could not find container \"d7a25105a96b38bddb83a1d2e6e2e9aa49701e1ba7c80838e215b31171162100\": container with ID starting with d7a25105a96b38bddb83a1d2e6e2e9aa49701e1ba7c80838e215b31171162100 not found: ID does not exist" Dec 07 16:58:16 crc kubenswrapper[4716]: I1207 16:58:16.796900 4716 scope.go:117] "RemoveContainer" containerID="ff1d1f2b0dd0b1754b302dcea15629caff396cda8a06d4d2c1d082dbc1cfd9ad" Dec 07 16:58:16 crc kubenswrapper[4716]: E1207 16:58:16.797313 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff1d1f2b0dd0b1754b302dcea15629caff396cda8a06d4d2c1d082dbc1cfd9ad\": container with ID starting with ff1d1f2b0dd0b1754b302dcea15629caff396cda8a06d4d2c1d082dbc1cfd9ad not found: ID does not exist" containerID="ff1d1f2b0dd0b1754b302dcea15629caff396cda8a06d4d2c1d082dbc1cfd9ad" Dec 07 16:58:16 crc kubenswrapper[4716]: I1207 16:58:16.797342 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff1d1f2b0dd0b1754b302dcea15629caff396cda8a06d4d2c1d082dbc1cfd9ad"} err="failed to get container status \"ff1d1f2b0dd0b1754b302dcea15629caff396cda8a06d4d2c1d082dbc1cfd9ad\": rpc error: code = NotFound desc = could not find container \"ff1d1f2b0dd0b1754b302dcea15629caff396cda8a06d4d2c1d082dbc1cfd9ad\": container with ID starting with ff1d1f2b0dd0b1754b302dcea15629caff396cda8a06d4d2c1d082dbc1cfd9ad not found: ID does not exist" Dec 07 16:58:16 crc kubenswrapper[4716]: I1207 16:58:16.797359 4716 scope.go:117] "RemoveContainer" containerID="69a4c38acf2d20e75f8b080764c05560b6fcf6ce679ecf199d3217723d313ed8" Dec 07 16:58:16 crc kubenswrapper[4716]: E1207 16:58:16.798112 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69a4c38acf2d20e75f8b080764c05560b6fcf6ce679ecf199d3217723d313ed8\": container with ID starting with 69a4c38acf2d20e75f8b080764c05560b6fcf6ce679ecf199d3217723d313ed8 not found: ID does not exist" containerID="69a4c38acf2d20e75f8b080764c05560b6fcf6ce679ecf199d3217723d313ed8" Dec 07 16:58:16 crc kubenswrapper[4716]: I1207 16:58:16.798155 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69a4c38acf2d20e75f8b080764c05560b6fcf6ce679ecf199d3217723d313ed8"} err="failed to get container status \"69a4c38acf2d20e75f8b080764c05560b6fcf6ce679ecf199d3217723d313ed8\": rpc error: code = NotFound desc = could not find container \"69a4c38acf2d20e75f8b080764c05560b6fcf6ce679ecf199d3217723d313ed8\": container with ID starting with 69a4c38acf2d20e75f8b080764c05560b6fcf6ce679ecf199d3217723d313ed8 not found: ID does not exist" Dec 07 16:58:17 crc kubenswrapper[4716]: I1207 16:58:17.667506 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="89b3c2a2-81ac-4f41-8c50-7d9a8a4d2041" path="/var/lib/kubelet/pods/89b3c2a2-81ac-4f41-8c50-7d9a8a4d2041/volumes" Dec 07 16:58:18 crc kubenswrapper[4716]: I1207 16:58:18.658312 4716 scope.go:117] "RemoveContainer" containerID="320114e03a98e202a688ac2d6205631a9f8d9952773c6722837647cd3ea2b74d" Dec 07 16:58:18 crc kubenswrapper[4716]: E1207 16:58:18.659226 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 16:58:30 crc kubenswrapper[4716]: I1207 16:58:30.657997 4716 scope.go:117] "RemoveContainer" containerID="320114e03a98e202a688ac2d6205631a9f8d9952773c6722837647cd3ea2b74d" Dec 07 16:58:30 crc kubenswrapper[4716]: E1207 16:58:30.658852 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 16:58:43 crc kubenswrapper[4716]: I1207 16:58:43.276751 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-xkclh"] Dec 07 16:58:43 crc kubenswrapper[4716]: E1207 16:58:43.277698 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89b3c2a2-81ac-4f41-8c50-7d9a8a4d2041" containerName="extract-utilities" Dec 07 16:58:43 crc kubenswrapper[4716]: I1207 16:58:43.277713 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="89b3c2a2-81ac-4f41-8c50-7d9a8a4d2041" containerName="extract-utilities" Dec 07 16:58:43 crc kubenswrapper[4716]: E1207 16:58:43.277736 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89b3c2a2-81ac-4f41-8c50-7d9a8a4d2041" containerName="extract-content" Dec 07 16:58:43 crc kubenswrapper[4716]: I1207 16:58:43.277742 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="89b3c2a2-81ac-4f41-8c50-7d9a8a4d2041" containerName="extract-content" Dec 07 16:58:43 crc kubenswrapper[4716]: E1207 16:58:43.277760 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89b3c2a2-81ac-4f41-8c50-7d9a8a4d2041" containerName="registry-server" Dec 07 16:58:43 crc kubenswrapper[4716]: I1207 16:58:43.277766 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="89b3c2a2-81ac-4f41-8c50-7d9a8a4d2041" containerName="registry-server" Dec 07 16:58:43 crc kubenswrapper[4716]: I1207 16:58:43.277942 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="89b3c2a2-81ac-4f41-8c50-7d9a8a4d2041" containerName="registry-server" Dec 07 16:58:43 crc kubenswrapper[4716]: I1207 16:58:43.279651 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xkclh" Dec 07 16:58:43 crc kubenswrapper[4716]: I1207 16:58:43.291048 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xkclh"] Dec 07 16:58:43 crc kubenswrapper[4716]: I1207 16:58:43.408424 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2d690dcc-677f-45a1-8ef0-fb1ecb4a3b4f-utilities\") pod \"redhat-marketplace-xkclh\" (UID: \"2d690dcc-677f-45a1-8ef0-fb1ecb4a3b4f\") " pod="openshift-marketplace/redhat-marketplace-xkclh" Dec 07 16:58:43 crc kubenswrapper[4716]: I1207 16:58:43.408497 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2d690dcc-677f-45a1-8ef0-fb1ecb4a3b4f-catalog-content\") pod \"redhat-marketplace-xkclh\" (UID: \"2d690dcc-677f-45a1-8ef0-fb1ecb4a3b4f\") " pod="openshift-marketplace/redhat-marketplace-xkclh" Dec 07 16:58:43 crc kubenswrapper[4716]: I1207 16:58:43.408646 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ftlx9\" (UniqueName: \"kubernetes.io/projected/2d690dcc-677f-45a1-8ef0-fb1ecb4a3b4f-kube-api-access-ftlx9\") pod \"redhat-marketplace-xkclh\" (UID: \"2d690dcc-677f-45a1-8ef0-fb1ecb4a3b4f\") " pod="openshift-marketplace/redhat-marketplace-xkclh" Dec 07 16:58:43 crc kubenswrapper[4716]: I1207 16:58:43.510327 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ftlx9\" (UniqueName: \"kubernetes.io/projected/2d690dcc-677f-45a1-8ef0-fb1ecb4a3b4f-kube-api-access-ftlx9\") pod \"redhat-marketplace-xkclh\" (UID: \"2d690dcc-677f-45a1-8ef0-fb1ecb4a3b4f\") " pod="openshift-marketplace/redhat-marketplace-xkclh" Dec 07 16:58:43 crc kubenswrapper[4716]: I1207 16:58:43.510422 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2d690dcc-677f-45a1-8ef0-fb1ecb4a3b4f-utilities\") pod \"redhat-marketplace-xkclh\" (UID: \"2d690dcc-677f-45a1-8ef0-fb1ecb4a3b4f\") " pod="openshift-marketplace/redhat-marketplace-xkclh" Dec 07 16:58:43 crc kubenswrapper[4716]: I1207 16:58:43.510446 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2d690dcc-677f-45a1-8ef0-fb1ecb4a3b4f-catalog-content\") pod \"redhat-marketplace-xkclh\" (UID: \"2d690dcc-677f-45a1-8ef0-fb1ecb4a3b4f\") " pod="openshift-marketplace/redhat-marketplace-xkclh" Dec 07 16:58:43 crc kubenswrapper[4716]: I1207 16:58:43.510958 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2d690dcc-677f-45a1-8ef0-fb1ecb4a3b4f-utilities\") pod \"redhat-marketplace-xkclh\" (UID: \"2d690dcc-677f-45a1-8ef0-fb1ecb4a3b4f\") " pod="openshift-marketplace/redhat-marketplace-xkclh" Dec 07 16:58:43 crc kubenswrapper[4716]: I1207 16:58:43.510998 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2d690dcc-677f-45a1-8ef0-fb1ecb4a3b4f-catalog-content\") pod \"redhat-marketplace-xkclh\" (UID: \"2d690dcc-677f-45a1-8ef0-fb1ecb4a3b4f\") " pod="openshift-marketplace/redhat-marketplace-xkclh" Dec 07 16:58:43 crc kubenswrapper[4716]: I1207 16:58:43.535924 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ftlx9\" (UniqueName: \"kubernetes.io/projected/2d690dcc-677f-45a1-8ef0-fb1ecb4a3b4f-kube-api-access-ftlx9\") pod \"redhat-marketplace-xkclh\" (UID: \"2d690dcc-677f-45a1-8ef0-fb1ecb4a3b4f\") " pod="openshift-marketplace/redhat-marketplace-xkclh" Dec 07 16:58:43 crc kubenswrapper[4716]: I1207 16:58:43.626686 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xkclh" Dec 07 16:58:44 crc kubenswrapper[4716]: I1207 16:58:44.162643 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xkclh"] Dec 07 16:58:44 crc kubenswrapper[4716]: I1207 16:58:44.657688 4716 scope.go:117] "RemoveContainer" containerID="320114e03a98e202a688ac2d6205631a9f8d9952773c6722837647cd3ea2b74d" Dec 07 16:58:44 crc kubenswrapper[4716]: E1207 16:58:44.658181 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 16:58:44 crc kubenswrapper[4716]: I1207 16:58:44.988300 4716 generic.go:334] "Generic (PLEG): container finished" podID="2d690dcc-677f-45a1-8ef0-fb1ecb4a3b4f" containerID="af0f23bb18a48382b5165fcec913ec503c7c3fc81455a8d13adbec7f6e5bab06" exitCode=0 Dec 07 16:58:44 crc kubenswrapper[4716]: I1207 16:58:44.988362 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xkclh" event={"ID":"2d690dcc-677f-45a1-8ef0-fb1ecb4a3b4f","Type":"ContainerDied","Data":"af0f23bb18a48382b5165fcec913ec503c7c3fc81455a8d13adbec7f6e5bab06"} Dec 07 16:58:44 crc kubenswrapper[4716]: I1207 16:58:44.988393 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xkclh" event={"ID":"2d690dcc-677f-45a1-8ef0-fb1ecb4a3b4f","Type":"ContainerStarted","Data":"28d9a3d4aed80a6a616b3498633a21f3063cde9c34c3d6df12d737a8c6f31a34"} Dec 07 16:58:46 crc kubenswrapper[4716]: I1207 16:58:45.999610 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xkclh" event={"ID":"2d690dcc-677f-45a1-8ef0-fb1ecb4a3b4f","Type":"ContainerDied","Data":"c69c51a78085d94ab6127abcc025f4b2c26a9e713b8f46f0b4c1bb3b521d6645"} Dec 07 16:58:46 crc kubenswrapper[4716]: I1207 16:58:45.999579 4716 generic.go:334] "Generic (PLEG): container finished" podID="2d690dcc-677f-45a1-8ef0-fb1ecb4a3b4f" containerID="c69c51a78085d94ab6127abcc025f4b2c26a9e713b8f46f0b4c1bb3b521d6645" exitCode=0 Dec 07 16:58:47 crc kubenswrapper[4716]: I1207 16:58:47.010567 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xkclh" event={"ID":"2d690dcc-677f-45a1-8ef0-fb1ecb4a3b4f","Type":"ContainerStarted","Data":"f6fc90968ae0bcf6f94d458d9d0603dc88c9b42580b366018a0c402714ccc817"} Dec 07 16:58:53 crc kubenswrapper[4716]: I1207 16:58:53.627652 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-xkclh" Dec 07 16:58:53 crc kubenswrapper[4716]: I1207 16:58:53.628284 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-xkclh" Dec 07 16:58:53 crc kubenswrapper[4716]: I1207 16:58:53.675998 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-xkclh" Dec 07 16:58:53 crc kubenswrapper[4716]: I1207 16:58:53.708214 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-xkclh" podStartSLOduration=9.287652816 podStartE2EDuration="10.70819295s" podCreationTimestamp="2025-12-07 16:58:43 +0000 UTC" firstStartedPulling="2025-12-07 16:58:44.990341883 +0000 UTC m=+3387.680626795" lastFinishedPulling="2025-12-07 16:58:46.410882027 +0000 UTC m=+3389.101166929" observedRunningTime="2025-12-07 16:58:47.034462271 +0000 UTC m=+3389.724747183" watchObservedRunningTime="2025-12-07 16:58:53.70819295 +0000 UTC m=+3396.398477882" Dec 07 16:58:54 crc kubenswrapper[4716]: I1207 16:58:54.125623 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-xkclh" Dec 07 16:58:54 crc kubenswrapper[4716]: I1207 16:58:54.184673 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xkclh"] Dec 07 16:58:56 crc kubenswrapper[4716]: I1207 16:58:56.103021 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-xkclh" podUID="2d690dcc-677f-45a1-8ef0-fb1ecb4a3b4f" containerName="registry-server" containerID="cri-o://f6fc90968ae0bcf6f94d458d9d0603dc88c9b42580b366018a0c402714ccc817" gracePeriod=2 Dec 07 16:58:56 crc kubenswrapper[4716]: I1207 16:58:56.676025 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xkclh" Dec 07 16:58:56 crc kubenswrapper[4716]: I1207 16:58:56.773177 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2d690dcc-677f-45a1-8ef0-fb1ecb4a3b4f-catalog-content\") pod \"2d690dcc-677f-45a1-8ef0-fb1ecb4a3b4f\" (UID: \"2d690dcc-677f-45a1-8ef0-fb1ecb4a3b4f\") " Dec 07 16:58:56 crc kubenswrapper[4716]: I1207 16:58:56.773228 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2d690dcc-677f-45a1-8ef0-fb1ecb4a3b4f-utilities\") pod \"2d690dcc-677f-45a1-8ef0-fb1ecb4a3b4f\" (UID: \"2d690dcc-677f-45a1-8ef0-fb1ecb4a3b4f\") " Dec 07 16:58:56 crc kubenswrapper[4716]: I1207 16:58:56.773328 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ftlx9\" (UniqueName: \"kubernetes.io/projected/2d690dcc-677f-45a1-8ef0-fb1ecb4a3b4f-kube-api-access-ftlx9\") pod \"2d690dcc-677f-45a1-8ef0-fb1ecb4a3b4f\" (UID: \"2d690dcc-677f-45a1-8ef0-fb1ecb4a3b4f\") " Dec 07 16:58:56 crc kubenswrapper[4716]: I1207 16:58:56.774158 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2d690dcc-677f-45a1-8ef0-fb1ecb4a3b4f-utilities" (OuterVolumeSpecName: "utilities") pod "2d690dcc-677f-45a1-8ef0-fb1ecb4a3b4f" (UID: "2d690dcc-677f-45a1-8ef0-fb1ecb4a3b4f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:58:56 crc kubenswrapper[4716]: I1207 16:58:56.783659 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d690dcc-677f-45a1-8ef0-fb1ecb4a3b4f-kube-api-access-ftlx9" (OuterVolumeSpecName: "kube-api-access-ftlx9") pod "2d690dcc-677f-45a1-8ef0-fb1ecb4a3b4f" (UID: "2d690dcc-677f-45a1-8ef0-fb1ecb4a3b4f"). InnerVolumeSpecName "kube-api-access-ftlx9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:58:56 crc kubenswrapper[4716]: I1207 16:58:56.810830 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2d690dcc-677f-45a1-8ef0-fb1ecb4a3b4f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2d690dcc-677f-45a1-8ef0-fb1ecb4a3b4f" (UID: "2d690dcc-677f-45a1-8ef0-fb1ecb4a3b4f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:58:56 crc kubenswrapper[4716]: I1207 16:58:56.875239 4716 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2d690dcc-677f-45a1-8ef0-fb1ecb4a3b4f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 07 16:58:56 crc kubenswrapper[4716]: I1207 16:58:56.875275 4716 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2d690dcc-677f-45a1-8ef0-fb1ecb4a3b4f-utilities\") on node \"crc\" DevicePath \"\"" Dec 07 16:58:56 crc kubenswrapper[4716]: I1207 16:58:56.875286 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ftlx9\" (UniqueName: \"kubernetes.io/projected/2d690dcc-677f-45a1-8ef0-fb1ecb4a3b4f-kube-api-access-ftlx9\") on node \"crc\" DevicePath \"\"" Dec 07 16:58:57 crc kubenswrapper[4716]: I1207 16:58:57.113190 4716 generic.go:334] "Generic (PLEG): container finished" podID="2d690dcc-677f-45a1-8ef0-fb1ecb4a3b4f" containerID="f6fc90968ae0bcf6f94d458d9d0603dc88c9b42580b366018a0c402714ccc817" exitCode=0 Dec 07 16:58:57 crc kubenswrapper[4716]: I1207 16:58:57.113239 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xkclh" Dec 07 16:58:57 crc kubenswrapper[4716]: I1207 16:58:57.113243 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xkclh" event={"ID":"2d690dcc-677f-45a1-8ef0-fb1ecb4a3b4f","Type":"ContainerDied","Data":"f6fc90968ae0bcf6f94d458d9d0603dc88c9b42580b366018a0c402714ccc817"} Dec 07 16:58:57 crc kubenswrapper[4716]: I1207 16:58:57.113420 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xkclh" event={"ID":"2d690dcc-677f-45a1-8ef0-fb1ecb4a3b4f","Type":"ContainerDied","Data":"28d9a3d4aed80a6a616b3498633a21f3063cde9c34c3d6df12d737a8c6f31a34"} Dec 07 16:58:57 crc kubenswrapper[4716]: I1207 16:58:57.113474 4716 scope.go:117] "RemoveContainer" containerID="f6fc90968ae0bcf6f94d458d9d0603dc88c9b42580b366018a0c402714ccc817" Dec 07 16:58:57 crc kubenswrapper[4716]: I1207 16:58:57.157434 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xkclh"] Dec 07 16:58:57 crc kubenswrapper[4716]: I1207 16:58:57.158648 4716 scope.go:117] "RemoveContainer" containerID="c69c51a78085d94ab6127abcc025f4b2c26a9e713b8f46f0b4c1bb3b521d6645" Dec 07 16:58:57 crc kubenswrapper[4716]: I1207 16:58:57.165986 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-xkclh"] Dec 07 16:58:57 crc kubenswrapper[4716]: I1207 16:58:57.181196 4716 scope.go:117] "RemoveContainer" containerID="af0f23bb18a48382b5165fcec913ec503c7c3fc81455a8d13adbec7f6e5bab06" Dec 07 16:58:57 crc kubenswrapper[4716]: I1207 16:58:57.241222 4716 scope.go:117] "RemoveContainer" containerID="f6fc90968ae0bcf6f94d458d9d0603dc88c9b42580b366018a0c402714ccc817" Dec 07 16:58:57 crc kubenswrapper[4716]: E1207 16:58:57.241777 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f6fc90968ae0bcf6f94d458d9d0603dc88c9b42580b366018a0c402714ccc817\": container with ID starting with f6fc90968ae0bcf6f94d458d9d0603dc88c9b42580b366018a0c402714ccc817 not found: ID does not exist" containerID="f6fc90968ae0bcf6f94d458d9d0603dc88c9b42580b366018a0c402714ccc817" Dec 07 16:58:57 crc kubenswrapper[4716]: I1207 16:58:57.241825 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6fc90968ae0bcf6f94d458d9d0603dc88c9b42580b366018a0c402714ccc817"} err="failed to get container status \"f6fc90968ae0bcf6f94d458d9d0603dc88c9b42580b366018a0c402714ccc817\": rpc error: code = NotFound desc = could not find container \"f6fc90968ae0bcf6f94d458d9d0603dc88c9b42580b366018a0c402714ccc817\": container with ID starting with f6fc90968ae0bcf6f94d458d9d0603dc88c9b42580b366018a0c402714ccc817 not found: ID does not exist" Dec 07 16:58:57 crc kubenswrapper[4716]: I1207 16:58:57.241853 4716 scope.go:117] "RemoveContainer" containerID="c69c51a78085d94ab6127abcc025f4b2c26a9e713b8f46f0b4c1bb3b521d6645" Dec 07 16:58:57 crc kubenswrapper[4716]: E1207 16:58:57.242260 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c69c51a78085d94ab6127abcc025f4b2c26a9e713b8f46f0b4c1bb3b521d6645\": container with ID starting with c69c51a78085d94ab6127abcc025f4b2c26a9e713b8f46f0b4c1bb3b521d6645 not found: ID does not exist" containerID="c69c51a78085d94ab6127abcc025f4b2c26a9e713b8f46f0b4c1bb3b521d6645" Dec 07 16:58:57 crc kubenswrapper[4716]: I1207 16:58:57.242322 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c69c51a78085d94ab6127abcc025f4b2c26a9e713b8f46f0b4c1bb3b521d6645"} err="failed to get container status \"c69c51a78085d94ab6127abcc025f4b2c26a9e713b8f46f0b4c1bb3b521d6645\": rpc error: code = NotFound desc = could not find container \"c69c51a78085d94ab6127abcc025f4b2c26a9e713b8f46f0b4c1bb3b521d6645\": container with ID starting with c69c51a78085d94ab6127abcc025f4b2c26a9e713b8f46f0b4c1bb3b521d6645 not found: ID does not exist" Dec 07 16:58:57 crc kubenswrapper[4716]: I1207 16:58:57.242365 4716 scope.go:117] "RemoveContainer" containerID="af0f23bb18a48382b5165fcec913ec503c7c3fc81455a8d13adbec7f6e5bab06" Dec 07 16:58:57 crc kubenswrapper[4716]: E1207 16:58:57.242704 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af0f23bb18a48382b5165fcec913ec503c7c3fc81455a8d13adbec7f6e5bab06\": container with ID starting with af0f23bb18a48382b5165fcec913ec503c7c3fc81455a8d13adbec7f6e5bab06 not found: ID does not exist" containerID="af0f23bb18a48382b5165fcec913ec503c7c3fc81455a8d13adbec7f6e5bab06" Dec 07 16:58:57 crc kubenswrapper[4716]: I1207 16:58:57.242748 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af0f23bb18a48382b5165fcec913ec503c7c3fc81455a8d13adbec7f6e5bab06"} err="failed to get container status \"af0f23bb18a48382b5165fcec913ec503c7c3fc81455a8d13adbec7f6e5bab06\": rpc error: code = NotFound desc = could not find container \"af0f23bb18a48382b5165fcec913ec503c7c3fc81455a8d13adbec7f6e5bab06\": container with ID starting with af0f23bb18a48382b5165fcec913ec503c7c3fc81455a8d13adbec7f6e5bab06 not found: ID does not exist" Dec 07 16:58:57 crc kubenswrapper[4716]: I1207 16:58:57.669256 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d690dcc-677f-45a1-8ef0-fb1ecb4a3b4f" path="/var/lib/kubelet/pods/2d690dcc-677f-45a1-8ef0-fb1ecb4a3b4f/volumes" Dec 07 16:58:59 crc kubenswrapper[4716]: I1207 16:58:59.658362 4716 scope.go:117] "RemoveContainer" containerID="320114e03a98e202a688ac2d6205631a9f8d9952773c6722837647cd3ea2b74d" Dec 07 16:58:59 crc kubenswrapper[4716]: E1207 16:58:59.658965 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 16:59:11 crc kubenswrapper[4716]: I1207 16:59:11.657903 4716 scope.go:117] "RemoveContainer" containerID="320114e03a98e202a688ac2d6205631a9f8d9952773c6722837647cd3ea2b74d" Dec 07 16:59:11 crc kubenswrapper[4716]: E1207 16:59:11.658870 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 16:59:17 crc kubenswrapper[4716]: I1207 16:59:17.330455 4716 generic.go:334] "Generic (PLEG): container finished" podID="7c4bb2e7-29d0-4548-863b-ea89d11d68a7" containerID="73fea1b696cd9f27104920314586e5afab71cee2c49a27ca498d06a78f57428d" exitCode=0 Dec 07 16:59:17 crc kubenswrapper[4716]: I1207 16:59:17.330528 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"7c4bb2e7-29d0-4548-863b-ea89d11d68a7","Type":"ContainerDied","Data":"73fea1b696cd9f27104920314586e5afab71cee2c49a27ca498d06a78f57428d"} Dec 07 16:59:18 crc kubenswrapper[4716]: I1207 16:59:18.717355 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 07 16:59:18 crc kubenswrapper[4716]: I1207 16:59:18.819918 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"7c4bb2e7-29d0-4548-863b-ea89d11d68a7\" (UID: \"7c4bb2e7-29d0-4548-863b-ea89d11d68a7\") " Dec 07 16:59:18 crc kubenswrapper[4716]: I1207 16:59:18.819982 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/7c4bb2e7-29d0-4548-863b-ea89d11d68a7-test-operator-ephemeral-workdir\") pod \"7c4bb2e7-29d0-4548-863b-ea89d11d68a7\" (UID: \"7c4bb2e7-29d0-4548-863b-ea89d11d68a7\") " Dec 07 16:59:18 crc kubenswrapper[4716]: I1207 16:59:18.820040 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/7c4bb2e7-29d0-4548-863b-ea89d11d68a7-openstack-config-secret\") pod \"7c4bb2e7-29d0-4548-863b-ea89d11d68a7\" (UID: \"7c4bb2e7-29d0-4548-863b-ea89d11d68a7\") " Dec 07 16:59:18 crc kubenswrapper[4716]: I1207 16:59:18.820221 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/7c4bb2e7-29d0-4548-863b-ea89d11d68a7-ca-certs\") pod \"7c4bb2e7-29d0-4548-863b-ea89d11d68a7\" (UID: \"7c4bb2e7-29d0-4548-863b-ea89d11d68a7\") " Dec 07 16:59:18 crc kubenswrapper[4716]: I1207 16:59:18.820251 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/7c4bb2e7-29d0-4548-863b-ea89d11d68a7-test-operator-ephemeral-temporary\") pod \"7c4bb2e7-29d0-4548-863b-ea89d11d68a7\" (UID: \"7c4bb2e7-29d0-4548-863b-ea89d11d68a7\") " Dec 07 16:59:18 crc kubenswrapper[4716]: I1207 16:59:18.820281 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7c4bb2e7-29d0-4548-863b-ea89d11d68a7-config-data\") pod \"7c4bb2e7-29d0-4548-863b-ea89d11d68a7\" (UID: \"7c4bb2e7-29d0-4548-863b-ea89d11d68a7\") " Dec 07 16:59:18 crc kubenswrapper[4716]: I1207 16:59:18.820300 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/7c4bb2e7-29d0-4548-863b-ea89d11d68a7-openstack-config\") pod \"7c4bb2e7-29d0-4548-863b-ea89d11d68a7\" (UID: \"7c4bb2e7-29d0-4548-863b-ea89d11d68a7\") " Dec 07 16:59:18 crc kubenswrapper[4716]: I1207 16:59:18.820322 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7c4bb2e7-29d0-4548-863b-ea89d11d68a7-ssh-key\") pod \"7c4bb2e7-29d0-4548-863b-ea89d11d68a7\" (UID: \"7c4bb2e7-29d0-4548-863b-ea89d11d68a7\") " Dec 07 16:59:18 crc kubenswrapper[4716]: I1207 16:59:18.820369 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-476jr\" (UniqueName: \"kubernetes.io/projected/7c4bb2e7-29d0-4548-863b-ea89d11d68a7-kube-api-access-476jr\") pod \"7c4bb2e7-29d0-4548-863b-ea89d11d68a7\" (UID: \"7c4bb2e7-29d0-4548-863b-ea89d11d68a7\") " Dec 07 16:59:18 crc kubenswrapper[4716]: I1207 16:59:18.821188 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c4bb2e7-29d0-4548-863b-ea89d11d68a7-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "7c4bb2e7-29d0-4548-863b-ea89d11d68a7" (UID: "7c4bb2e7-29d0-4548-863b-ea89d11d68a7"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:59:18 crc kubenswrapper[4716]: I1207 16:59:18.822202 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c4bb2e7-29d0-4548-863b-ea89d11d68a7-config-data" (OuterVolumeSpecName: "config-data") pod "7c4bb2e7-29d0-4548-863b-ea89d11d68a7" (UID: "7c4bb2e7-29d0-4548-863b-ea89d11d68a7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:59:18 crc kubenswrapper[4716]: I1207 16:59:18.822355 4716 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/7c4bb2e7-29d0-4548-863b-ea89d11d68a7-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Dec 07 16:59:18 crc kubenswrapper[4716]: I1207 16:59:18.827366 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "test-operator-logs") pod "7c4bb2e7-29d0-4548-863b-ea89d11d68a7" (UID: "7c4bb2e7-29d0-4548-863b-ea89d11d68a7"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 07 16:59:18 crc kubenswrapper[4716]: I1207 16:59:18.827910 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c4bb2e7-29d0-4548-863b-ea89d11d68a7-kube-api-access-476jr" (OuterVolumeSpecName: "kube-api-access-476jr") pod "7c4bb2e7-29d0-4548-863b-ea89d11d68a7" (UID: "7c4bb2e7-29d0-4548-863b-ea89d11d68a7"). InnerVolumeSpecName "kube-api-access-476jr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:59:18 crc kubenswrapper[4716]: I1207 16:59:18.828314 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c4bb2e7-29d0-4548-863b-ea89d11d68a7-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "7c4bb2e7-29d0-4548-863b-ea89d11d68a7" (UID: "7c4bb2e7-29d0-4548-863b-ea89d11d68a7"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:59:18 crc kubenswrapper[4716]: I1207 16:59:18.849622 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c4bb2e7-29d0-4548-863b-ea89d11d68a7-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7c4bb2e7-29d0-4548-863b-ea89d11d68a7" (UID: "7c4bb2e7-29d0-4548-863b-ea89d11d68a7"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:59:18 crc kubenswrapper[4716]: I1207 16:59:18.866116 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c4bb2e7-29d0-4548-863b-ea89d11d68a7-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "7c4bb2e7-29d0-4548-863b-ea89d11d68a7" (UID: "7c4bb2e7-29d0-4548-863b-ea89d11d68a7"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:59:18 crc kubenswrapper[4716]: I1207 16:59:18.882527 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c4bb2e7-29d0-4548-863b-ea89d11d68a7-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "7c4bb2e7-29d0-4548-863b-ea89d11d68a7" (UID: "7c4bb2e7-29d0-4548-863b-ea89d11d68a7"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 16:59:18 crc kubenswrapper[4716]: I1207 16:59:18.891114 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c4bb2e7-29d0-4548-863b-ea89d11d68a7-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "7c4bb2e7-29d0-4548-863b-ea89d11d68a7" (UID: "7c4bb2e7-29d0-4548-863b-ea89d11d68a7"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 16:59:18 crc kubenswrapper[4716]: I1207 16:59:18.924571 4716 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/7c4bb2e7-29d0-4548-863b-ea89d11d68a7-ca-certs\") on node \"crc\" DevicePath \"\"" Dec 07 16:59:18 crc kubenswrapper[4716]: I1207 16:59:18.924614 4716 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/7c4bb2e7-29d0-4548-863b-ea89d11d68a7-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 07 16:59:18 crc kubenswrapper[4716]: I1207 16:59:18.924628 4716 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7c4bb2e7-29d0-4548-863b-ea89d11d68a7-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 16:59:18 crc kubenswrapper[4716]: I1207 16:59:18.924640 4716 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7c4bb2e7-29d0-4548-863b-ea89d11d68a7-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 07 16:59:18 crc kubenswrapper[4716]: I1207 16:59:18.924656 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-476jr\" (UniqueName: \"kubernetes.io/projected/7c4bb2e7-29d0-4548-863b-ea89d11d68a7-kube-api-access-476jr\") on node \"crc\" DevicePath \"\"" Dec 07 16:59:18 crc kubenswrapper[4716]: I1207 16:59:18.924694 4716 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Dec 07 16:59:18 crc kubenswrapper[4716]: I1207 16:59:18.924707 4716 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/7c4bb2e7-29d0-4548-863b-ea89d11d68a7-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Dec 07 16:59:18 crc kubenswrapper[4716]: I1207 16:59:18.924720 4716 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/7c4bb2e7-29d0-4548-863b-ea89d11d68a7-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 07 16:59:18 crc kubenswrapper[4716]: I1207 16:59:18.954031 4716 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Dec 07 16:59:19 crc kubenswrapper[4716]: I1207 16:59:19.026302 4716 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Dec 07 16:59:19 crc kubenswrapper[4716]: I1207 16:59:19.350442 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"7c4bb2e7-29d0-4548-863b-ea89d11d68a7","Type":"ContainerDied","Data":"9b4ab7ee63628f7976099cf1e42c2babc5b245cd708504e41763ba408bb35b43"} Dec 07 16:59:19 crc kubenswrapper[4716]: I1207 16:59:19.350694 4716 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9b4ab7ee63628f7976099cf1e42c2babc5b245cd708504e41763ba408bb35b43" Dec 07 16:59:19 crc kubenswrapper[4716]: I1207 16:59:19.350746 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 07 16:59:19 crc kubenswrapper[4716]: I1207 16:59:19.886674 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-krdhw"] Dec 07 16:59:19 crc kubenswrapper[4716]: E1207 16:59:19.887341 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d690dcc-677f-45a1-8ef0-fb1ecb4a3b4f" containerName="extract-utilities" Dec 07 16:59:19 crc kubenswrapper[4716]: I1207 16:59:19.887363 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d690dcc-677f-45a1-8ef0-fb1ecb4a3b4f" containerName="extract-utilities" Dec 07 16:59:19 crc kubenswrapper[4716]: E1207 16:59:19.887381 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d690dcc-677f-45a1-8ef0-fb1ecb4a3b4f" containerName="registry-server" Dec 07 16:59:19 crc kubenswrapper[4716]: I1207 16:59:19.887392 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d690dcc-677f-45a1-8ef0-fb1ecb4a3b4f" containerName="registry-server" Dec 07 16:59:19 crc kubenswrapper[4716]: E1207 16:59:19.887442 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c4bb2e7-29d0-4548-863b-ea89d11d68a7" containerName="tempest-tests-tempest-tests-runner" Dec 07 16:59:19 crc kubenswrapper[4716]: I1207 16:59:19.887455 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c4bb2e7-29d0-4548-863b-ea89d11d68a7" containerName="tempest-tests-tempest-tests-runner" Dec 07 16:59:19 crc kubenswrapper[4716]: E1207 16:59:19.887486 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d690dcc-677f-45a1-8ef0-fb1ecb4a3b4f" containerName="extract-content" Dec 07 16:59:19 crc kubenswrapper[4716]: I1207 16:59:19.887498 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d690dcc-677f-45a1-8ef0-fb1ecb4a3b4f" containerName="extract-content" Dec 07 16:59:19 crc kubenswrapper[4716]: I1207 16:59:19.887811 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d690dcc-677f-45a1-8ef0-fb1ecb4a3b4f" containerName="registry-server" Dec 07 16:59:19 crc kubenswrapper[4716]: I1207 16:59:19.887848 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c4bb2e7-29d0-4548-863b-ea89d11d68a7" containerName="tempest-tests-tempest-tests-runner" Dec 07 16:59:19 crc kubenswrapper[4716]: I1207 16:59:19.889829 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-krdhw" Dec 07 16:59:19 crc kubenswrapper[4716]: I1207 16:59:19.901392 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-krdhw"] Dec 07 16:59:19 crc kubenswrapper[4716]: I1207 16:59:19.946052 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ceceb39-b078-4dd9-9ca7-3ac9a216e75b-catalog-content\") pod \"certified-operators-krdhw\" (UID: \"7ceceb39-b078-4dd9-9ca7-3ac9a216e75b\") " pod="openshift-marketplace/certified-operators-krdhw" Dec 07 16:59:19 crc kubenswrapper[4716]: I1207 16:59:19.946357 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vb7mn\" (UniqueName: \"kubernetes.io/projected/7ceceb39-b078-4dd9-9ca7-3ac9a216e75b-kube-api-access-vb7mn\") pod \"certified-operators-krdhw\" (UID: \"7ceceb39-b078-4dd9-9ca7-3ac9a216e75b\") " pod="openshift-marketplace/certified-operators-krdhw" Dec 07 16:59:19 crc kubenswrapper[4716]: I1207 16:59:19.946638 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ceceb39-b078-4dd9-9ca7-3ac9a216e75b-utilities\") pod \"certified-operators-krdhw\" (UID: \"7ceceb39-b078-4dd9-9ca7-3ac9a216e75b\") " pod="openshift-marketplace/certified-operators-krdhw" Dec 07 16:59:20 crc kubenswrapper[4716]: I1207 16:59:20.048215 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ceceb39-b078-4dd9-9ca7-3ac9a216e75b-catalog-content\") pod \"certified-operators-krdhw\" (UID: \"7ceceb39-b078-4dd9-9ca7-3ac9a216e75b\") " pod="openshift-marketplace/certified-operators-krdhw" Dec 07 16:59:20 crc kubenswrapper[4716]: I1207 16:59:20.048457 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vb7mn\" (UniqueName: \"kubernetes.io/projected/7ceceb39-b078-4dd9-9ca7-3ac9a216e75b-kube-api-access-vb7mn\") pod \"certified-operators-krdhw\" (UID: \"7ceceb39-b078-4dd9-9ca7-3ac9a216e75b\") " pod="openshift-marketplace/certified-operators-krdhw" Dec 07 16:59:20 crc kubenswrapper[4716]: I1207 16:59:20.048682 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ceceb39-b078-4dd9-9ca7-3ac9a216e75b-utilities\") pod \"certified-operators-krdhw\" (UID: \"7ceceb39-b078-4dd9-9ca7-3ac9a216e75b\") " pod="openshift-marketplace/certified-operators-krdhw" Dec 07 16:59:20 crc kubenswrapper[4716]: I1207 16:59:20.048836 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ceceb39-b078-4dd9-9ca7-3ac9a216e75b-catalog-content\") pod \"certified-operators-krdhw\" (UID: \"7ceceb39-b078-4dd9-9ca7-3ac9a216e75b\") " pod="openshift-marketplace/certified-operators-krdhw" Dec 07 16:59:20 crc kubenswrapper[4716]: I1207 16:59:20.049275 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ceceb39-b078-4dd9-9ca7-3ac9a216e75b-utilities\") pod \"certified-operators-krdhw\" (UID: \"7ceceb39-b078-4dd9-9ca7-3ac9a216e75b\") " pod="openshift-marketplace/certified-operators-krdhw" Dec 07 16:59:20 crc kubenswrapper[4716]: I1207 16:59:20.072956 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vb7mn\" (UniqueName: \"kubernetes.io/projected/7ceceb39-b078-4dd9-9ca7-3ac9a216e75b-kube-api-access-vb7mn\") pod \"certified-operators-krdhw\" (UID: \"7ceceb39-b078-4dd9-9ca7-3ac9a216e75b\") " pod="openshift-marketplace/certified-operators-krdhw" Dec 07 16:59:20 crc kubenswrapper[4716]: I1207 16:59:20.215387 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-krdhw" Dec 07 16:59:20 crc kubenswrapper[4716]: I1207 16:59:20.705007 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-krdhw"] Dec 07 16:59:21 crc kubenswrapper[4716]: I1207 16:59:21.373192 4716 generic.go:334] "Generic (PLEG): container finished" podID="7ceceb39-b078-4dd9-9ca7-3ac9a216e75b" containerID="e65d50073654acd3b615b62ea25e46611d25d301deb35d224b2de9bc0fb55ee3" exitCode=0 Dec 07 16:59:21 crc kubenswrapper[4716]: I1207 16:59:21.373309 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-krdhw" event={"ID":"7ceceb39-b078-4dd9-9ca7-3ac9a216e75b","Type":"ContainerDied","Data":"e65d50073654acd3b615b62ea25e46611d25d301deb35d224b2de9bc0fb55ee3"} Dec 07 16:59:21 crc kubenswrapper[4716]: I1207 16:59:21.373544 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-krdhw" event={"ID":"7ceceb39-b078-4dd9-9ca7-3ac9a216e75b","Type":"ContainerStarted","Data":"588e36d4f74b1760a8a4d766162da6e1f9db4987f5317d282638f19fa189d16a"} Dec 07 16:59:22 crc kubenswrapper[4716]: I1207 16:59:22.389589 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-krdhw" event={"ID":"7ceceb39-b078-4dd9-9ca7-3ac9a216e75b","Type":"ContainerStarted","Data":"63c5c3fc559afbb8c0d7e3fef08fbdcc0368cc2d6f4eb30e3383c399d982e14e"} Dec 07 16:59:23 crc kubenswrapper[4716]: I1207 16:59:23.402744 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-krdhw" event={"ID":"7ceceb39-b078-4dd9-9ca7-3ac9a216e75b","Type":"ContainerDied","Data":"63c5c3fc559afbb8c0d7e3fef08fbdcc0368cc2d6f4eb30e3383c399d982e14e"} Dec 07 16:59:23 crc kubenswrapper[4716]: I1207 16:59:23.402672 4716 generic.go:334] "Generic (PLEG): container finished" podID="7ceceb39-b078-4dd9-9ca7-3ac9a216e75b" containerID="63c5c3fc559afbb8c0d7e3fef08fbdcc0368cc2d6f4eb30e3383c399d982e14e" exitCode=0 Dec 07 16:59:23 crc kubenswrapper[4716]: I1207 16:59:23.468019 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 07 16:59:23 crc kubenswrapper[4716]: I1207 16:59:23.469285 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 07 16:59:23 crc kubenswrapper[4716]: I1207 16:59:23.472740 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-mq8wt" Dec 07 16:59:23 crc kubenswrapper[4716]: I1207 16:59:23.487116 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 07 16:59:23 crc kubenswrapper[4716]: I1207 16:59:23.621303 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"fa15bae7-7145-486c-8f55-53797c60c01c\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 07 16:59:23 crc kubenswrapper[4716]: I1207 16:59:23.621630 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9lgfk\" (UniqueName: \"kubernetes.io/projected/fa15bae7-7145-486c-8f55-53797c60c01c-kube-api-access-9lgfk\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"fa15bae7-7145-486c-8f55-53797c60c01c\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 07 16:59:23 crc kubenswrapper[4716]: I1207 16:59:23.723576 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"fa15bae7-7145-486c-8f55-53797c60c01c\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 07 16:59:23 crc kubenswrapper[4716]: I1207 16:59:23.723616 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9lgfk\" (UniqueName: \"kubernetes.io/projected/fa15bae7-7145-486c-8f55-53797c60c01c-kube-api-access-9lgfk\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"fa15bae7-7145-486c-8f55-53797c60c01c\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 07 16:59:23 crc kubenswrapper[4716]: I1207 16:59:23.724034 4716 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"fa15bae7-7145-486c-8f55-53797c60c01c\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 07 16:59:23 crc kubenswrapper[4716]: I1207 16:59:23.747158 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9lgfk\" (UniqueName: \"kubernetes.io/projected/fa15bae7-7145-486c-8f55-53797c60c01c-kube-api-access-9lgfk\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"fa15bae7-7145-486c-8f55-53797c60c01c\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 07 16:59:23 crc kubenswrapper[4716]: I1207 16:59:23.750037 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"fa15bae7-7145-486c-8f55-53797c60c01c\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 07 16:59:23 crc kubenswrapper[4716]: I1207 16:59:23.836755 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 07 16:59:24 crc kubenswrapper[4716]: I1207 16:59:24.353813 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 07 16:59:24 crc kubenswrapper[4716]: W1207 16:59:24.357209 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfa15bae7_7145_486c_8f55_53797c60c01c.slice/crio-a73d31e59df829004e46ca097dae84df1bd9b89ac0682ebe98eaa4f125502165 WatchSource:0}: Error finding container a73d31e59df829004e46ca097dae84df1bd9b89ac0682ebe98eaa4f125502165: Status 404 returned error can't find the container with id a73d31e59df829004e46ca097dae84df1bd9b89ac0682ebe98eaa4f125502165 Dec 07 16:59:24 crc kubenswrapper[4716]: I1207 16:59:24.415261 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-krdhw" event={"ID":"7ceceb39-b078-4dd9-9ca7-3ac9a216e75b","Type":"ContainerStarted","Data":"743d827848da7997b2df28e7ef4131ae923a652ec2b3b9a15dab4b6799067548"} Dec 07 16:59:24 crc kubenswrapper[4716]: I1207 16:59:24.417907 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"fa15bae7-7145-486c-8f55-53797c60c01c","Type":"ContainerStarted","Data":"a73d31e59df829004e46ca097dae84df1bd9b89ac0682ebe98eaa4f125502165"} Dec 07 16:59:24 crc kubenswrapper[4716]: I1207 16:59:24.433276 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-krdhw" podStartSLOduration=2.936238211 podStartE2EDuration="5.433257592s" podCreationTimestamp="2025-12-07 16:59:19 +0000 UTC" firstStartedPulling="2025-12-07 16:59:21.374894777 +0000 UTC m=+3424.065179709" lastFinishedPulling="2025-12-07 16:59:23.871914158 +0000 UTC m=+3426.562199090" observedRunningTime="2025-12-07 16:59:24.432397559 +0000 UTC m=+3427.122682481" watchObservedRunningTime="2025-12-07 16:59:24.433257592 +0000 UTC m=+3427.123542504" Dec 07 16:59:24 crc kubenswrapper[4716]: I1207 16:59:24.657561 4716 scope.go:117] "RemoveContainer" containerID="320114e03a98e202a688ac2d6205631a9f8d9952773c6722837647cd3ea2b74d" Dec 07 16:59:25 crc kubenswrapper[4716]: I1207 16:59:25.438211 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" event={"ID":"c15b59eb-565d-4556-a4ce-75afdf159dc8","Type":"ContainerStarted","Data":"79a8e8c2fa02cd8c837e2fe1d95aefde9b087d4864372423b2cc8afd41b4df75"} Dec 07 16:59:26 crc kubenswrapper[4716]: I1207 16:59:26.446484 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"fa15bae7-7145-486c-8f55-53797c60c01c","Type":"ContainerStarted","Data":"4db92f4f0d04c17efff9ddc382008839ad62e223df440c0060b40570ad8a756d"} Dec 07 16:59:26 crc kubenswrapper[4716]: I1207 16:59:26.465741 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=2.556335773 podStartE2EDuration="3.465724575s" podCreationTimestamp="2025-12-07 16:59:23 +0000 UTC" firstStartedPulling="2025-12-07 16:59:24.359287195 +0000 UTC m=+3427.049572107" lastFinishedPulling="2025-12-07 16:59:25.268676007 +0000 UTC m=+3427.958960909" observedRunningTime="2025-12-07 16:59:26.457320507 +0000 UTC m=+3429.147605499" watchObservedRunningTime="2025-12-07 16:59:26.465724575 +0000 UTC m=+3429.156009487" Dec 07 16:59:30 crc kubenswrapper[4716]: I1207 16:59:30.216024 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-krdhw" Dec 07 16:59:30 crc kubenswrapper[4716]: I1207 16:59:30.216677 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-krdhw" Dec 07 16:59:30 crc kubenswrapper[4716]: I1207 16:59:30.275397 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-krdhw" Dec 07 16:59:30 crc kubenswrapper[4716]: I1207 16:59:30.525680 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-krdhw" Dec 07 16:59:30 crc kubenswrapper[4716]: I1207 16:59:30.584257 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-krdhw"] Dec 07 16:59:32 crc kubenswrapper[4716]: I1207 16:59:32.560502 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-krdhw" podUID="7ceceb39-b078-4dd9-9ca7-3ac9a216e75b" containerName="registry-server" containerID="cri-o://743d827848da7997b2df28e7ef4131ae923a652ec2b3b9a15dab4b6799067548" gracePeriod=2 Dec 07 16:59:33 crc kubenswrapper[4716]: I1207 16:59:33.001093 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-krdhw" Dec 07 16:59:33 crc kubenswrapper[4716]: I1207 16:59:33.113187 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vb7mn\" (UniqueName: \"kubernetes.io/projected/7ceceb39-b078-4dd9-9ca7-3ac9a216e75b-kube-api-access-vb7mn\") pod \"7ceceb39-b078-4dd9-9ca7-3ac9a216e75b\" (UID: \"7ceceb39-b078-4dd9-9ca7-3ac9a216e75b\") " Dec 07 16:59:33 crc kubenswrapper[4716]: I1207 16:59:33.113711 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ceceb39-b078-4dd9-9ca7-3ac9a216e75b-catalog-content\") pod \"7ceceb39-b078-4dd9-9ca7-3ac9a216e75b\" (UID: \"7ceceb39-b078-4dd9-9ca7-3ac9a216e75b\") " Dec 07 16:59:33 crc kubenswrapper[4716]: I1207 16:59:33.113861 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ceceb39-b078-4dd9-9ca7-3ac9a216e75b-utilities\") pod \"7ceceb39-b078-4dd9-9ca7-3ac9a216e75b\" (UID: \"7ceceb39-b078-4dd9-9ca7-3ac9a216e75b\") " Dec 07 16:59:33 crc kubenswrapper[4716]: I1207 16:59:33.115546 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ceceb39-b078-4dd9-9ca7-3ac9a216e75b-utilities" (OuterVolumeSpecName: "utilities") pod "7ceceb39-b078-4dd9-9ca7-3ac9a216e75b" (UID: "7ceceb39-b078-4dd9-9ca7-3ac9a216e75b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:59:33 crc kubenswrapper[4716]: I1207 16:59:33.126248 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ceceb39-b078-4dd9-9ca7-3ac9a216e75b-kube-api-access-vb7mn" (OuterVolumeSpecName: "kube-api-access-vb7mn") pod "7ceceb39-b078-4dd9-9ca7-3ac9a216e75b" (UID: "7ceceb39-b078-4dd9-9ca7-3ac9a216e75b"). InnerVolumeSpecName "kube-api-access-vb7mn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 16:59:33 crc kubenswrapper[4716]: I1207 16:59:33.169477 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ceceb39-b078-4dd9-9ca7-3ac9a216e75b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7ceceb39-b078-4dd9-9ca7-3ac9a216e75b" (UID: "7ceceb39-b078-4dd9-9ca7-3ac9a216e75b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 16:59:33 crc kubenswrapper[4716]: I1207 16:59:33.216640 4716 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ceceb39-b078-4dd9-9ca7-3ac9a216e75b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 07 16:59:33 crc kubenswrapper[4716]: I1207 16:59:33.216840 4716 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ceceb39-b078-4dd9-9ca7-3ac9a216e75b-utilities\") on node \"crc\" DevicePath \"\"" Dec 07 16:59:33 crc kubenswrapper[4716]: I1207 16:59:33.216911 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vb7mn\" (UniqueName: \"kubernetes.io/projected/7ceceb39-b078-4dd9-9ca7-3ac9a216e75b-kube-api-access-vb7mn\") on node \"crc\" DevicePath \"\"" Dec 07 16:59:33 crc kubenswrapper[4716]: I1207 16:59:33.574115 4716 generic.go:334] "Generic (PLEG): container finished" podID="7ceceb39-b078-4dd9-9ca7-3ac9a216e75b" containerID="743d827848da7997b2df28e7ef4131ae923a652ec2b3b9a15dab4b6799067548" exitCode=0 Dec 07 16:59:33 crc kubenswrapper[4716]: I1207 16:59:33.574171 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-krdhw" event={"ID":"7ceceb39-b078-4dd9-9ca7-3ac9a216e75b","Type":"ContainerDied","Data":"743d827848da7997b2df28e7ef4131ae923a652ec2b3b9a15dab4b6799067548"} Dec 07 16:59:33 crc kubenswrapper[4716]: I1207 16:59:33.574248 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-krdhw" Dec 07 16:59:33 crc kubenswrapper[4716]: I1207 16:59:33.574272 4716 scope.go:117] "RemoveContainer" containerID="743d827848da7997b2df28e7ef4131ae923a652ec2b3b9a15dab4b6799067548" Dec 07 16:59:33 crc kubenswrapper[4716]: I1207 16:59:33.574254 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-krdhw" event={"ID":"7ceceb39-b078-4dd9-9ca7-3ac9a216e75b","Type":"ContainerDied","Data":"588e36d4f74b1760a8a4d766162da6e1f9db4987f5317d282638f19fa189d16a"} Dec 07 16:59:33 crc kubenswrapper[4716]: I1207 16:59:33.603767 4716 scope.go:117] "RemoveContainer" containerID="63c5c3fc559afbb8c0d7e3fef08fbdcc0368cc2d6f4eb30e3383c399d982e14e" Dec 07 16:59:33 crc kubenswrapper[4716]: I1207 16:59:33.629873 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-krdhw"] Dec 07 16:59:33 crc kubenswrapper[4716]: I1207 16:59:33.639786 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-krdhw"] Dec 07 16:59:33 crc kubenswrapper[4716]: I1207 16:59:33.646430 4716 scope.go:117] "RemoveContainer" containerID="e65d50073654acd3b615b62ea25e46611d25d301deb35d224b2de9bc0fb55ee3" Dec 07 16:59:33 crc kubenswrapper[4716]: I1207 16:59:33.678052 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ceceb39-b078-4dd9-9ca7-3ac9a216e75b" path="/var/lib/kubelet/pods/7ceceb39-b078-4dd9-9ca7-3ac9a216e75b/volumes" Dec 07 16:59:33 crc kubenswrapper[4716]: I1207 16:59:33.681431 4716 scope.go:117] "RemoveContainer" containerID="743d827848da7997b2df28e7ef4131ae923a652ec2b3b9a15dab4b6799067548" Dec 07 16:59:33 crc kubenswrapper[4716]: E1207 16:59:33.681789 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"743d827848da7997b2df28e7ef4131ae923a652ec2b3b9a15dab4b6799067548\": container with ID starting with 743d827848da7997b2df28e7ef4131ae923a652ec2b3b9a15dab4b6799067548 not found: ID does not exist" containerID="743d827848da7997b2df28e7ef4131ae923a652ec2b3b9a15dab4b6799067548" Dec 07 16:59:33 crc kubenswrapper[4716]: I1207 16:59:33.681818 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"743d827848da7997b2df28e7ef4131ae923a652ec2b3b9a15dab4b6799067548"} err="failed to get container status \"743d827848da7997b2df28e7ef4131ae923a652ec2b3b9a15dab4b6799067548\": rpc error: code = NotFound desc = could not find container \"743d827848da7997b2df28e7ef4131ae923a652ec2b3b9a15dab4b6799067548\": container with ID starting with 743d827848da7997b2df28e7ef4131ae923a652ec2b3b9a15dab4b6799067548 not found: ID does not exist" Dec 07 16:59:33 crc kubenswrapper[4716]: I1207 16:59:33.681834 4716 scope.go:117] "RemoveContainer" containerID="63c5c3fc559afbb8c0d7e3fef08fbdcc0368cc2d6f4eb30e3383c399d982e14e" Dec 07 16:59:33 crc kubenswrapper[4716]: E1207 16:59:33.682152 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"63c5c3fc559afbb8c0d7e3fef08fbdcc0368cc2d6f4eb30e3383c399d982e14e\": container with ID starting with 63c5c3fc559afbb8c0d7e3fef08fbdcc0368cc2d6f4eb30e3383c399d982e14e not found: ID does not exist" containerID="63c5c3fc559afbb8c0d7e3fef08fbdcc0368cc2d6f4eb30e3383c399d982e14e" Dec 07 16:59:33 crc kubenswrapper[4716]: I1207 16:59:33.682179 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"63c5c3fc559afbb8c0d7e3fef08fbdcc0368cc2d6f4eb30e3383c399d982e14e"} err="failed to get container status \"63c5c3fc559afbb8c0d7e3fef08fbdcc0368cc2d6f4eb30e3383c399d982e14e\": rpc error: code = NotFound desc = could not find container \"63c5c3fc559afbb8c0d7e3fef08fbdcc0368cc2d6f4eb30e3383c399d982e14e\": container with ID starting with 63c5c3fc559afbb8c0d7e3fef08fbdcc0368cc2d6f4eb30e3383c399d982e14e not found: ID does not exist" Dec 07 16:59:33 crc kubenswrapper[4716]: I1207 16:59:33.682191 4716 scope.go:117] "RemoveContainer" containerID="e65d50073654acd3b615b62ea25e46611d25d301deb35d224b2de9bc0fb55ee3" Dec 07 16:59:33 crc kubenswrapper[4716]: E1207 16:59:33.682381 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e65d50073654acd3b615b62ea25e46611d25d301deb35d224b2de9bc0fb55ee3\": container with ID starting with e65d50073654acd3b615b62ea25e46611d25d301deb35d224b2de9bc0fb55ee3 not found: ID does not exist" containerID="e65d50073654acd3b615b62ea25e46611d25d301deb35d224b2de9bc0fb55ee3" Dec 07 16:59:33 crc kubenswrapper[4716]: I1207 16:59:33.682403 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e65d50073654acd3b615b62ea25e46611d25d301deb35d224b2de9bc0fb55ee3"} err="failed to get container status \"e65d50073654acd3b615b62ea25e46611d25d301deb35d224b2de9bc0fb55ee3\": rpc error: code = NotFound desc = could not find container \"e65d50073654acd3b615b62ea25e46611d25d301deb35d224b2de9bc0fb55ee3\": container with ID starting with e65d50073654acd3b615b62ea25e46611d25d301deb35d224b2de9bc0fb55ee3 not found: ID does not exist" Dec 07 16:59:48 crc kubenswrapper[4716]: I1207 16:59:48.049311 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-zlz52/must-gather-nx26c"] Dec 07 16:59:48 crc kubenswrapper[4716]: E1207 16:59:48.050397 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ceceb39-b078-4dd9-9ca7-3ac9a216e75b" containerName="registry-server" Dec 07 16:59:48 crc kubenswrapper[4716]: I1207 16:59:48.050420 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ceceb39-b078-4dd9-9ca7-3ac9a216e75b" containerName="registry-server" Dec 07 16:59:48 crc kubenswrapper[4716]: E1207 16:59:48.050442 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ceceb39-b078-4dd9-9ca7-3ac9a216e75b" containerName="extract-utilities" Dec 07 16:59:48 crc kubenswrapper[4716]: I1207 16:59:48.050450 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ceceb39-b078-4dd9-9ca7-3ac9a216e75b" containerName="extract-utilities" Dec 07 16:59:48 crc kubenswrapper[4716]: E1207 16:59:48.050474 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ceceb39-b078-4dd9-9ca7-3ac9a216e75b" containerName="extract-content" Dec 07 16:59:48 crc kubenswrapper[4716]: I1207 16:59:48.050482 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ceceb39-b078-4dd9-9ca7-3ac9a216e75b" containerName="extract-content" Dec 07 16:59:48 crc kubenswrapper[4716]: I1207 16:59:48.050780 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ceceb39-b078-4dd9-9ca7-3ac9a216e75b" containerName="registry-server" Dec 07 16:59:48 crc kubenswrapper[4716]: I1207 16:59:48.054308 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zlz52/must-gather-nx26c" Dec 07 16:59:48 crc kubenswrapper[4716]: I1207 16:59:48.056052 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-zlz52"/"default-dockercfg-zfg2v" Dec 07 16:59:48 crc kubenswrapper[4716]: I1207 16:59:48.057359 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-zlz52"/"kube-root-ca.crt" Dec 07 16:59:48 crc kubenswrapper[4716]: I1207 16:59:48.057446 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-zlz52"/"openshift-service-ca.crt" Dec 07 16:59:48 crc kubenswrapper[4716]: I1207 16:59:48.072153 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-zlz52/must-gather-nx26c"] Dec 07 16:59:48 crc kubenswrapper[4716]: I1207 16:59:48.146147 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4vc6\" (UniqueName: \"kubernetes.io/projected/68a27046-ae9a-4651-98fd-d5633838e785-kube-api-access-l4vc6\") pod \"must-gather-nx26c\" (UID: \"68a27046-ae9a-4651-98fd-d5633838e785\") " pod="openshift-must-gather-zlz52/must-gather-nx26c" Dec 07 16:59:48 crc kubenswrapper[4716]: I1207 16:59:48.146204 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/68a27046-ae9a-4651-98fd-d5633838e785-must-gather-output\") pod \"must-gather-nx26c\" (UID: \"68a27046-ae9a-4651-98fd-d5633838e785\") " pod="openshift-must-gather-zlz52/must-gather-nx26c" Dec 07 16:59:48 crc kubenswrapper[4716]: I1207 16:59:48.247544 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4vc6\" (UniqueName: \"kubernetes.io/projected/68a27046-ae9a-4651-98fd-d5633838e785-kube-api-access-l4vc6\") pod \"must-gather-nx26c\" (UID: \"68a27046-ae9a-4651-98fd-d5633838e785\") " pod="openshift-must-gather-zlz52/must-gather-nx26c" Dec 07 16:59:48 crc kubenswrapper[4716]: I1207 16:59:48.247884 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/68a27046-ae9a-4651-98fd-d5633838e785-must-gather-output\") pod \"must-gather-nx26c\" (UID: \"68a27046-ae9a-4651-98fd-d5633838e785\") " pod="openshift-must-gather-zlz52/must-gather-nx26c" Dec 07 16:59:48 crc kubenswrapper[4716]: I1207 16:59:48.248522 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/68a27046-ae9a-4651-98fd-d5633838e785-must-gather-output\") pod \"must-gather-nx26c\" (UID: \"68a27046-ae9a-4651-98fd-d5633838e785\") " pod="openshift-must-gather-zlz52/must-gather-nx26c" Dec 07 16:59:48 crc kubenswrapper[4716]: I1207 16:59:48.271380 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4vc6\" (UniqueName: \"kubernetes.io/projected/68a27046-ae9a-4651-98fd-d5633838e785-kube-api-access-l4vc6\") pod \"must-gather-nx26c\" (UID: \"68a27046-ae9a-4651-98fd-d5633838e785\") " pod="openshift-must-gather-zlz52/must-gather-nx26c" Dec 07 16:59:48 crc kubenswrapper[4716]: I1207 16:59:48.381792 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zlz52/must-gather-nx26c" Dec 07 16:59:48 crc kubenswrapper[4716]: I1207 16:59:48.822198 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-zlz52/must-gather-nx26c"] Dec 07 16:59:49 crc kubenswrapper[4716]: I1207 16:59:49.977139 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zlz52/must-gather-nx26c" event={"ID":"68a27046-ae9a-4651-98fd-d5633838e785","Type":"ContainerStarted","Data":"90fa38373aeec0a9b9c0ae2ef7cad382f8674422d21630e0b7db19e4580ecd6c"} Dec 07 16:59:56 crc kubenswrapper[4716]: I1207 16:59:56.014202 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zlz52/must-gather-nx26c" event={"ID":"68a27046-ae9a-4651-98fd-d5633838e785","Type":"ContainerStarted","Data":"7c6bdbfe919aa4baaebaa598b7bbb8f5e7594bb501a25a21f7e642540ca5eb48"} Dec 07 16:59:57 crc kubenswrapper[4716]: I1207 16:59:57.024018 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zlz52/must-gather-nx26c" event={"ID":"68a27046-ae9a-4651-98fd-d5633838e785","Type":"ContainerStarted","Data":"3313afc325e5011a91e073a0f91f05c04984c5290f76cca9a6a62a3c75663fa7"} Dec 07 16:59:59 crc kubenswrapper[4716]: I1207 16:59:59.416445 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-zlz52/must-gather-nx26c" podStartSLOduration=4.704518178 podStartE2EDuration="11.416427393s" podCreationTimestamp="2025-12-07 16:59:48 +0000 UTC" firstStartedPulling="2025-12-07 16:59:48.832783197 +0000 UTC m=+3451.523068109" lastFinishedPulling="2025-12-07 16:59:55.544692412 +0000 UTC m=+3458.234977324" observedRunningTime="2025-12-07 16:59:57.048360063 +0000 UTC m=+3459.738644975" watchObservedRunningTime="2025-12-07 16:59:59.416427393 +0000 UTC m=+3462.106712295" Dec 07 16:59:59 crc kubenswrapper[4716]: I1207 16:59:59.418426 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-zlz52/crc-debug-wnjwc"] Dec 07 16:59:59 crc kubenswrapper[4716]: I1207 16:59:59.419553 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zlz52/crc-debug-wnjwc" Dec 07 16:59:59 crc kubenswrapper[4716]: I1207 16:59:59.530574 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7ed612df-44de-47e2-9939-4947359889f5-host\") pod \"crc-debug-wnjwc\" (UID: \"7ed612df-44de-47e2-9939-4947359889f5\") " pod="openshift-must-gather-zlz52/crc-debug-wnjwc" Dec 07 16:59:59 crc kubenswrapper[4716]: I1207 16:59:59.530689 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vl5fl\" (UniqueName: \"kubernetes.io/projected/7ed612df-44de-47e2-9939-4947359889f5-kube-api-access-vl5fl\") pod \"crc-debug-wnjwc\" (UID: \"7ed612df-44de-47e2-9939-4947359889f5\") " pod="openshift-must-gather-zlz52/crc-debug-wnjwc" Dec 07 16:59:59 crc kubenswrapper[4716]: I1207 16:59:59.632422 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7ed612df-44de-47e2-9939-4947359889f5-host\") pod \"crc-debug-wnjwc\" (UID: \"7ed612df-44de-47e2-9939-4947359889f5\") " pod="openshift-must-gather-zlz52/crc-debug-wnjwc" Dec 07 16:59:59 crc kubenswrapper[4716]: I1207 16:59:59.632582 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7ed612df-44de-47e2-9939-4947359889f5-host\") pod \"crc-debug-wnjwc\" (UID: \"7ed612df-44de-47e2-9939-4947359889f5\") " pod="openshift-must-gather-zlz52/crc-debug-wnjwc" Dec 07 16:59:59 crc kubenswrapper[4716]: I1207 16:59:59.632609 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vl5fl\" (UniqueName: \"kubernetes.io/projected/7ed612df-44de-47e2-9939-4947359889f5-kube-api-access-vl5fl\") pod \"crc-debug-wnjwc\" (UID: \"7ed612df-44de-47e2-9939-4947359889f5\") " pod="openshift-must-gather-zlz52/crc-debug-wnjwc" Dec 07 16:59:59 crc kubenswrapper[4716]: I1207 16:59:59.651819 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vl5fl\" (UniqueName: \"kubernetes.io/projected/7ed612df-44de-47e2-9939-4947359889f5-kube-api-access-vl5fl\") pod \"crc-debug-wnjwc\" (UID: \"7ed612df-44de-47e2-9939-4947359889f5\") " pod="openshift-must-gather-zlz52/crc-debug-wnjwc" Dec 07 16:59:59 crc kubenswrapper[4716]: I1207 16:59:59.740248 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zlz52/crc-debug-wnjwc" Dec 07 16:59:59 crc kubenswrapper[4716]: W1207 16:59:59.791319 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7ed612df_44de_47e2_9939_4947359889f5.slice/crio-3510e4b86604639db0fe0251e6949d8fc0867a2a71a97b6c514ce5f46f77ff29 WatchSource:0}: Error finding container 3510e4b86604639db0fe0251e6949d8fc0867a2a71a97b6c514ce5f46f77ff29: Status 404 returned error can't find the container with id 3510e4b86604639db0fe0251e6949d8fc0867a2a71a97b6c514ce5f46f77ff29 Dec 07 17:00:00 crc kubenswrapper[4716]: I1207 17:00:00.055066 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zlz52/crc-debug-wnjwc" event={"ID":"7ed612df-44de-47e2-9939-4947359889f5","Type":"ContainerStarted","Data":"3510e4b86604639db0fe0251e6949d8fc0867a2a71a97b6c514ce5f46f77ff29"} Dec 07 17:00:00 crc kubenswrapper[4716]: I1207 17:00:00.150721 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29418780-wtdzr"] Dec 07 17:00:00 crc kubenswrapper[4716]: I1207 17:00:00.157186 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29418780-wtdzr" Dec 07 17:00:00 crc kubenswrapper[4716]: I1207 17:00:00.159497 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 07 17:00:00 crc kubenswrapper[4716]: I1207 17:00:00.159688 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 07 17:00:00 crc kubenswrapper[4716]: I1207 17:00:00.165571 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29418780-wtdzr"] Dec 07 17:00:00 crc kubenswrapper[4716]: I1207 17:00:00.346362 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/060923f1-6da9-427f-af9f-d6a365600623-secret-volume\") pod \"collect-profiles-29418780-wtdzr\" (UID: \"060923f1-6da9-427f-af9f-d6a365600623\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418780-wtdzr" Dec 07 17:00:00 crc kubenswrapper[4716]: I1207 17:00:00.346776 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/060923f1-6da9-427f-af9f-d6a365600623-config-volume\") pod \"collect-profiles-29418780-wtdzr\" (UID: \"060923f1-6da9-427f-af9f-d6a365600623\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418780-wtdzr" Dec 07 17:00:00 crc kubenswrapper[4716]: I1207 17:00:00.347161 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f29bb\" (UniqueName: \"kubernetes.io/projected/060923f1-6da9-427f-af9f-d6a365600623-kube-api-access-f29bb\") pod \"collect-profiles-29418780-wtdzr\" (UID: \"060923f1-6da9-427f-af9f-d6a365600623\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418780-wtdzr" Dec 07 17:00:00 crc kubenswrapper[4716]: I1207 17:00:00.449147 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f29bb\" (UniqueName: \"kubernetes.io/projected/060923f1-6da9-427f-af9f-d6a365600623-kube-api-access-f29bb\") pod \"collect-profiles-29418780-wtdzr\" (UID: \"060923f1-6da9-427f-af9f-d6a365600623\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418780-wtdzr" Dec 07 17:00:00 crc kubenswrapper[4716]: I1207 17:00:00.449213 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/060923f1-6da9-427f-af9f-d6a365600623-secret-volume\") pod \"collect-profiles-29418780-wtdzr\" (UID: \"060923f1-6da9-427f-af9f-d6a365600623\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418780-wtdzr" Dec 07 17:00:00 crc kubenswrapper[4716]: I1207 17:00:00.449264 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/060923f1-6da9-427f-af9f-d6a365600623-config-volume\") pod \"collect-profiles-29418780-wtdzr\" (UID: \"060923f1-6da9-427f-af9f-d6a365600623\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418780-wtdzr" Dec 07 17:00:00 crc kubenswrapper[4716]: I1207 17:00:00.450189 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/060923f1-6da9-427f-af9f-d6a365600623-config-volume\") pod \"collect-profiles-29418780-wtdzr\" (UID: \"060923f1-6da9-427f-af9f-d6a365600623\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418780-wtdzr" Dec 07 17:00:00 crc kubenswrapper[4716]: I1207 17:00:00.456606 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/060923f1-6da9-427f-af9f-d6a365600623-secret-volume\") pod \"collect-profiles-29418780-wtdzr\" (UID: \"060923f1-6da9-427f-af9f-d6a365600623\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418780-wtdzr" Dec 07 17:00:00 crc kubenswrapper[4716]: I1207 17:00:00.470779 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f29bb\" (UniqueName: \"kubernetes.io/projected/060923f1-6da9-427f-af9f-d6a365600623-kube-api-access-f29bb\") pod \"collect-profiles-29418780-wtdzr\" (UID: \"060923f1-6da9-427f-af9f-d6a365600623\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418780-wtdzr" Dec 07 17:00:00 crc kubenswrapper[4716]: I1207 17:00:00.476745 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29418780-wtdzr" Dec 07 17:00:00 crc kubenswrapper[4716]: I1207 17:00:00.978537 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29418780-wtdzr"] Dec 07 17:00:01 crc kubenswrapper[4716]: I1207 17:00:01.066544 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29418780-wtdzr" event={"ID":"060923f1-6da9-427f-af9f-d6a365600623","Type":"ContainerStarted","Data":"6227e25ca6328a89b79ec7ab6e747e2857e634ac7aa8e412b103fb236c6b9807"} Dec 07 17:00:02 crc kubenswrapper[4716]: I1207 17:00:02.078009 4716 generic.go:334] "Generic (PLEG): container finished" podID="060923f1-6da9-427f-af9f-d6a365600623" containerID="ab30be16e21ecbd617a095de226415e00a92c8f3efb1eb59ee03899359c46b00" exitCode=0 Dec 07 17:00:02 crc kubenswrapper[4716]: I1207 17:00:02.078130 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29418780-wtdzr" event={"ID":"060923f1-6da9-427f-af9f-d6a365600623","Type":"ContainerDied","Data":"ab30be16e21ecbd617a095de226415e00a92c8f3efb1eb59ee03899359c46b00"} Dec 07 17:00:03 crc kubenswrapper[4716]: I1207 17:00:03.445068 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29418780-wtdzr" Dec 07 17:00:03 crc kubenswrapper[4716]: I1207 17:00:03.604604 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f29bb\" (UniqueName: \"kubernetes.io/projected/060923f1-6da9-427f-af9f-d6a365600623-kube-api-access-f29bb\") pod \"060923f1-6da9-427f-af9f-d6a365600623\" (UID: \"060923f1-6da9-427f-af9f-d6a365600623\") " Dec 07 17:00:03 crc kubenswrapper[4716]: I1207 17:00:03.604783 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/060923f1-6da9-427f-af9f-d6a365600623-config-volume\") pod \"060923f1-6da9-427f-af9f-d6a365600623\" (UID: \"060923f1-6da9-427f-af9f-d6a365600623\") " Dec 07 17:00:03 crc kubenswrapper[4716]: I1207 17:00:03.604853 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/060923f1-6da9-427f-af9f-d6a365600623-secret-volume\") pod \"060923f1-6da9-427f-af9f-d6a365600623\" (UID: \"060923f1-6da9-427f-af9f-d6a365600623\") " Dec 07 17:00:03 crc kubenswrapper[4716]: I1207 17:00:03.606665 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/060923f1-6da9-427f-af9f-d6a365600623-config-volume" (OuterVolumeSpecName: "config-volume") pod "060923f1-6da9-427f-af9f-d6a365600623" (UID: "060923f1-6da9-427f-af9f-d6a365600623"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 17:00:03 crc kubenswrapper[4716]: I1207 17:00:03.611552 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/060923f1-6da9-427f-af9f-d6a365600623-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "060923f1-6da9-427f-af9f-d6a365600623" (UID: "060923f1-6da9-427f-af9f-d6a365600623"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 17:00:03 crc kubenswrapper[4716]: I1207 17:00:03.616330 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/060923f1-6da9-427f-af9f-d6a365600623-kube-api-access-f29bb" (OuterVolumeSpecName: "kube-api-access-f29bb") pod "060923f1-6da9-427f-af9f-d6a365600623" (UID: "060923f1-6da9-427f-af9f-d6a365600623"). InnerVolumeSpecName "kube-api-access-f29bb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 17:00:03 crc kubenswrapper[4716]: I1207 17:00:03.706795 4716 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/060923f1-6da9-427f-af9f-d6a365600623-config-volume\") on node \"crc\" DevicePath \"\"" Dec 07 17:00:03 crc kubenswrapper[4716]: I1207 17:00:03.706827 4716 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/060923f1-6da9-427f-af9f-d6a365600623-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 07 17:00:03 crc kubenswrapper[4716]: I1207 17:00:03.706837 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f29bb\" (UniqueName: \"kubernetes.io/projected/060923f1-6da9-427f-af9f-d6a365600623-kube-api-access-f29bb\") on node \"crc\" DevicePath \"\"" Dec 07 17:00:04 crc kubenswrapper[4716]: I1207 17:00:04.106993 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29418780-wtdzr" event={"ID":"060923f1-6da9-427f-af9f-d6a365600623","Type":"ContainerDied","Data":"6227e25ca6328a89b79ec7ab6e747e2857e634ac7aa8e412b103fb236c6b9807"} Dec 07 17:00:04 crc kubenswrapper[4716]: I1207 17:00:04.107068 4716 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6227e25ca6328a89b79ec7ab6e747e2857e634ac7aa8e412b103fb236c6b9807" Dec 07 17:00:04 crc kubenswrapper[4716]: I1207 17:00:04.107033 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29418780-wtdzr" Dec 07 17:00:04 crc kubenswrapper[4716]: I1207 17:00:04.533391 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29418735-swmgk"] Dec 07 17:00:04 crc kubenswrapper[4716]: I1207 17:00:04.541520 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29418735-swmgk"] Dec 07 17:00:05 crc kubenswrapper[4716]: I1207 17:00:05.668701 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e624e76-4467-44cc-b0d0-520ced143ba2" path="/var/lib/kubelet/pods/9e624e76-4467-44cc-b0d0-520ced143ba2/volumes" Dec 07 17:00:11 crc kubenswrapper[4716]: I1207 17:00:11.173011 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zlz52/crc-debug-wnjwc" event={"ID":"7ed612df-44de-47e2-9939-4947359889f5","Type":"ContainerStarted","Data":"efd6def30ea2b205dacfcda73da7b1141d929863697261c072b99be752c3caab"} Dec 07 17:00:11 crc kubenswrapper[4716]: I1207 17:00:11.196619 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-zlz52/crc-debug-wnjwc" podStartSLOduration=1.434720379 podStartE2EDuration="12.196604333s" podCreationTimestamp="2025-12-07 16:59:59 +0000 UTC" firstStartedPulling="2025-12-07 16:59:59.793786515 +0000 UTC m=+3462.484071427" lastFinishedPulling="2025-12-07 17:00:10.555670469 +0000 UTC m=+3473.245955381" observedRunningTime="2025-12-07 17:00:11.192224684 +0000 UTC m=+3473.882509596" watchObservedRunningTime="2025-12-07 17:00:11.196604333 +0000 UTC m=+3473.886889245" Dec 07 17:00:35 crc kubenswrapper[4716]: I1207 17:00:35.275768 4716 scope.go:117] "RemoveContainer" containerID="c5cf5db4504ee2f9b599aae7d13d4c19913d206ce7bf19a6754a54d7185b83ce" Dec 07 17:00:52 crc kubenswrapper[4716]: I1207 17:00:52.561217 4716 generic.go:334] "Generic (PLEG): container finished" podID="7ed612df-44de-47e2-9939-4947359889f5" containerID="efd6def30ea2b205dacfcda73da7b1141d929863697261c072b99be752c3caab" exitCode=0 Dec 07 17:00:52 crc kubenswrapper[4716]: I1207 17:00:52.561957 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zlz52/crc-debug-wnjwc" event={"ID":"7ed612df-44de-47e2-9939-4947359889f5","Type":"ContainerDied","Data":"efd6def30ea2b205dacfcda73da7b1141d929863697261c072b99be752c3caab"} Dec 07 17:00:53 crc kubenswrapper[4716]: I1207 17:00:53.688502 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zlz52/crc-debug-wnjwc" Dec 07 17:00:53 crc kubenswrapper[4716]: I1207 17:00:53.739614 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-zlz52/crc-debug-wnjwc"] Dec 07 17:00:53 crc kubenswrapper[4716]: I1207 17:00:53.756990 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-zlz52/crc-debug-wnjwc"] Dec 07 17:00:53 crc kubenswrapper[4716]: I1207 17:00:53.858059 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7ed612df-44de-47e2-9939-4947359889f5-host\") pod \"7ed612df-44de-47e2-9939-4947359889f5\" (UID: \"7ed612df-44de-47e2-9939-4947359889f5\") " Dec 07 17:00:53 crc kubenswrapper[4716]: I1207 17:00:53.858150 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7ed612df-44de-47e2-9939-4947359889f5-host" (OuterVolumeSpecName: "host") pod "7ed612df-44de-47e2-9939-4947359889f5" (UID: "7ed612df-44de-47e2-9939-4947359889f5"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 17:00:53 crc kubenswrapper[4716]: I1207 17:00:53.858187 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vl5fl\" (UniqueName: \"kubernetes.io/projected/7ed612df-44de-47e2-9939-4947359889f5-kube-api-access-vl5fl\") pod \"7ed612df-44de-47e2-9939-4947359889f5\" (UID: \"7ed612df-44de-47e2-9939-4947359889f5\") " Dec 07 17:00:53 crc kubenswrapper[4716]: I1207 17:00:53.858811 4716 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7ed612df-44de-47e2-9939-4947359889f5-host\") on node \"crc\" DevicePath \"\"" Dec 07 17:00:53 crc kubenswrapper[4716]: I1207 17:00:53.863840 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ed612df-44de-47e2-9939-4947359889f5-kube-api-access-vl5fl" (OuterVolumeSpecName: "kube-api-access-vl5fl") pod "7ed612df-44de-47e2-9939-4947359889f5" (UID: "7ed612df-44de-47e2-9939-4947359889f5"). InnerVolumeSpecName "kube-api-access-vl5fl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 17:00:53 crc kubenswrapper[4716]: I1207 17:00:53.960148 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vl5fl\" (UniqueName: \"kubernetes.io/projected/7ed612df-44de-47e2-9939-4947359889f5-kube-api-access-vl5fl\") on node \"crc\" DevicePath \"\"" Dec 07 17:00:54 crc kubenswrapper[4716]: I1207 17:00:54.583319 4716 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3510e4b86604639db0fe0251e6949d8fc0867a2a71a97b6c514ce5f46f77ff29" Dec 07 17:00:54 crc kubenswrapper[4716]: I1207 17:00:54.583416 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zlz52/crc-debug-wnjwc" Dec 07 17:00:54 crc kubenswrapper[4716]: I1207 17:00:54.940912 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-zlz52/crc-debug-7psng"] Dec 07 17:00:54 crc kubenswrapper[4716]: E1207 17:00:54.941377 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ed612df-44de-47e2-9939-4947359889f5" containerName="container-00" Dec 07 17:00:54 crc kubenswrapper[4716]: I1207 17:00:54.941393 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ed612df-44de-47e2-9939-4947359889f5" containerName="container-00" Dec 07 17:00:54 crc kubenswrapper[4716]: E1207 17:00:54.941432 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="060923f1-6da9-427f-af9f-d6a365600623" containerName="collect-profiles" Dec 07 17:00:54 crc kubenswrapper[4716]: I1207 17:00:54.941440 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="060923f1-6da9-427f-af9f-d6a365600623" containerName="collect-profiles" Dec 07 17:00:54 crc kubenswrapper[4716]: I1207 17:00:54.941682 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="060923f1-6da9-427f-af9f-d6a365600623" containerName="collect-profiles" Dec 07 17:00:54 crc kubenswrapper[4716]: I1207 17:00:54.941720 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ed612df-44de-47e2-9939-4947359889f5" containerName="container-00" Dec 07 17:00:54 crc kubenswrapper[4716]: I1207 17:00:54.942452 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zlz52/crc-debug-7psng" Dec 07 17:00:55 crc kubenswrapper[4716]: I1207 17:00:55.082107 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thnf8\" (UniqueName: \"kubernetes.io/projected/94d1a32b-4c67-47a4-989c-854936a2a6ac-kube-api-access-thnf8\") pod \"crc-debug-7psng\" (UID: \"94d1a32b-4c67-47a4-989c-854936a2a6ac\") " pod="openshift-must-gather-zlz52/crc-debug-7psng" Dec 07 17:00:55 crc kubenswrapper[4716]: I1207 17:00:55.082496 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/94d1a32b-4c67-47a4-989c-854936a2a6ac-host\") pod \"crc-debug-7psng\" (UID: \"94d1a32b-4c67-47a4-989c-854936a2a6ac\") " pod="openshift-must-gather-zlz52/crc-debug-7psng" Dec 07 17:00:55 crc kubenswrapper[4716]: I1207 17:00:55.184204 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/94d1a32b-4c67-47a4-989c-854936a2a6ac-host\") pod \"crc-debug-7psng\" (UID: \"94d1a32b-4c67-47a4-989c-854936a2a6ac\") " pod="openshift-must-gather-zlz52/crc-debug-7psng" Dec 07 17:00:55 crc kubenswrapper[4716]: I1207 17:00:55.184285 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-thnf8\" (UniqueName: \"kubernetes.io/projected/94d1a32b-4c67-47a4-989c-854936a2a6ac-kube-api-access-thnf8\") pod \"crc-debug-7psng\" (UID: \"94d1a32b-4c67-47a4-989c-854936a2a6ac\") " pod="openshift-must-gather-zlz52/crc-debug-7psng" Dec 07 17:00:55 crc kubenswrapper[4716]: I1207 17:00:55.184361 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/94d1a32b-4c67-47a4-989c-854936a2a6ac-host\") pod \"crc-debug-7psng\" (UID: \"94d1a32b-4c67-47a4-989c-854936a2a6ac\") " pod="openshift-must-gather-zlz52/crc-debug-7psng" Dec 07 17:00:55 crc kubenswrapper[4716]: I1207 17:00:55.205329 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-thnf8\" (UniqueName: \"kubernetes.io/projected/94d1a32b-4c67-47a4-989c-854936a2a6ac-kube-api-access-thnf8\") pod \"crc-debug-7psng\" (UID: \"94d1a32b-4c67-47a4-989c-854936a2a6ac\") " pod="openshift-must-gather-zlz52/crc-debug-7psng" Dec 07 17:00:55 crc kubenswrapper[4716]: I1207 17:00:55.275100 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zlz52/crc-debug-7psng" Dec 07 17:00:55 crc kubenswrapper[4716]: W1207 17:00:55.333726 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod94d1a32b_4c67_47a4_989c_854936a2a6ac.slice/crio-954b377aa3b1e6f987d8d1fdd93bbdc33b92c371df9806684a7b0c020c6ecba1 WatchSource:0}: Error finding container 954b377aa3b1e6f987d8d1fdd93bbdc33b92c371df9806684a7b0c020c6ecba1: Status 404 returned error can't find the container with id 954b377aa3b1e6f987d8d1fdd93bbdc33b92c371df9806684a7b0c020c6ecba1 Dec 07 17:00:55 crc kubenswrapper[4716]: I1207 17:00:55.596807 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zlz52/crc-debug-7psng" event={"ID":"94d1a32b-4c67-47a4-989c-854936a2a6ac","Type":"ContainerStarted","Data":"c251dea702de488ff8a2aab279c7651d0476faf38cd2a4eea7fc83db2d05ad2f"} Dec 07 17:00:55 crc kubenswrapper[4716]: I1207 17:00:55.597481 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zlz52/crc-debug-7psng" event={"ID":"94d1a32b-4c67-47a4-989c-854936a2a6ac","Type":"ContainerStarted","Data":"954b377aa3b1e6f987d8d1fdd93bbdc33b92c371df9806684a7b0c020c6ecba1"} Dec 07 17:00:55 crc kubenswrapper[4716]: I1207 17:00:55.618882 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-zlz52/crc-debug-7psng" podStartSLOduration=1.618862607 podStartE2EDuration="1.618862607s" podCreationTimestamp="2025-12-07 17:00:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 17:00:55.618430555 +0000 UTC m=+3518.308715477" watchObservedRunningTime="2025-12-07 17:00:55.618862607 +0000 UTC m=+3518.309147539" Dec 07 17:00:55 crc kubenswrapper[4716]: I1207 17:00:55.683915 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ed612df-44de-47e2-9939-4947359889f5" path="/var/lib/kubelet/pods/7ed612df-44de-47e2-9939-4947359889f5/volumes" Dec 07 17:00:56 crc kubenswrapper[4716]: I1207 17:00:56.635326 4716 generic.go:334] "Generic (PLEG): container finished" podID="94d1a32b-4c67-47a4-989c-854936a2a6ac" containerID="c251dea702de488ff8a2aab279c7651d0476faf38cd2a4eea7fc83db2d05ad2f" exitCode=0 Dec 07 17:00:56 crc kubenswrapper[4716]: I1207 17:00:56.635408 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zlz52/crc-debug-7psng" event={"ID":"94d1a32b-4c67-47a4-989c-854936a2a6ac","Type":"ContainerDied","Data":"c251dea702de488ff8a2aab279c7651d0476faf38cd2a4eea7fc83db2d05ad2f"} Dec 07 17:00:57 crc kubenswrapper[4716]: I1207 17:00:57.805554 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zlz52/crc-debug-7psng" Dec 07 17:00:57 crc kubenswrapper[4716]: I1207 17:00:57.853852 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-zlz52/crc-debug-7psng"] Dec 07 17:00:57 crc kubenswrapper[4716]: I1207 17:00:57.864442 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-zlz52/crc-debug-7psng"] Dec 07 17:00:57 crc kubenswrapper[4716]: I1207 17:00:57.952208 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/94d1a32b-4c67-47a4-989c-854936a2a6ac-host\") pod \"94d1a32b-4c67-47a4-989c-854936a2a6ac\" (UID: \"94d1a32b-4c67-47a4-989c-854936a2a6ac\") " Dec 07 17:00:57 crc kubenswrapper[4716]: I1207 17:00:57.952296 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/94d1a32b-4c67-47a4-989c-854936a2a6ac-host" (OuterVolumeSpecName: "host") pod "94d1a32b-4c67-47a4-989c-854936a2a6ac" (UID: "94d1a32b-4c67-47a4-989c-854936a2a6ac"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 17:00:57 crc kubenswrapper[4716]: I1207 17:00:57.952321 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-thnf8\" (UniqueName: \"kubernetes.io/projected/94d1a32b-4c67-47a4-989c-854936a2a6ac-kube-api-access-thnf8\") pod \"94d1a32b-4c67-47a4-989c-854936a2a6ac\" (UID: \"94d1a32b-4c67-47a4-989c-854936a2a6ac\") " Dec 07 17:00:57 crc kubenswrapper[4716]: I1207 17:00:57.953557 4716 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/94d1a32b-4c67-47a4-989c-854936a2a6ac-host\") on node \"crc\" DevicePath \"\"" Dec 07 17:00:57 crc kubenswrapper[4716]: I1207 17:00:57.961780 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94d1a32b-4c67-47a4-989c-854936a2a6ac-kube-api-access-thnf8" (OuterVolumeSpecName: "kube-api-access-thnf8") pod "94d1a32b-4c67-47a4-989c-854936a2a6ac" (UID: "94d1a32b-4c67-47a4-989c-854936a2a6ac"). InnerVolumeSpecName "kube-api-access-thnf8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 17:00:58 crc kubenswrapper[4716]: I1207 17:00:58.055149 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-thnf8\" (UniqueName: \"kubernetes.io/projected/94d1a32b-4c67-47a4-989c-854936a2a6ac-kube-api-access-thnf8\") on node \"crc\" DevicePath \"\"" Dec 07 17:00:58 crc kubenswrapper[4716]: I1207 17:00:58.655384 4716 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="954b377aa3b1e6f987d8d1fdd93bbdc33b92c371df9806684a7b0c020c6ecba1" Dec 07 17:00:58 crc kubenswrapper[4716]: I1207 17:00:58.655442 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zlz52/crc-debug-7psng" Dec 07 17:00:59 crc kubenswrapper[4716]: I1207 17:00:59.027047 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-zlz52/crc-debug-wns8q"] Dec 07 17:00:59 crc kubenswrapper[4716]: E1207 17:00:59.027484 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94d1a32b-4c67-47a4-989c-854936a2a6ac" containerName="container-00" Dec 07 17:00:59 crc kubenswrapper[4716]: I1207 17:00:59.027496 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="94d1a32b-4c67-47a4-989c-854936a2a6ac" containerName="container-00" Dec 07 17:00:59 crc kubenswrapper[4716]: I1207 17:00:59.027724 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="94d1a32b-4c67-47a4-989c-854936a2a6ac" containerName="container-00" Dec 07 17:00:59 crc kubenswrapper[4716]: I1207 17:00:59.028282 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zlz52/crc-debug-wns8q" Dec 07 17:00:59 crc kubenswrapper[4716]: I1207 17:00:59.176333 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ed852328-7eb0-44be-a1fa-5ce371b4038f-host\") pod \"crc-debug-wns8q\" (UID: \"ed852328-7eb0-44be-a1fa-5ce371b4038f\") " pod="openshift-must-gather-zlz52/crc-debug-wns8q" Dec 07 17:00:59 crc kubenswrapper[4716]: I1207 17:00:59.176788 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bmfhn\" (UniqueName: \"kubernetes.io/projected/ed852328-7eb0-44be-a1fa-5ce371b4038f-kube-api-access-bmfhn\") pod \"crc-debug-wns8q\" (UID: \"ed852328-7eb0-44be-a1fa-5ce371b4038f\") " pod="openshift-must-gather-zlz52/crc-debug-wns8q" Dec 07 17:00:59 crc kubenswrapper[4716]: I1207 17:00:59.278651 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bmfhn\" (UniqueName: \"kubernetes.io/projected/ed852328-7eb0-44be-a1fa-5ce371b4038f-kube-api-access-bmfhn\") pod \"crc-debug-wns8q\" (UID: \"ed852328-7eb0-44be-a1fa-5ce371b4038f\") " pod="openshift-must-gather-zlz52/crc-debug-wns8q" Dec 07 17:00:59 crc kubenswrapper[4716]: I1207 17:00:59.278723 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ed852328-7eb0-44be-a1fa-5ce371b4038f-host\") pod \"crc-debug-wns8q\" (UID: \"ed852328-7eb0-44be-a1fa-5ce371b4038f\") " pod="openshift-must-gather-zlz52/crc-debug-wns8q" Dec 07 17:00:59 crc kubenswrapper[4716]: I1207 17:00:59.278915 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ed852328-7eb0-44be-a1fa-5ce371b4038f-host\") pod \"crc-debug-wns8q\" (UID: \"ed852328-7eb0-44be-a1fa-5ce371b4038f\") " pod="openshift-must-gather-zlz52/crc-debug-wns8q" Dec 07 17:00:59 crc kubenswrapper[4716]: I1207 17:00:59.297478 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bmfhn\" (UniqueName: \"kubernetes.io/projected/ed852328-7eb0-44be-a1fa-5ce371b4038f-kube-api-access-bmfhn\") pod \"crc-debug-wns8q\" (UID: \"ed852328-7eb0-44be-a1fa-5ce371b4038f\") " pod="openshift-must-gather-zlz52/crc-debug-wns8q" Dec 07 17:00:59 crc kubenswrapper[4716]: I1207 17:00:59.350752 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zlz52/crc-debug-wns8q" Dec 07 17:00:59 crc kubenswrapper[4716]: W1207 17:00:59.376484 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poded852328_7eb0_44be_a1fa_5ce371b4038f.slice/crio-eb3aaa30cedacfbe05001d20348f0098c7e2f1d59c6d1ca242fd9b699ddc7a18 WatchSource:0}: Error finding container eb3aaa30cedacfbe05001d20348f0098c7e2f1d59c6d1ca242fd9b699ddc7a18: Status 404 returned error can't find the container with id eb3aaa30cedacfbe05001d20348f0098c7e2f1d59c6d1ca242fd9b699ddc7a18 Dec 07 17:00:59 crc kubenswrapper[4716]: I1207 17:00:59.668375 4716 generic.go:334] "Generic (PLEG): container finished" podID="ed852328-7eb0-44be-a1fa-5ce371b4038f" containerID="1f7db631db955da7ef6e25e50d0d9445ddda47be8939cb501cdab76d51473cfe" exitCode=0 Dec 07 17:00:59 crc kubenswrapper[4716]: I1207 17:00:59.669779 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94d1a32b-4c67-47a4-989c-854936a2a6ac" path="/var/lib/kubelet/pods/94d1a32b-4c67-47a4-989c-854936a2a6ac/volumes" Dec 07 17:00:59 crc kubenswrapper[4716]: I1207 17:00:59.670674 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zlz52/crc-debug-wns8q" event={"ID":"ed852328-7eb0-44be-a1fa-5ce371b4038f","Type":"ContainerDied","Data":"1f7db631db955da7ef6e25e50d0d9445ddda47be8939cb501cdab76d51473cfe"} Dec 07 17:00:59 crc kubenswrapper[4716]: I1207 17:00:59.670712 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zlz52/crc-debug-wns8q" event={"ID":"ed852328-7eb0-44be-a1fa-5ce371b4038f","Type":"ContainerStarted","Data":"eb3aaa30cedacfbe05001d20348f0098c7e2f1d59c6d1ca242fd9b699ddc7a18"} Dec 07 17:00:59 crc kubenswrapper[4716]: I1207 17:00:59.735054 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-zlz52/crc-debug-wns8q"] Dec 07 17:00:59 crc kubenswrapper[4716]: I1207 17:00:59.745639 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-zlz52/crc-debug-wns8q"] Dec 07 17:01:00 crc kubenswrapper[4716]: I1207 17:01:00.191353 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29418781-qddct"] Dec 07 17:01:00 crc kubenswrapper[4716]: E1207 17:01:00.192147 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed852328-7eb0-44be-a1fa-5ce371b4038f" containerName="container-00" Dec 07 17:01:00 crc kubenswrapper[4716]: I1207 17:01:00.192167 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed852328-7eb0-44be-a1fa-5ce371b4038f" containerName="container-00" Dec 07 17:01:00 crc kubenswrapper[4716]: I1207 17:01:00.192457 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed852328-7eb0-44be-a1fa-5ce371b4038f" containerName="container-00" Dec 07 17:01:00 crc kubenswrapper[4716]: I1207 17:01:00.193259 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29418781-qddct" Dec 07 17:01:00 crc kubenswrapper[4716]: I1207 17:01:00.199554 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29418781-qddct"] Dec 07 17:01:00 crc kubenswrapper[4716]: I1207 17:01:00.328681 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c7bbcbc-7ce9-419d-b63d-8bcc377e4264-config-data\") pod \"keystone-cron-29418781-qddct\" (UID: \"1c7bbcbc-7ce9-419d-b63d-8bcc377e4264\") " pod="openstack/keystone-cron-29418781-qddct" Dec 07 17:01:00 crc kubenswrapper[4716]: I1207 17:01:00.328766 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c7bbcbc-7ce9-419d-b63d-8bcc377e4264-combined-ca-bundle\") pod \"keystone-cron-29418781-qddct\" (UID: \"1c7bbcbc-7ce9-419d-b63d-8bcc377e4264\") " pod="openstack/keystone-cron-29418781-qddct" Dec 07 17:01:00 crc kubenswrapper[4716]: I1207 17:01:00.328799 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n9989\" (UniqueName: \"kubernetes.io/projected/1c7bbcbc-7ce9-419d-b63d-8bcc377e4264-kube-api-access-n9989\") pod \"keystone-cron-29418781-qddct\" (UID: \"1c7bbcbc-7ce9-419d-b63d-8bcc377e4264\") " pod="openstack/keystone-cron-29418781-qddct" Dec 07 17:01:00 crc kubenswrapper[4716]: I1207 17:01:00.328820 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1c7bbcbc-7ce9-419d-b63d-8bcc377e4264-fernet-keys\") pod \"keystone-cron-29418781-qddct\" (UID: \"1c7bbcbc-7ce9-419d-b63d-8bcc377e4264\") " pod="openstack/keystone-cron-29418781-qddct" Dec 07 17:01:00 crc kubenswrapper[4716]: I1207 17:01:00.430692 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c7bbcbc-7ce9-419d-b63d-8bcc377e4264-config-data\") pod \"keystone-cron-29418781-qddct\" (UID: \"1c7bbcbc-7ce9-419d-b63d-8bcc377e4264\") " pod="openstack/keystone-cron-29418781-qddct" Dec 07 17:01:00 crc kubenswrapper[4716]: I1207 17:01:00.430801 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c7bbcbc-7ce9-419d-b63d-8bcc377e4264-combined-ca-bundle\") pod \"keystone-cron-29418781-qddct\" (UID: \"1c7bbcbc-7ce9-419d-b63d-8bcc377e4264\") " pod="openstack/keystone-cron-29418781-qddct" Dec 07 17:01:00 crc kubenswrapper[4716]: I1207 17:01:00.430847 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n9989\" (UniqueName: \"kubernetes.io/projected/1c7bbcbc-7ce9-419d-b63d-8bcc377e4264-kube-api-access-n9989\") pod \"keystone-cron-29418781-qddct\" (UID: \"1c7bbcbc-7ce9-419d-b63d-8bcc377e4264\") " pod="openstack/keystone-cron-29418781-qddct" Dec 07 17:01:00 crc kubenswrapper[4716]: I1207 17:01:00.430890 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1c7bbcbc-7ce9-419d-b63d-8bcc377e4264-fernet-keys\") pod \"keystone-cron-29418781-qddct\" (UID: \"1c7bbcbc-7ce9-419d-b63d-8bcc377e4264\") " pod="openstack/keystone-cron-29418781-qddct" Dec 07 17:01:00 crc kubenswrapper[4716]: I1207 17:01:00.438114 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1c7bbcbc-7ce9-419d-b63d-8bcc377e4264-fernet-keys\") pod \"keystone-cron-29418781-qddct\" (UID: \"1c7bbcbc-7ce9-419d-b63d-8bcc377e4264\") " pod="openstack/keystone-cron-29418781-qddct" Dec 07 17:01:00 crc kubenswrapper[4716]: I1207 17:01:00.438505 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c7bbcbc-7ce9-419d-b63d-8bcc377e4264-config-data\") pod \"keystone-cron-29418781-qddct\" (UID: \"1c7bbcbc-7ce9-419d-b63d-8bcc377e4264\") " pod="openstack/keystone-cron-29418781-qddct" Dec 07 17:01:00 crc kubenswrapper[4716]: I1207 17:01:00.442328 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c7bbcbc-7ce9-419d-b63d-8bcc377e4264-combined-ca-bundle\") pod \"keystone-cron-29418781-qddct\" (UID: \"1c7bbcbc-7ce9-419d-b63d-8bcc377e4264\") " pod="openstack/keystone-cron-29418781-qddct" Dec 07 17:01:00 crc kubenswrapper[4716]: I1207 17:01:00.454856 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n9989\" (UniqueName: \"kubernetes.io/projected/1c7bbcbc-7ce9-419d-b63d-8bcc377e4264-kube-api-access-n9989\") pod \"keystone-cron-29418781-qddct\" (UID: \"1c7bbcbc-7ce9-419d-b63d-8bcc377e4264\") " pod="openstack/keystone-cron-29418781-qddct" Dec 07 17:01:00 crc kubenswrapper[4716]: I1207 17:01:00.547642 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29418781-qddct" Dec 07 17:01:00 crc kubenswrapper[4716]: I1207 17:01:00.749233 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zlz52/crc-debug-wns8q" Dec 07 17:01:00 crc kubenswrapper[4716]: I1207 17:01:00.837026 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ed852328-7eb0-44be-a1fa-5ce371b4038f-host\") pod \"ed852328-7eb0-44be-a1fa-5ce371b4038f\" (UID: \"ed852328-7eb0-44be-a1fa-5ce371b4038f\") " Dec 07 17:01:00 crc kubenswrapper[4716]: I1207 17:01:00.837223 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bmfhn\" (UniqueName: \"kubernetes.io/projected/ed852328-7eb0-44be-a1fa-5ce371b4038f-kube-api-access-bmfhn\") pod \"ed852328-7eb0-44be-a1fa-5ce371b4038f\" (UID: \"ed852328-7eb0-44be-a1fa-5ce371b4038f\") " Dec 07 17:01:00 crc kubenswrapper[4716]: I1207 17:01:00.837593 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ed852328-7eb0-44be-a1fa-5ce371b4038f-host" (OuterVolumeSpecName: "host") pod "ed852328-7eb0-44be-a1fa-5ce371b4038f" (UID: "ed852328-7eb0-44be-a1fa-5ce371b4038f"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 17:01:00 crc kubenswrapper[4716]: I1207 17:01:00.837824 4716 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ed852328-7eb0-44be-a1fa-5ce371b4038f-host\") on node \"crc\" DevicePath \"\"" Dec 07 17:01:00 crc kubenswrapper[4716]: I1207 17:01:00.850323 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed852328-7eb0-44be-a1fa-5ce371b4038f-kube-api-access-bmfhn" (OuterVolumeSpecName: "kube-api-access-bmfhn") pod "ed852328-7eb0-44be-a1fa-5ce371b4038f" (UID: "ed852328-7eb0-44be-a1fa-5ce371b4038f"). InnerVolumeSpecName "kube-api-access-bmfhn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 17:01:00 crc kubenswrapper[4716]: I1207 17:01:00.939290 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bmfhn\" (UniqueName: \"kubernetes.io/projected/ed852328-7eb0-44be-a1fa-5ce371b4038f-kube-api-access-bmfhn\") on node \"crc\" DevicePath \"\"" Dec 07 17:01:01 crc kubenswrapper[4716]: I1207 17:01:01.025739 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29418781-qddct"] Dec 07 17:01:01 crc kubenswrapper[4716]: I1207 17:01:01.668237 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed852328-7eb0-44be-a1fa-5ce371b4038f" path="/var/lib/kubelet/pods/ed852328-7eb0-44be-a1fa-5ce371b4038f/volumes" Dec 07 17:01:01 crc kubenswrapper[4716]: I1207 17:01:01.695883 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29418781-qddct" event={"ID":"1c7bbcbc-7ce9-419d-b63d-8bcc377e4264","Type":"ContainerStarted","Data":"cebccbca4ab002c2dfccec76b84aa23e92232f51844ef2cedab7e4a2978781a1"} Dec 07 17:01:01 crc kubenswrapper[4716]: I1207 17:01:01.695923 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29418781-qddct" event={"ID":"1c7bbcbc-7ce9-419d-b63d-8bcc377e4264","Type":"ContainerStarted","Data":"c6f7ad02c1f19971caf8bb57b393d5c76f52036c452481177174a52ba27ed653"} Dec 07 17:01:01 crc kubenswrapper[4716]: I1207 17:01:01.697805 4716 scope.go:117] "RemoveContainer" containerID="1f7db631db955da7ef6e25e50d0d9445ddda47be8939cb501cdab76d51473cfe" Dec 07 17:01:01 crc kubenswrapper[4716]: I1207 17:01:01.697838 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zlz52/crc-debug-wns8q" Dec 07 17:01:01 crc kubenswrapper[4716]: I1207 17:01:01.717257 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29418781-qddct" podStartSLOduration=1.7172370099999998 podStartE2EDuration="1.71723701s" podCreationTimestamp="2025-12-07 17:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 17:01:01.711275348 +0000 UTC m=+3524.401560260" watchObservedRunningTime="2025-12-07 17:01:01.71723701 +0000 UTC m=+3524.407521922" Dec 07 17:01:03 crc kubenswrapper[4716]: I1207 17:01:03.717033 4716 generic.go:334] "Generic (PLEG): container finished" podID="1c7bbcbc-7ce9-419d-b63d-8bcc377e4264" containerID="cebccbca4ab002c2dfccec76b84aa23e92232f51844ef2cedab7e4a2978781a1" exitCode=0 Dec 07 17:01:03 crc kubenswrapper[4716]: I1207 17:01:03.717253 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29418781-qddct" event={"ID":"1c7bbcbc-7ce9-419d-b63d-8bcc377e4264","Type":"ContainerDied","Data":"cebccbca4ab002c2dfccec76b84aa23e92232f51844ef2cedab7e4a2978781a1"} Dec 07 17:01:05 crc kubenswrapper[4716]: I1207 17:01:05.048659 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29418781-qddct" Dec 07 17:01:05 crc kubenswrapper[4716]: I1207 17:01:05.212340 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n9989\" (UniqueName: \"kubernetes.io/projected/1c7bbcbc-7ce9-419d-b63d-8bcc377e4264-kube-api-access-n9989\") pod \"1c7bbcbc-7ce9-419d-b63d-8bcc377e4264\" (UID: \"1c7bbcbc-7ce9-419d-b63d-8bcc377e4264\") " Dec 07 17:01:05 crc kubenswrapper[4716]: I1207 17:01:05.212445 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c7bbcbc-7ce9-419d-b63d-8bcc377e4264-combined-ca-bundle\") pod \"1c7bbcbc-7ce9-419d-b63d-8bcc377e4264\" (UID: \"1c7bbcbc-7ce9-419d-b63d-8bcc377e4264\") " Dec 07 17:01:05 crc kubenswrapper[4716]: I1207 17:01:05.212646 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1c7bbcbc-7ce9-419d-b63d-8bcc377e4264-fernet-keys\") pod \"1c7bbcbc-7ce9-419d-b63d-8bcc377e4264\" (UID: \"1c7bbcbc-7ce9-419d-b63d-8bcc377e4264\") " Dec 07 17:01:05 crc kubenswrapper[4716]: I1207 17:01:05.212671 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c7bbcbc-7ce9-419d-b63d-8bcc377e4264-config-data\") pod \"1c7bbcbc-7ce9-419d-b63d-8bcc377e4264\" (UID: \"1c7bbcbc-7ce9-419d-b63d-8bcc377e4264\") " Dec 07 17:01:05 crc kubenswrapper[4716]: I1207 17:01:05.218755 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1c7bbcbc-7ce9-419d-b63d-8bcc377e4264-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "1c7bbcbc-7ce9-419d-b63d-8bcc377e4264" (UID: "1c7bbcbc-7ce9-419d-b63d-8bcc377e4264"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 17:01:05 crc kubenswrapper[4716]: I1207 17:01:05.220776 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c7bbcbc-7ce9-419d-b63d-8bcc377e4264-kube-api-access-n9989" (OuterVolumeSpecName: "kube-api-access-n9989") pod "1c7bbcbc-7ce9-419d-b63d-8bcc377e4264" (UID: "1c7bbcbc-7ce9-419d-b63d-8bcc377e4264"). InnerVolumeSpecName "kube-api-access-n9989". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 17:01:05 crc kubenswrapper[4716]: I1207 17:01:05.255704 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1c7bbcbc-7ce9-419d-b63d-8bcc377e4264-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1c7bbcbc-7ce9-419d-b63d-8bcc377e4264" (UID: "1c7bbcbc-7ce9-419d-b63d-8bcc377e4264"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 17:01:05 crc kubenswrapper[4716]: I1207 17:01:05.271578 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1c7bbcbc-7ce9-419d-b63d-8bcc377e4264-config-data" (OuterVolumeSpecName: "config-data") pod "1c7bbcbc-7ce9-419d-b63d-8bcc377e4264" (UID: "1c7bbcbc-7ce9-419d-b63d-8bcc377e4264"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 17:01:05 crc kubenswrapper[4716]: I1207 17:01:05.315247 4716 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c7bbcbc-7ce9-419d-b63d-8bcc377e4264-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 07 17:01:05 crc kubenswrapper[4716]: I1207 17:01:05.315318 4716 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1c7bbcbc-7ce9-419d-b63d-8bcc377e4264-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 07 17:01:05 crc kubenswrapper[4716]: I1207 17:01:05.315343 4716 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c7bbcbc-7ce9-419d-b63d-8bcc377e4264-config-data\") on node \"crc\" DevicePath \"\"" Dec 07 17:01:05 crc kubenswrapper[4716]: I1207 17:01:05.315365 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n9989\" (UniqueName: \"kubernetes.io/projected/1c7bbcbc-7ce9-419d-b63d-8bcc377e4264-kube-api-access-n9989\") on node \"crc\" DevicePath \"\"" Dec 07 17:01:05 crc kubenswrapper[4716]: I1207 17:01:05.736086 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29418781-qddct" event={"ID":"1c7bbcbc-7ce9-419d-b63d-8bcc377e4264","Type":"ContainerDied","Data":"c6f7ad02c1f19971caf8bb57b393d5c76f52036c452481177174a52ba27ed653"} Dec 07 17:01:05 crc kubenswrapper[4716]: I1207 17:01:05.736119 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29418781-qddct" Dec 07 17:01:05 crc kubenswrapper[4716]: I1207 17:01:05.736127 4716 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c6f7ad02c1f19971caf8bb57b393d5c76f52036c452481177174a52ba27ed653" Dec 07 17:01:15 crc kubenswrapper[4716]: I1207 17:01:15.758235 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-64c898bd6-wgw8m_19efaa0a-3cda-4ef8-a845-bbcc78c747ec/barbican-api/0.log" Dec 07 17:01:15 crc kubenswrapper[4716]: I1207 17:01:15.765242 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-64c898bd6-wgw8m_19efaa0a-3cda-4ef8-a845-bbcc78c747ec/barbican-api-log/0.log" Dec 07 17:01:15 crc kubenswrapper[4716]: I1207 17:01:15.895581 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-75c8ccd7fb-bhzds_214231d4-11de-4827-afd3-e5169b138b7b/barbican-keystone-listener/0.log" Dec 07 17:01:15 crc kubenswrapper[4716]: I1207 17:01:15.968201 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-75c8ccd7fb-bhzds_214231d4-11de-4827-afd3-e5169b138b7b/barbican-keystone-listener-log/0.log" Dec 07 17:01:16 crc kubenswrapper[4716]: I1207 17:01:16.020618 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-587c8bb479-4d8m7_cc929236-8385-4f05-8ca5-37315e852be6/barbican-worker/0.log" Dec 07 17:01:16 crc kubenswrapper[4716]: I1207 17:01:16.095412 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-587c8bb479-4d8m7_cc929236-8385-4f05-8ca5-37315e852be6/barbican-worker-log/0.log" Dec 07 17:01:16 crc kubenswrapper[4716]: I1207 17:01:16.209469 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-bvvcz_1c545d8b-e4c4-411c-b122-68f22c6befa4/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 07 17:01:16 crc kubenswrapper[4716]: I1207 17:01:16.355763 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_5f86be30-3a49-4354-a6e9-059592ee5ebc/ceilometer-central-agent/0.log" Dec 07 17:01:16 crc kubenswrapper[4716]: I1207 17:01:16.428707 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_5f86be30-3a49-4354-a6e9-059592ee5ebc/ceilometer-notification-agent/0.log" Dec 07 17:01:16 crc kubenswrapper[4716]: I1207 17:01:16.456553 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_5f86be30-3a49-4354-a6e9-059592ee5ebc/proxy-httpd/0.log" Dec 07 17:01:16 crc kubenswrapper[4716]: I1207 17:01:16.498702 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_5f86be30-3a49-4354-a6e9-059592ee5ebc/sg-core/0.log" Dec 07 17:01:16 crc kubenswrapper[4716]: I1207 17:01:16.649628 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_9237e175-c046-4f00-9535-474448e79076/cinder-api/0.log" Dec 07 17:01:16 crc kubenswrapper[4716]: I1207 17:01:16.667534 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_9237e175-c046-4f00-9535-474448e79076/cinder-api-log/0.log" Dec 07 17:01:16 crc kubenswrapper[4716]: I1207 17:01:16.863178 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_6db216a8-54ce-4596-a4ab-bb24a787b027/cinder-scheduler/0.log" Dec 07 17:01:16 crc kubenswrapper[4716]: I1207 17:01:16.872585 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_6db216a8-54ce-4596-a4ab-bb24a787b027/probe/0.log" Dec 07 17:01:16 crc kubenswrapper[4716]: I1207 17:01:16.957845 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-zsnd4_d2fcfd64-6b34-4f9c-9e8b-3715df35f92d/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 07 17:01:17 crc kubenswrapper[4716]: I1207 17:01:17.150427 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-kcj2p_9d00d0ca-8b9a-455b-a7a5-f3d17a6e363a/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 07 17:01:17 crc kubenswrapper[4716]: I1207 17:01:17.344438 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-sw78r_27293e5d-5955-4ccb-b78b-63433622f073/init/0.log" Dec 07 17:01:17 crc kubenswrapper[4716]: I1207 17:01:17.539654 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-sw78r_27293e5d-5955-4ccb-b78b-63433622f073/init/0.log" Dec 07 17:01:17 crc kubenswrapper[4716]: I1207 17:01:17.549437 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-sw78r_27293e5d-5955-4ccb-b78b-63433622f073/dnsmasq-dns/0.log" Dec 07 17:01:17 crc kubenswrapper[4716]: I1207 17:01:17.644042 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-9d8h7_7495cf42-56aa-43c9-9d24-a022a3e50505/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Dec 07 17:01:17 crc kubenswrapper[4716]: I1207 17:01:17.812177 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_9f4f8c7b-38d4-4054-bb66-c51f32448567/glance-httpd/0.log" Dec 07 17:01:17 crc kubenswrapper[4716]: I1207 17:01:17.815887 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_9f4f8c7b-38d4-4054-bb66-c51f32448567/glance-log/0.log" Dec 07 17:01:17 crc kubenswrapper[4716]: I1207 17:01:17.991382 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_0f287d7e-4b08-4c16-8a23-cf9e69225d3c/glance-httpd/0.log" Dec 07 17:01:18 crc kubenswrapper[4716]: I1207 17:01:18.179438 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_0f287d7e-4b08-4c16-8a23-cf9e69225d3c/glance-log/0.log" Dec 07 17:01:18 crc kubenswrapper[4716]: I1207 17:01:18.403009 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-52qx7_6dd92737-9433-4fe4-ab78-03ebaeb31a24/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 07 17:01:18 crc kubenswrapper[4716]: I1207 17:01:18.448070 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7f9b78866d-9g9kv_d845ae0b-5d45-4021-a1e9-4b124298b65b/horizon/0.log" Dec 07 17:01:18 crc kubenswrapper[4716]: I1207 17:01:18.619059 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7f9b78866d-9g9kv_d845ae0b-5d45-4021-a1e9-4b124298b65b/horizon-log/0.log" Dec 07 17:01:18 crc kubenswrapper[4716]: I1207 17:01:18.696569 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-bd27b_9e126137-6235-4a75-8d46-05bb3cf60acc/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 07 17:01:18 crc kubenswrapper[4716]: I1207 17:01:18.916724 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29418781-qddct_1c7bbcbc-7ce9-419d-b63d-8bcc377e4264/keystone-cron/0.log" Dec 07 17:01:18 crc kubenswrapper[4716]: I1207 17:01:18.968363 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-6c4445878c-bnpkh_ade9101f-5435-455f-807c-d277918cbb46/keystone-api/0.log" Dec 07 17:01:19 crc kubenswrapper[4716]: I1207 17:01:19.075400 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_7953b7b7-a0c6-452c-868a-dcc4a1bda1fa/kube-state-metrics/0.log" Dec 07 17:01:19 crc kubenswrapper[4716]: I1207 17:01:19.227314 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-66mjc_0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 07 17:01:19 crc kubenswrapper[4716]: I1207 17:01:19.509562 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-bd7b79585-wwp6b_dd279760-9178-4ad8-ae1e-ae4e6fab3f3c/neutron-httpd/0.log" Dec 07 17:01:19 crc kubenswrapper[4716]: I1207 17:01:19.518980 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-bd7b79585-wwp6b_dd279760-9178-4ad8-ae1e-ae4e6fab3f3c/neutron-api/0.log" Dec 07 17:01:19 crc kubenswrapper[4716]: I1207 17:01:19.560810 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-n265m_b4f56131-2845-43e1-84ca-52db74bf2b08/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 07 17:01:20 crc kubenswrapper[4716]: I1207 17:01:20.069167 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_b3643cef-3c95-4ecc-9225-ebf929947127/nova-cell0-conductor-conductor/0.log" Dec 07 17:01:20 crc kubenswrapper[4716]: I1207 17:01:20.069423 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_3953832e-aa6d-49f2-980d-e2308f0d1ece/nova-api-log/0.log" Dec 07 17:01:20 crc kubenswrapper[4716]: I1207 17:01:20.328863 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_3953832e-aa6d-49f2-980d-e2308f0d1ece/nova-api-api/0.log" Dec 07 17:01:20 crc kubenswrapper[4716]: I1207 17:01:20.378936 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_098f28a4-4852-4c3a-ada9-7b8b9272c3ae/nova-cell1-conductor-conductor/0.log" Dec 07 17:01:20 crc kubenswrapper[4716]: I1207 17:01:20.523863 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_c0c9c313-94be-404d-849d-3eb23c5eb5e5/nova-cell1-novncproxy-novncproxy/0.log" Dec 07 17:01:20 crc kubenswrapper[4716]: I1207 17:01:20.724280 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-8jm5n_01472d87-913d-4565-8d83-40966b88a630/nova-edpm-deployment-openstack-edpm-ipam/0.log" Dec 07 17:01:20 crc kubenswrapper[4716]: I1207 17:01:20.845984 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_2ac894fd-e293-4f28-bdb5-e80e01601903/nova-metadata-log/0.log" Dec 07 17:01:21 crc kubenswrapper[4716]: I1207 17:01:21.147425 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_794a9d02-75dc-4ad6-bdb1-51bccaab7c9a/nova-scheduler-scheduler/0.log" Dec 07 17:01:21 crc kubenswrapper[4716]: I1207 17:01:21.171566 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_f5bcaa65-381c-40e7-97bd-4fc0eafc572a/mysql-bootstrap/0.log" Dec 07 17:01:21 crc kubenswrapper[4716]: I1207 17:01:21.324862 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_f5bcaa65-381c-40e7-97bd-4fc0eafc572a/galera/0.log" Dec 07 17:01:21 crc kubenswrapper[4716]: I1207 17:01:21.335800 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_f5bcaa65-381c-40e7-97bd-4fc0eafc572a/mysql-bootstrap/0.log" Dec 07 17:01:21 crc kubenswrapper[4716]: I1207 17:01:21.536960 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_c450ef40-3a23-4bab-9d64-4475f12d0490/mysql-bootstrap/0.log" Dec 07 17:01:21 crc kubenswrapper[4716]: I1207 17:01:21.771255 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_c450ef40-3a23-4bab-9d64-4475f12d0490/galera/0.log" Dec 07 17:01:21 crc kubenswrapper[4716]: I1207 17:01:21.825037 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_c450ef40-3a23-4bab-9d64-4475f12d0490/mysql-bootstrap/0.log" Dec 07 17:01:21 crc kubenswrapper[4716]: I1207 17:01:21.961635 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_6a6ecf77-dd60-497a-8670-919d23db95dd/openstackclient/0.log" Dec 07 17:01:21 crc kubenswrapper[4716]: I1207 17:01:21.975025 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_2ac894fd-e293-4f28-bdb5-e80e01601903/nova-metadata-metadata/0.log" Dec 07 17:01:22 crc kubenswrapper[4716]: I1207 17:01:22.094032 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-8r8gk_1fcf6527-b54d-4239-a4c4-a9cd6fe35bd2/openstack-network-exporter/0.log" Dec 07 17:01:22 crc kubenswrapper[4716]: I1207 17:01:22.251501 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-g6nwl_0fc0f0b8-b60a-4aff-84dd-bbec37b12d81/ovsdb-server-init/0.log" Dec 07 17:01:22 crc kubenswrapper[4716]: I1207 17:01:22.439208 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-g6nwl_0fc0f0b8-b60a-4aff-84dd-bbec37b12d81/ovs-vswitchd/0.log" Dec 07 17:01:22 crc kubenswrapper[4716]: I1207 17:01:22.447573 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-g6nwl_0fc0f0b8-b60a-4aff-84dd-bbec37b12d81/ovsdb-server-init/0.log" Dec 07 17:01:22 crc kubenswrapper[4716]: I1207 17:01:22.483776 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-g6nwl_0fc0f0b8-b60a-4aff-84dd-bbec37b12d81/ovsdb-server/0.log" Dec 07 17:01:22 crc kubenswrapper[4716]: I1207 17:01:22.683431 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-tgl27_399984bf-94af-4630-a7c3-9375a388159a/ovn-controller/0.log" Dec 07 17:01:22 crc kubenswrapper[4716]: I1207 17:01:22.777816 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-26hn5_b9e8242a-2734-47b9-9f14-7eea8d9ad3c4/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 07 17:01:22 crc kubenswrapper[4716]: I1207 17:01:22.871932 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_3ae6bded-8084-4e07-afcc-a7fafb718e3a/openstack-network-exporter/0.log" Dec 07 17:01:22 crc kubenswrapper[4716]: I1207 17:01:22.916635 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_3ae6bded-8084-4e07-afcc-a7fafb718e3a/ovn-northd/0.log" Dec 07 17:01:23 crc kubenswrapper[4716]: I1207 17:01:23.040365 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_771a3c3e-34e4-43f6-a3e2-8e7156e081d1/ovsdbserver-nb/0.log" Dec 07 17:01:23 crc kubenswrapper[4716]: I1207 17:01:23.047001 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_771a3c3e-34e4-43f6-a3e2-8e7156e081d1/openstack-network-exporter/0.log" Dec 07 17:01:23 crc kubenswrapper[4716]: I1207 17:01:23.199457 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_5c11a9b9-34de-42f7-9399-3cddce1af7c9/openstack-network-exporter/0.log" Dec 07 17:01:23 crc kubenswrapper[4716]: I1207 17:01:23.273171 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_5c11a9b9-34de-42f7-9399-3cddce1af7c9/ovsdbserver-sb/0.log" Dec 07 17:01:23 crc kubenswrapper[4716]: I1207 17:01:23.342922 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-746cf47744-w9nm7_0320eb6a-f473-4c4c-ae36-f74080dcdaa5/placement-api/0.log" Dec 07 17:01:23 crc kubenswrapper[4716]: I1207 17:01:23.668689 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-746cf47744-w9nm7_0320eb6a-f473-4c4c-ae36-f74080dcdaa5/placement-log/0.log" Dec 07 17:01:23 crc kubenswrapper[4716]: I1207 17:01:23.673258 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_bb3eb7dd-3210-452a-adf4-92bb2eb2447f/setup-container/0.log" Dec 07 17:01:23 crc kubenswrapper[4716]: I1207 17:01:23.906221 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_bb3eb7dd-3210-452a-adf4-92bb2eb2447f/setup-container/0.log" Dec 07 17:01:23 crc kubenswrapper[4716]: I1207 17:01:23.933452 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_db1d1fd3-b1ed-472b-b615-b866706b28a9/setup-container/0.log" Dec 07 17:01:23 crc kubenswrapper[4716]: I1207 17:01:23.947002 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_bb3eb7dd-3210-452a-adf4-92bb2eb2447f/rabbitmq/0.log" Dec 07 17:01:24 crc kubenswrapper[4716]: I1207 17:01:24.161734 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_db1d1fd3-b1ed-472b-b615-b866706b28a9/setup-container/0.log" Dec 07 17:01:24 crc kubenswrapper[4716]: I1207 17:01:24.200924 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_db1d1fd3-b1ed-472b-b615-b866706b28a9/rabbitmq/0.log" Dec 07 17:01:24 crc kubenswrapper[4716]: I1207 17:01:24.217967 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-xsl26_3a9def06-fe00-45a7-9599-2fada793ba25/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 07 17:01:24 crc kubenswrapper[4716]: I1207 17:01:24.381364 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-l5phd_2d823bb1-cb32-4c09-af7c-6ca81d0277bf/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Dec 07 17:01:24 crc kubenswrapper[4716]: I1207 17:01:24.466858 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-dv7q9_2c9bedba-2a2e-4e68-839a-646d5747911c/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 07 17:01:24 crc kubenswrapper[4716]: I1207 17:01:24.630581 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-9c27f_0d4b7b56-033c-41be-b46f-e4090775ab5c/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 07 17:01:24 crc kubenswrapper[4716]: I1207 17:01:24.658434 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-vpm5k_92420ce0-b95e-4857-b6f3-f6d4519c3d86/ssh-known-hosts-edpm-deployment/0.log" Dec 07 17:01:24 crc kubenswrapper[4716]: I1207 17:01:24.852913 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-5488d7fd99-pwmf5_826c0644-3046-453f-a139-5bd9c95216d0/proxy-server/0.log" Dec 07 17:01:24 crc kubenswrapper[4716]: I1207 17:01:24.958334 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-5488d7fd99-pwmf5_826c0644-3046-453f-a139-5bd9c95216d0/proxy-httpd/0.log" Dec 07 17:01:25 crc kubenswrapper[4716]: I1207 17:01:25.114650 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a714bbaa-9626-4b49-92ae-2abb029408d3/account-auditor/0.log" Dec 07 17:01:25 crc kubenswrapper[4716]: I1207 17:01:25.118661 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-jgrbf_2c371669-443a-492e-ad04-ab79ae978e5f/swift-ring-rebalance/0.log" Dec 07 17:01:25 crc kubenswrapper[4716]: I1207 17:01:25.202996 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a714bbaa-9626-4b49-92ae-2abb029408d3/account-reaper/0.log" Dec 07 17:01:25 crc kubenswrapper[4716]: I1207 17:01:25.290231 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a714bbaa-9626-4b49-92ae-2abb029408d3/account-server/0.log" Dec 07 17:01:25 crc kubenswrapper[4716]: I1207 17:01:25.365105 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a714bbaa-9626-4b49-92ae-2abb029408d3/account-replicator/0.log" Dec 07 17:01:25 crc kubenswrapper[4716]: I1207 17:01:25.383791 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a714bbaa-9626-4b49-92ae-2abb029408d3/container-auditor/0.log" Dec 07 17:01:25 crc kubenswrapper[4716]: I1207 17:01:25.438884 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a714bbaa-9626-4b49-92ae-2abb029408d3/container-replicator/0.log" Dec 07 17:01:25 crc kubenswrapper[4716]: I1207 17:01:25.479437 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a714bbaa-9626-4b49-92ae-2abb029408d3/container-server/0.log" Dec 07 17:01:25 crc kubenswrapper[4716]: I1207 17:01:25.591787 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a714bbaa-9626-4b49-92ae-2abb029408d3/container-updater/0.log" Dec 07 17:01:25 crc kubenswrapper[4716]: I1207 17:01:25.633439 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a714bbaa-9626-4b49-92ae-2abb029408d3/object-expirer/0.log" Dec 07 17:01:25 crc kubenswrapper[4716]: I1207 17:01:25.671432 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a714bbaa-9626-4b49-92ae-2abb029408d3/object-auditor/0.log" Dec 07 17:01:25 crc kubenswrapper[4716]: I1207 17:01:25.714185 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a714bbaa-9626-4b49-92ae-2abb029408d3/object-replicator/0.log" Dec 07 17:01:25 crc kubenswrapper[4716]: I1207 17:01:25.801471 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a714bbaa-9626-4b49-92ae-2abb029408d3/object-server/0.log" Dec 07 17:01:25 crc kubenswrapper[4716]: I1207 17:01:25.837503 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a714bbaa-9626-4b49-92ae-2abb029408d3/object-updater/0.log" Dec 07 17:01:25 crc kubenswrapper[4716]: I1207 17:01:25.868142 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a714bbaa-9626-4b49-92ae-2abb029408d3/rsync/0.log" Dec 07 17:01:25 crc kubenswrapper[4716]: I1207 17:01:25.928720 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a714bbaa-9626-4b49-92ae-2abb029408d3/swift-recon-cron/0.log" Dec 07 17:01:26 crc kubenswrapper[4716]: I1207 17:01:26.049954 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-69qz6_c9735df5-27ef-41b1-84c1-4cab55f23b3a/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 07 17:01:26 crc kubenswrapper[4716]: I1207 17:01:26.149883 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_7c4bb2e7-29d0-4548-863b-ea89d11d68a7/tempest-tests-tempest-tests-runner/0.log" Dec 07 17:01:26 crc kubenswrapper[4716]: I1207 17:01:26.294923 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_fa15bae7-7145-486c-8f55-53797c60c01c/test-operator-logs-container/0.log" Dec 07 17:01:26 crc kubenswrapper[4716]: I1207 17:01:26.515005 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-kwl9v_fa50585d-9042-4f6e-b926-fdff8344e0ad/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 07 17:01:32 crc kubenswrapper[4716]: I1207 17:01:32.728968 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_0f4b5de9-9760-4930-bc8e-d7e47a13ad2e/memcached/0.log" Dec 07 17:01:51 crc kubenswrapper[4716]: I1207 17:01:51.146456 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4cb94d7f59637b410c6ea06aeac57a499ffab903391b8415798aca23986vcrp_5bdf0591-fc57-4545-8ab4-e89de972df50/util/0.log" Dec 07 17:01:51 crc kubenswrapper[4716]: I1207 17:01:51.318129 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4cb94d7f59637b410c6ea06aeac57a499ffab903391b8415798aca23986vcrp_5bdf0591-fc57-4545-8ab4-e89de972df50/util/0.log" Dec 07 17:01:51 crc kubenswrapper[4716]: I1207 17:01:51.352488 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4cb94d7f59637b410c6ea06aeac57a499ffab903391b8415798aca23986vcrp_5bdf0591-fc57-4545-8ab4-e89de972df50/pull/0.log" Dec 07 17:01:51 crc kubenswrapper[4716]: I1207 17:01:51.354333 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4cb94d7f59637b410c6ea06aeac57a499ffab903391b8415798aca23986vcrp_5bdf0591-fc57-4545-8ab4-e89de972df50/pull/0.log" Dec 07 17:01:51 crc kubenswrapper[4716]: I1207 17:01:51.556307 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4cb94d7f59637b410c6ea06aeac57a499ffab903391b8415798aca23986vcrp_5bdf0591-fc57-4545-8ab4-e89de972df50/extract/0.log" Dec 07 17:01:51 crc kubenswrapper[4716]: I1207 17:01:51.575880 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4cb94d7f59637b410c6ea06aeac57a499ffab903391b8415798aca23986vcrp_5bdf0591-fc57-4545-8ab4-e89de972df50/util/0.log" Dec 07 17:01:51 crc kubenswrapper[4716]: I1207 17:01:51.579281 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4cb94d7f59637b410c6ea06aeac57a499ffab903391b8415798aca23986vcrp_5bdf0591-fc57-4545-8ab4-e89de972df50/pull/0.log" Dec 07 17:01:51 crc kubenswrapper[4716]: I1207 17:01:51.757436 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-m957c_8765c8b2-a50b-4417-adbe-1174dcdfe172/kube-rbac-proxy/0.log" Dec 07 17:01:51 crc kubenswrapper[4716]: I1207 17:01:51.819565 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6c677c69b-4nrlz_62368956-27d2-41fa-a2b2-0fb49d869f11/kube-rbac-proxy/0.log" Dec 07 17:01:51 crc kubenswrapper[4716]: I1207 17:01:51.821472 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-m957c_8765c8b2-a50b-4417-adbe-1174dcdfe172/manager/0.log" Dec 07 17:01:51 crc kubenswrapper[4716]: I1207 17:01:51.946433 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6c677c69b-4nrlz_62368956-27d2-41fa-a2b2-0fb49d869f11/manager/0.log" Dec 07 17:01:52 crc kubenswrapper[4716]: I1207 17:01:52.013821 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-697fb699cf-zts6p_79562d20-950f-428e-ac3a-f78979053266/kube-rbac-proxy/0.log" Dec 07 17:01:52 crc kubenswrapper[4716]: I1207 17:01:52.051704 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-697fb699cf-zts6p_79562d20-950f-428e-ac3a-f78979053266/manager/0.log" Dec 07 17:01:52 crc kubenswrapper[4716]: I1207 17:01:52.190866 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5697bb5779-rcmz5_6f24594b-ebe7-4518-a067-45891924abe5/kube-rbac-proxy/0.log" Dec 07 17:01:52 crc kubenswrapper[4716]: I1207 17:01:52.306050 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5697bb5779-rcmz5_6f24594b-ebe7-4518-a067-45891924abe5/manager/0.log" Dec 07 17:01:52 crc kubenswrapper[4716]: I1207 17:01:52.365118 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-xqnzw_d9d5c8b8-240b-4c88-ad6e-b04881842f0b/kube-rbac-proxy/0.log" Dec 07 17:01:52 crc kubenswrapper[4716]: I1207 17:01:52.369184 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-xqnzw_d9d5c8b8-240b-4c88-ad6e-b04881842f0b/manager/0.log" Dec 07 17:01:52 crc kubenswrapper[4716]: I1207 17:01:52.483360 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-nx5mv_20067e63-712d-4f94-9019-627219d4299f/kube-rbac-proxy/0.log" Dec 07 17:01:52 crc kubenswrapper[4716]: I1207 17:01:52.523495 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-nx5mv_20067e63-712d-4f94-9019-627219d4299f/manager/0.log" Dec 07 17:01:52 crc kubenswrapper[4716]: I1207 17:01:52.636188 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-78d48bff9d-pxs5r_01686458-debd-419a-90ec-cd27cc6953ec/kube-rbac-proxy/0.log" Dec 07 17:01:52 crc kubenswrapper[4716]: I1207 17:01:52.760712 4716 patch_prober.go:28] interesting pod/machine-config-daemon-zcxxp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 17:01:52 crc kubenswrapper[4716]: I1207 17:01:52.760771 4716 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 17:01:52 crc kubenswrapper[4716]: I1207 17:01:52.912540 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-78d48bff9d-pxs5r_01686458-debd-419a-90ec-cd27cc6953ec/manager/0.log" Dec 07 17:01:52 crc kubenswrapper[4716]: I1207 17:01:52.995779 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-967d97867-hx48c_a929d511-1d07-429d-b302-8843a6834f52/manager/0.log" Dec 07 17:01:53 crc kubenswrapper[4716]: I1207 17:01:53.022002 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-967d97867-hx48c_a929d511-1d07-429d-b302-8843a6834f52/kube-rbac-proxy/0.log" Dec 07 17:01:53 crc kubenswrapper[4716]: I1207 17:01:53.213161 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-wctdp_2babb856-a5c2-4b60-a6bc-00d3f510f014/manager/0.log" Dec 07 17:01:53 crc kubenswrapper[4716]: I1207 17:01:53.220721 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-wctdp_2babb856-a5c2-4b60-a6bc-00d3f510f014/kube-rbac-proxy/0.log" Dec 07 17:01:53 crc kubenswrapper[4716]: I1207 17:01:53.326711 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5b5fd79c9c-kbzpp_d3ef7831-ac99-410d-90bf-42d87042964b/kube-rbac-proxy/0.log" Dec 07 17:01:53 crc kubenswrapper[4716]: I1207 17:01:53.381639 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5b5fd79c9c-kbzpp_d3ef7831-ac99-410d-90bf-42d87042964b/manager/0.log" Dec 07 17:01:53 crc kubenswrapper[4716]: I1207 17:01:53.416539 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-79c8c4686c-9dsmr_5be39543-890e-479c-a041-d864922e038f/kube-rbac-proxy/0.log" Dec 07 17:01:53 crc kubenswrapper[4716]: I1207 17:01:53.589737 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-79c8c4686c-9dsmr_5be39543-890e-479c-a041-d864922e038f/manager/0.log" Dec 07 17:01:53 crc kubenswrapper[4716]: I1207 17:01:53.666765 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-srq27_463e6bed-ce60-4e35-98a3-cd837b9066f2/kube-rbac-proxy/0.log" Dec 07 17:01:53 crc kubenswrapper[4716]: I1207 17:01:53.686199 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-srq27_463e6bed-ce60-4e35-98a3-cd837b9066f2/manager/0.log" Dec 07 17:01:53 crc kubenswrapper[4716]: I1207 17:01:53.780273 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-z7lcv_69c2e54f-9af6-44e8-abdd-226a29b64da6/kube-rbac-proxy/0.log" Dec 07 17:01:53 crc kubenswrapper[4716]: I1207 17:01:53.899911 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-z7lcv_69c2e54f-9af6-44e8-abdd-226a29b64da6/manager/0.log" Dec 07 17:01:54 crc kubenswrapper[4716]: I1207 17:01:54.030284 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-kkdgl_57c9705a-9bf1-45b4-aefb-d224a9d72a4a/manager/0.log" Dec 07 17:01:54 crc kubenswrapper[4716]: I1207 17:01:54.034350 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-kkdgl_57c9705a-9bf1-45b4-aefb-d224a9d72a4a/kube-rbac-proxy/0.log" Dec 07 17:01:54 crc kubenswrapper[4716]: I1207 17:01:54.093750 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-84b575879fdkpkx_34e0d69c-2f36-4ba6-a4e3-114cabe016c2/kube-rbac-proxy/0.log" Dec 07 17:01:54 crc kubenswrapper[4716]: I1207 17:01:54.242793 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-84b575879fdkpkx_34e0d69c-2f36-4ba6-a4e3-114cabe016c2/manager/0.log" Dec 07 17:01:54 crc kubenswrapper[4716]: I1207 17:01:54.533207 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-qcnsg_12e91e68-d5ea-4b85-81ca-88f8d00e06bc/registry-server/0.log" Dec 07 17:01:54 crc kubenswrapper[4716]: I1207 17:01:54.630380 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-64d95d5f5f-6t4rp_93975489-d46c-4ea7-a11b-65202433b51f/operator/0.log" Dec 07 17:01:54 crc kubenswrapper[4716]: I1207 17:01:54.702186 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-h95ls_1bf3b0b2-ea38-4fa3-b07c-4467dc111476/kube-rbac-proxy/0.log" Dec 07 17:01:54 crc kubenswrapper[4716]: I1207 17:01:54.862915 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-h95ls_1bf3b0b2-ea38-4fa3-b07c-4467dc111476/manager/0.log" Dec 07 17:01:54 crc kubenswrapper[4716]: I1207 17:01:54.924769 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-8xrch_0f05e35d-f818-4f65-bac6-f6a22006627e/kube-rbac-proxy/0.log" Dec 07 17:01:55 crc kubenswrapper[4716]: I1207 17:01:55.134512 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-8xrch_0f05e35d-f818-4f65-bac6-f6a22006627e/manager/0.log" Dec 07 17:01:55 crc kubenswrapper[4716]: I1207 17:01:55.209218 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-7mjmm_79aa5696-34a1-45cf-b965-56f64fe63f4e/operator/0.log" Dec 07 17:01:55 crc kubenswrapper[4716]: I1207 17:01:55.255329 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-7db7d7886d-z6n25_9d840ba3-d5ef-48fc-87b2-6af4d164c9ba/manager/0.log" Dec 07 17:01:55 crc kubenswrapper[4716]: I1207 17:01:55.348450 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-9d58d64bc-p2hbf_d9d5a1f4-edb7-4630-8907-5d29c4678f24/manager/0.log" Dec 07 17:01:55 crc kubenswrapper[4716]: I1207 17:01:55.371324 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-9d58d64bc-p2hbf_d9d5a1f4-edb7-4630-8907-5d29c4678f24/kube-rbac-proxy/0.log" Dec 07 17:01:55 crc kubenswrapper[4716]: I1207 17:01:55.422501 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-58d5ff84df-ggcr7_e4c2f2f6-3285-4f48-a428-b2065f33c046/kube-rbac-proxy/0.log" Dec 07 17:01:55 crc kubenswrapper[4716]: I1207 17:01:55.463955 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-58d5ff84df-ggcr7_e4c2f2f6-3285-4f48-a428-b2065f33c046/manager/0.log" Dec 07 17:01:55 crc kubenswrapper[4716]: I1207 17:01:55.564028 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-4dnl8_0c6a49cd-1cfe-49c8-a068-f85c735b46b1/kube-rbac-proxy/0.log" Dec 07 17:01:55 crc kubenswrapper[4716]: I1207 17:01:55.603212 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-4dnl8_0c6a49cd-1cfe-49c8-a068-f85c735b46b1/manager/0.log" Dec 07 17:01:55 crc kubenswrapper[4716]: I1207 17:01:55.662673 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-667bd8d554-8zwtk_2ef47713-a73c-4c63-8fa0-ffc530832285/kube-rbac-proxy/0.log" Dec 07 17:01:55 crc kubenswrapper[4716]: I1207 17:01:55.767637 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-667bd8d554-8zwtk_2ef47713-a73c-4c63-8fa0-ffc530832285/manager/0.log" Dec 07 17:02:14 crc kubenswrapper[4716]: I1207 17:02:14.648683 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-mmfjq_db812416-33e4-462c-a28f-ec40b629dc37/control-plane-machine-set-operator/0.log" Dec 07 17:02:14 crc kubenswrapper[4716]: I1207 17:02:14.836677 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-j7zhr_4df2ba74-d18e-4e94-bee1-703645b1a463/machine-api-operator/0.log" Dec 07 17:02:14 crc kubenswrapper[4716]: I1207 17:02:14.839222 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-j7zhr_4df2ba74-d18e-4e94-bee1-703645b1a463/kube-rbac-proxy/0.log" Dec 07 17:02:22 crc kubenswrapper[4716]: I1207 17:02:22.761118 4716 patch_prober.go:28] interesting pod/machine-config-daemon-zcxxp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 17:02:22 crc kubenswrapper[4716]: I1207 17:02:22.761716 4716 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 17:02:26 crc kubenswrapper[4716]: I1207 17:02:26.453320 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-xb4z7_55db1fd5-6c4f-4f27-888c-4d4b2445e603/cert-manager-controller/0.log" Dec 07 17:02:26 crc kubenswrapper[4716]: I1207 17:02:26.562267 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-cgnm7_64d9b2cc-19be-4cd5-81d8-fda3cbf2228e/cert-manager-cainjector/0.log" Dec 07 17:02:26 crc kubenswrapper[4716]: I1207 17:02:26.622919 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-ft9cq_badea5ba-19c9-4716-87df-a4ef7248b50a/cert-manager-webhook/0.log" Dec 07 17:02:37 crc kubenswrapper[4716]: I1207 17:02:37.970498 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-rqbh5_682c523b-c5ad-4e1c-9fd9-969911ef2242/nmstate-console-plugin/0.log" Dec 07 17:02:38 crc kubenswrapper[4716]: I1207 17:02:38.178350 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-89zc4_e4dfc168-4ec1-40d0-9d1a-d7f04063f189/nmstate-handler/0.log" Dec 07 17:02:38 crc kubenswrapper[4716]: I1207 17:02:38.182470 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-slf82_f54315f3-de9b-48b5-9baf-94e78414c0e2/kube-rbac-proxy/0.log" Dec 07 17:02:38 crc kubenswrapper[4716]: I1207 17:02:38.204365 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-slf82_f54315f3-de9b-48b5-9baf-94e78414c0e2/nmstate-metrics/0.log" Dec 07 17:02:38 crc kubenswrapper[4716]: I1207 17:02:38.524090 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-s55qn_08d6dea5-f7a4-4895-b768-f6feebceab45/nmstate-operator/0.log" Dec 07 17:02:38 crc kubenswrapper[4716]: I1207 17:02:38.543129 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-8tpxz_3546fcc6-820f-4601-9c0e-b652481582d3/nmstate-webhook/0.log" Dec 07 17:02:52 crc kubenswrapper[4716]: I1207 17:02:52.412569 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-t728x_f233fcdd-5c35-4ffb-bb33-7e2249f1c6c0/kube-rbac-proxy/0.log" Dec 07 17:02:52 crc kubenswrapper[4716]: I1207 17:02:52.481165 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-t728x_f233fcdd-5c35-4ffb-bb33-7e2249f1c6c0/controller/0.log" Dec 07 17:02:52 crc kubenswrapper[4716]: I1207 17:02:52.589661 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9zdjv_5e169ec0-a81a-45eb-b2a6-8ac97d905caa/cp-frr-files/0.log" Dec 07 17:02:52 crc kubenswrapper[4716]: I1207 17:02:52.717355 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9zdjv_5e169ec0-a81a-45eb-b2a6-8ac97d905caa/cp-frr-files/0.log" Dec 07 17:02:52 crc kubenswrapper[4716]: I1207 17:02:52.726127 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9zdjv_5e169ec0-a81a-45eb-b2a6-8ac97d905caa/cp-reloader/0.log" Dec 07 17:02:52 crc kubenswrapper[4716]: I1207 17:02:52.760777 4716 patch_prober.go:28] interesting pod/machine-config-daemon-zcxxp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 17:02:52 crc kubenswrapper[4716]: I1207 17:02:52.760949 4716 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 17:02:52 crc kubenswrapper[4716]: I1207 17:02:52.761007 4716 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" Dec 07 17:02:52 crc kubenswrapper[4716]: I1207 17:02:52.761346 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9zdjv_5e169ec0-a81a-45eb-b2a6-8ac97d905caa/cp-reloader/0.log" Dec 07 17:02:52 crc kubenswrapper[4716]: I1207 17:02:52.761867 4716 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"79a8e8c2fa02cd8c837e2fe1d95aefde9b087d4864372423b2cc8afd41b4df75"} pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 07 17:02:52 crc kubenswrapper[4716]: I1207 17:02:52.761944 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" containerName="machine-config-daemon" containerID="cri-o://79a8e8c2fa02cd8c837e2fe1d95aefde9b087d4864372423b2cc8afd41b4df75" gracePeriod=600 Dec 07 17:02:52 crc kubenswrapper[4716]: I1207 17:02:52.781457 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9zdjv_5e169ec0-a81a-45eb-b2a6-8ac97d905caa/cp-metrics/0.log" Dec 07 17:02:52 crc kubenswrapper[4716]: I1207 17:02:52.979108 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9zdjv_5e169ec0-a81a-45eb-b2a6-8ac97d905caa/cp-metrics/0.log" Dec 07 17:02:52 crc kubenswrapper[4716]: I1207 17:02:52.997781 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9zdjv_5e169ec0-a81a-45eb-b2a6-8ac97d905caa/cp-frr-files/0.log" Dec 07 17:02:53 crc kubenswrapper[4716]: I1207 17:02:53.021502 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9zdjv_5e169ec0-a81a-45eb-b2a6-8ac97d905caa/cp-metrics/0.log" Dec 07 17:02:53 crc kubenswrapper[4716]: I1207 17:02:53.050763 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9zdjv_5e169ec0-a81a-45eb-b2a6-8ac97d905caa/cp-reloader/0.log" Dec 07 17:02:53 crc kubenswrapper[4716]: I1207 17:02:53.247935 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9zdjv_5e169ec0-a81a-45eb-b2a6-8ac97d905caa/cp-metrics/0.log" Dec 07 17:02:53 crc kubenswrapper[4716]: I1207 17:02:53.279867 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9zdjv_5e169ec0-a81a-45eb-b2a6-8ac97d905caa/controller/0.log" Dec 07 17:02:53 crc kubenswrapper[4716]: I1207 17:02:53.297441 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9zdjv_5e169ec0-a81a-45eb-b2a6-8ac97d905caa/cp-reloader/0.log" Dec 07 17:02:53 crc kubenswrapper[4716]: I1207 17:02:53.305870 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9zdjv_5e169ec0-a81a-45eb-b2a6-8ac97d905caa/cp-frr-files/0.log" Dec 07 17:02:53 crc kubenswrapper[4716]: I1207 17:02:53.524036 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9zdjv_5e169ec0-a81a-45eb-b2a6-8ac97d905caa/frr-metrics/0.log" Dec 07 17:02:53 crc kubenswrapper[4716]: I1207 17:02:53.561947 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9zdjv_5e169ec0-a81a-45eb-b2a6-8ac97d905caa/kube-rbac-proxy-frr/0.log" Dec 07 17:02:53 crc kubenswrapper[4716]: I1207 17:02:53.569597 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9zdjv_5e169ec0-a81a-45eb-b2a6-8ac97d905caa/kube-rbac-proxy/0.log" Dec 07 17:02:53 crc kubenswrapper[4716]: I1207 17:02:53.690215 4716 generic.go:334] "Generic (PLEG): container finished" podID="c15b59eb-565d-4556-a4ce-75afdf159dc8" containerID="79a8e8c2fa02cd8c837e2fe1d95aefde9b087d4864372423b2cc8afd41b4df75" exitCode=0 Dec 07 17:02:53 crc kubenswrapper[4716]: I1207 17:02:53.690281 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" event={"ID":"c15b59eb-565d-4556-a4ce-75afdf159dc8","Type":"ContainerDied","Data":"79a8e8c2fa02cd8c837e2fe1d95aefde9b087d4864372423b2cc8afd41b4df75"} Dec 07 17:02:53 crc kubenswrapper[4716]: I1207 17:02:53.690311 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" event={"ID":"c15b59eb-565d-4556-a4ce-75afdf159dc8","Type":"ContainerStarted","Data":"0f9119a8d01b371a5ab06e6f944c6501025c243f659e26196309c46fc20c7a5d"} Dec 07 17:02:53 crc kubenswrapper[4716]: I1207 17:02:53.690328 4716 scope.go:117] "RemoveContainer" containerID="320114e03a98e202a688ac2d6205631a9f8d9952773c6722837647cd3ea2b74d" Dec 07 17:02:53 crc kubenswrapper[4716]: I1207 17:02:53.740221 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9zdjv_5e169ec0-a81a-45eb-b2a6-8ac97d905caa/reloader/0.log" Dec 07 17:02:53 crc kubenswrapper[4716]: I1207 17:02:53.782336 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-dqzl5_1a4973da-40f0-4277-a6ff-06ff961b4815/frr-k8s-webhook-server/0.log" Dec 07 17:02:54 crc kubenswrapper[4716]: I1207 17:02:54.020385 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-655c466c6d-8pvn6_049c277a-93fc-42db-952b-b1549f6a599c/manager/0.log" Dec 07 17:02:54 crc kubenswrapper[4716]: I1207 17:02:54.141123 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-65fbf78689-nbmlk_b6bf63fb-cc3e-4349-b7f6-efb11c1ff14f/webhook-server/0.log" Dec 07 17:02:54 crc kubenswrapper[4716]: I1207 17:02:54.334119 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-xnsb6_82d3c78d-32ff-47ed-854a-d3786c962ffd/kube-rbac-proxy/0.log" Dec 07 17:02:54 crc kubenswrapper[4716]: I1207 17:02:54.704865 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9zdjv_5e169ec0-a81a-45eb-b2a6-8ac97d905caa/frr/0.log" Dec 07 17:02:54 crc kubenswrapper[4716]: I1207 17:02:54.770330 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-xnsb6_82d3c78d-32ff-47ed-854a-d3786c962ffd/speaker/0.log" Dec 07 17:03:07 crc kubenswrapper[4716]: I1207 17:03:07.431608 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fh9qqw_7ecea701-6ded-4f27-844f-fd9dfcb92f29/util/0.log" Dec 07 17:03:07 crc kubenswrapper[4716]: I1207 17:03:07.562051 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fh9qqw_7ecea701-6ded-4f27-844f-fd9dfcb92f29/util/0.log" Dec 07 17:03:07 crc kubenswrapper[4716]: I1207 17:03:07.594811 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fh9qqw_7ecea701-6ded-4f27-844f-fd9dfcb92f29/pull/0.log" Dec 07 17:03:07 crc kubenswrapper[4716]: I1207 17:03:07.631588 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fh9qqw_7ecea701-6ded-4f27-844f-fd9dfcb92f29/pull/0.log" Dec 07 17:03:07 crc kubenswrapper[4716]: I1207 17:03:07.749822 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fh9qqw_7ecea701-6ded-4f27-844f-fd9dfcb92f29/pull/0.log" Dec 07 17:03:07 crc kubenswrapper[4716]: I1207 17:03:07.773509 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fh9qqw_7ecea701-6ded-4f27-844f-fd9dfcb92f29/util/0.log" Dec 07 17:03:07 crc kubenswrapper[4716]: I1207 17:03:07.845465 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fh9qqw_7ecea701-6ded-4f27-844f-fd9dfcb92f29/extract/0.log" Dec 07 17:03:07 crc kubenswrapper[4716]: I1207 17:03:07.994288 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c2vdd_19363c75-987f-4c3b-bbca-28a9cc33238b/util/0.log" Dec 07 17:03:08 crc kubenswrapper[4716]: I1207 17:03:08.171623 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c2vdd_19363c75-987f-4c3b-bbca-28a9cc33238b/pull/0.log" Dec 07 17:03:08 crc kubenswrapper[4716]: I1207 17:03:08.187540 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c2vdd_19363c75-987f-4c3b-bbca-28a9cc33238b/pull/0.log" Dec 07 17:03:08 crc kubenswrapper[4716]: I1207 17:03:08.204092 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c2vdd_19363c75-987f-4c3b-bbca-28a9cc33238b/util/0.log" Dec 07 17:03:08 crc kubenswrapper[4716]: I1207 17:03:08.379793 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c2vdd_19363c75-987f-4c3b-bbca-28a9cc33238b/pull/0.log" Dec 07 17:03:08 crc kubenswrapper[4716]: I1207 17:03:08.392231 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c2vdd_19363c75-987f-4c3b-bbca-28a9cc33238b/extract/0.log" Dec 07 17:03:08 crc kubenswrapper[4716]: I1207 17:03:08.424814 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c2vdd_19363c75-987f-4c3b-bbca-28a9cc33238b/util/0.log" Dec 07 17:03:08 crc kubenswrapper[4716]: I1207 17:03:08.558439 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tgn9w_fc342361-554a-41c0-a512-06e046ff03eb/extract-utilities/0.log" Dec 07 17:03:08 crc kubenswrapper[4716]: I1207 17:03:08.736160 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tgn9w_fc342361-554a-41c0-a512-06e046ff03eb/extract-content/0.log" Dec 07 17:03:08 crc kubenswrapper[4716]: I1207 17:03:08.745473 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tgn9w_fc342361-554a-41c0-a512-06e046ff03eb/extract-utilities/0.log" Dec 07 17:03:08 crc kubenswrapper[4716]: I1207 17:03:08.747483 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tgn9w_fc342361-554a-41c0-a512-06e046ff03eb/extract-content/0.log" Dec 07 17:03:08 crc kubenswrapper[4716]: I1207 17:03:08.915371 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tgn9w_fc342361-554a-41c0-a512-06e046ff03eb/extract-content/0.log" Dec 07 17:03:08 crc kubenswrapper[4716]: I1207 17:03:08.918177 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tgn9w_fc342361-554a-41c0-a512-06e046ff03eb/extract-utilities/0.log" Dec 07 17:03:09 crc kubenswrapper[4716]: I1207 17:03:09.104270 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-vmnk2_33db3f8a-1f66-44c4-b563-2f07dbc8267e/extract-utilities/0.log" Dec 07 17:03:09 crc kubenswrapper[4716]: I1207 17:03:09.296054 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tgn9w_fc342361-554a-41c0-a512-06e046ff03eb/registry-server/0.log" Dec 07 17:03:09 crc kubenswrapper[4716]: I1207 17:03:09.297178 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-vmnk2_33db3f8a-1f66-44c4-b563-2f07dbc8267e/extract-utilities/0.log" Dec 07 17:03:09 crc kubenswrapper[4716]: I1207 17:03:09.310436 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-vmnk2_33db3f8a-1f66-44c4-b563-2f07dbc8267e/extract-content/0.log" Dec 07 17:03:09 crc kubenswrapper[4716]: I1207 17:03:09.356688 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-vmnk2_33db3f8a-1f66-44c4-b563-2f07dbc8267e/extract-content/0.log" Dec 07 17:03:09 crc kubenswrapper[4716]: I1207 17:03:09.474569 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-vmnk2_33db3f8a-1f66-44c4-b563-2f07dbc8267e/extract-content/0.log" Dec 07 17:03:09 crc kubenswrapper[4716]: I1207 17:03:09.507780 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-vmnk2_33db3f8a-1f66-44c4-b563-2f07dbc8267e/extract-utilities/0.log" Dec 07 17:03:09 crc kubenswrapper[4716]: I1207 17:03:09.721892 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-dhjm4_845614d7-aa72-46de-b358-2e39f0209886/marketplace-operator/0.log" Dec 07 17:03:09 crc kubenswrapper[4716]: I1207 17:03:09.833658 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f7pqt_0f030783-1303-4815-8ac1-fbecf8e036fe/extract-utilities/0.log" Dec 07 17:03:09 crc kubenswrapper[4716]: I1207 17:03:09.874758 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-vmnk2_33db3f8a-1f66-44c4-b563-2f07dbc8267e/registry-server/0.log" Dec 07 17:03:09 crc kubenswrapper[4716]: I1207 17:03:09.991824 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f7pqt_0f030783-1303-4815-8ac1-fbecf8e036fe/extract-utilities/0.log" Dec 07 17:03:09 crc kubenswrapper[4716]: I1207 17:03:09.997348 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f7pqt_0f030783-1303-4815-8ac1-fbecf8e036fe/extract-content/0.log" Dec 07 17:03:10 crc kubenswrapper[4716]: I1207 17:03:10.040665 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f7pqt_0f030783-1303-4815-8ac1-fbecf8e036fe/extract-content/0.log" Dec 07 17:03:10 crc kubenswrapper[4716]: I1207 17:03:10.245138 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f7pqt_0f030783-1303-4815-8ac1-fbecf8e036fe/extract-utilities/0.log" Dec 07 17:03:10 crc kubenswrapper[4716]: I1207 17:03:10.252676 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f7pqt_0f030783-1303-4815-8ac1-fbecf8e036fe/extract-content/0.log" Dec 07 17:03:10 crc kubenswrapper[4716]: I1207 17:03:10.363585 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f7pqt_0f030783-1303-4815-8ac1-fbecf8e036fe/registry-server/0.log" Dec 07 17:03:10 crc kubenswrapper[4716]: I1207 17:03:10.413724 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-nkjl4_af948d72-6e6c-42d9-9d0b-1821d7f47176/extract-utilities/0.log" Dec 07 17:03:10 crc kubenswrapper[4716]: I1207 17:03:10.583139 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-nkjl4_af948d72-6e6c-42d9-9d0b-1821d7f47176/extract-content/0.log" Dec 07 17:03:10 crc kubenswrapper[4716]: I1207 17:03:10.586681 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-nkjl4_af948d72-6e6c-42d9-9d0b-1821d7f47176/extract-utilities/0.log" Dec 07 17:03:10 crc kubenswrapper[4716]: I1207 17:03:10.622542 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-nkjl4_af948d72-6e6c-42d9-9d0b-1821d7f47176/extract-content/0.log" Dec 07 17:03:10 crc kubenswrapper[4716]: I1207 17:03:10.824176 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-nkjl4_af948d72-6e6c-42d9-9d0b-1821d7f47176/extract-utilities/0.log" Dec 07 17:03:10 crc kubenswrapper[4716]: I1207 17:03:10.837041 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-nkjl4_af948d72-6e6c-42d9-9d0b-1821d7f47176/extract-content/0.log" Dec 07 17:03:11 crc kubenswrapper[4716]: I1207 17:03:11.255828 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-nkjl4_af948d72-6e6c-42d9-9d0b-1821d7f47176/registry-server/0.log" Dec 07 17:03:31 crc kubenswrapper[4716]: I1207 17:03:31.748431 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-2wfb4"] Dec 07 17:03:31 crc kubenswrapper[4716]: E1207 17:03:31.749305 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c7bbcbc-7ce9-419d-b63d-8bcc377e4264" containerName="keystone-cron" Dec 07 17:03:31 crc kubenswrapper[4716]: I1207 17:03:31.749317 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c7bbcbc-7ce9-419d-b63d-8bcc377e4264" containerName="keystone-cron" Dec 07 17:03:31 crc kubenswrapper[4716]: I1207 17:03:31.749494 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c7bbcbc-7ce9-419d-b63d-8bcc377e4264" containerName="keystone-cron" Dec 07 17:03:31 crc kubenswrapper[4716]: I1207 17:03:31.750750 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2wfb4" Dec 07 17:03:31 crc kubenswrapper[4716]: I1207 17:03:31.767976 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2wfb4"] Dec 07 17:03:31 crc kubenswrapper[4716]: I1207 17:03:31.885928 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/778133cd-af64-46f4-a87c-9c69a513a73b-utilities\") pod \"community-operators-2wfb4\" (UID: \"778133cd-af64-46f4-a87c-9c69a513a73b\") " pod="openshift-marketplace/community-operators-2wfb4" Dec 07 17:03:31 crc kubenswrapper[4716]: I1207 17:03:31.885995 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/778133cd-af64-46f4-a87c-9c69a513a73b-catalog-content\") pod \"community-operators-2wfb4\" (UID: \"778133cd-af64-46f4-a87c-9c69a513a73b\") " pod="openshift-marketplace/community-operators-2wfb4" Dec 07 17:03:31 crc kubenswrapper[4716]: I1207 17:03:31.886315 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbhkl\" (UniqueName: \"kubernetes.io/projected/778133cd-af64-46f4-a87c-9c69a513a73b-kube-api-access-jbhkl\") pod \"community-operators-2wfb4\" (UID: \"778133cd-af64-46f4-a87c-9c69a513a73b\") " pod="openshift-marketplace/community-operators-2wfb4" Dec 07 17:03:31 crc kubenswrapper[4716]: I1207 17:03:31.988064 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/778133cd-af64-46f4-a87c-9c69a513a73b-utilities\") pod \"community-operators-2wfb4\" (UID: \"778133cd-af64-46f4-a87c-9c69a513a73b\") " pod="openshift-marketplace/community-operators-2wfb4" Dec 07 17:03:31 crc kubenswrapper[4716]: I1207 17:03:31.988156 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/778133cd-af64-46f4-a87c-9c69a513a73b-catalog-content\") pod \"community-operators-2wfb4\" (UID: \"778133cd-af64-46f4-a87c-9c69a513a73b\") " pod="openshift-marketplace/community-operators-2wfb4" Dec 07 17:03:31 crc kubenswrapper[4716]: I1207 17:03:31.988235 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jbhkl\" (UniqueName: \"kubernetes.io/projected/778133cd-af64-46f4-a87c-9c69a513a73b-kube-api-access-jbhkl\") pod \"community-operators-2wfb4\" (UID: \"778133cd-af64-46f4-a87c-9c69a513a73b\") " pod="openshift-marketplace/community-operators-2wfb4" Dec 07 17:03:31 crc kubenswrapper[4716]: I1207 17:03:31.989164 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/778133cd-af64-46f4-a87c-9c69a513a73b-utilities\") pod \"community-operators-2wfb4\" (UID: \"778133cd-af64-46f4-a87c-9c69a513a73b\") " pod="openshift-marketplace/community-operators-2wfb4" Dec 07 17:03:31 crc kubenswrapper[4716]: I1207 17:03:31.989422 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/778133cd-af64-46f4-a87c-9c69a513a73b-catalog-content\") pod \"community-operators-2wfb4\" (UID: \"778133cd-af64-46f4-a87c-9c69a513a73b\") " pod="openshift-marketplace/community-operators-2wfb4" Dec 07 17:03:32 crc kubenswrapper[4716]: I1207 17:03:32.006829 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbhkl\" (UniqueName: \"kubernetes.io/projected/778133cd-af64-46f4-a87c-9c69a513a73b-kube-api-access-jbhkl\") pod \"community-operators-2wfb4\" (UID: \"778133cd-af64-46f4-a87c-9c69a513a73b\") " pod="openshift-marketplace/community-operators-2wfb4" Dec 07 17:03:32 crc kubenswrapper[4716]: I1207 17:03:32.093383 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2wfb4" Dec 07 17:03:32 crc kubenswrapper[4716]: I1207 17:03:32.754673 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2wfb4"] Dec 07 17:03:33 crc kubenswrapper[4716]: I1207 17:03:33.050112 4716 generic.go:334] "Generic (PLEG): container finished" podID="778133cd-af64-46f4-a87c-9c69a513a73b" containerID="1203f8950b701419596da899fd57ec96550697997d7272e5a24ca12e8854e667" exitCode=0 Dec 07 17:03:33 crc kubenswrapper[4716]: I1207 17:03:33.050214 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2wfb4" event={"ID":"778133cd-af64-46f4-a87c-9c69a513a73b","Type":"ContainerDied","Data":"1203f8950b701419596da899fd57ec96550697997d7272e5a24ca12e8854e667"} Dec 07 17:03:33 crc kubenswrapper[4716]: I1207 17:03:33.050393 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2wfb4" event={"ID":"778133cd-af64-46f4-a87c-9c69a513a73b","Type":"ContainerStarted","Data":"73f2ebc4e9a8d121268f65bd56d2b47b9c0e928ba5ac81f6354d92da35082444"} Dec 07 17:03:33 crc kubenswrapper[4716]: I1207 17:03:33.051984 4716 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 07 17:03:34 crc kubenswrapper[4716]: I1207 17:03:34.062656 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2wfb4" event={"ID":"778133cd-af64-46f4-a87c-9c69a513a73b","Type":"ContainerStarted","Data":"12c186c49834e242110e1b0aae47029fb501c7d28b78d1ee0258847de53b80d0"} Dec 07 17:03:35 crc kubenswrapper[4716]: I1207 17:03:35.082183 4716 generic.go:334] "Generic (PLEG): container finished" podID="778133cd-af64-46f4-a87c-9c69a513a73b" containerID="12c186c49834e242110e1b0aae47029fb501c7d28b78d1ee0258847de53b80d0" exitCode=0 Dec 07 17:03:35 crc kubenswrapper[4716]: I1207 17:03:35.082447 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2wfb4" event={"ID":"778133cd-af64-46f4-a87c-9c69a513a73b","Type":"ContainerDied","Data":"12c186c49834e242110e1b0aae47029fb501c7d28b78d1ee0258847de53b80d0"} Dec 07 17:03:36 crc kubenswrapper[4716]: I1207 17:03:36.096571 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2wfb4" event={"ID":"778133cd-af64-46f4-a87c-9c69a513a73b","Type":"ContainerStarted","Data":"cde8dc95052812ab03e40b6810821174285d605f9edc140528dab6fab6870d24"} Dec 07 17:03:36 crc kubenswrapper[4716]: I1207 17:03:36.122066 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-2wfb4" podStartSLOduration=2.690896254 podStartE2EDuration="5.122050037s" podCreationTimestamp="2025-12-07 17:03:31 +0000 UTC" firstStartedPulling="2025-12-07 17:03:33.051765388 +0000 UTC m=+3675.742050300" lastFinishedPulling="2025-12-07 17:03:35.482919171 +0000 UTC m=+3678.173204083" observedRunningTime="2025-12-07 17:03:36.119602931 +0000 UTC m=+3678.809887843" watchObservedRunningTime="2025-12-07 17:03:36.122050037 +0000 UTC m=+3678.812334949" Dec 07 17:03:42 crc kubenswrapper[4716]: I1207 17:03:42.094931 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-2wfb4" Dec 07 17:03:42 crc kubenswrapper[4716]: I1207 17:03:42.095393 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-2wfb4" Dec 07 17:03:42 crc kubenswrapper[4716]: I1207 17:03:42.138912 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-2wfb4" Dec 07 17:03:42 crc kubenswrapper[4716]: I1207 17:03:42.199750 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-2wfb4" Dec 07 17:03:42 crc kubenswrapper[4716]: I1207 17:03:42.385848 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2wfb4"] Dec 07 17:03:44 crc kubenswrapper[4716]: I1207 17:03:44.171092 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-2wfb4" podUID="778133cd-af64-46f4-a87c-9c69a513a73b" containerName="registry-server" containerID="cri-o://cde8dc95052812ab03e40b6810821174285d605f9edc140528dab6fab6870d24" gracePeriod=2 Dec 07 17:03:45 crc kubenswrapper[4716]: I1207 17:03:45.171272 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2wfb4" Dec 07 17:03:45 crc kubenswrapper[4716]: I1207 17:03:45.181101 4716 generic.go:334] "Generic (PLEG): container finished" podID="778133cd-af64-46f4-a87c-9c69a513a73b" containerID="cde8dc95052812ab03e40b6810821174285d605f9edc140528dab6fab6870d24" exitCode=0 Dec 07 17:03:45 crc kubenswrapper[4716]: I1207 17:03:45.181136 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2wfb4" event={"ID":"778133cd-af64-46f4-a87c-9c69a513a73b","Type":"ContainerDied","Data":"cde8dc95052812ab03e40b6810821174285d605f9edc140528dab6fab6870d24"} Dec 07 17:03:45 crc kubenswrapper[4716]: I1207 17:03:45.181159 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2wfb4" event={"ID":"778133cd-af64-46f4-a87c-9c69a513a73b","Type":"ContainerDied","Data":"73f2ebc4e9a8d121268f65bd56d2b47b9c0e928ba5ac81f6354d92da35082444"} Dec 07 17:03:45 crc kubenswrapper[4716]: I1207 17:03:45.181177 4716 scope.go:117] "RemoveContainer" containerID="cde8dc95052812ab03e40b6810821174285d605f9edc140528dab6fab6870d24" Dec 07 17:03:45 crc kubenswrapper[4716]: I1207 17:03:45.181458 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2wfb4" Dec 07 17:03:45 crc kubenswrapper[4716]: I1207 17:03:45.212514 4716 scope.go:117] "RemoveContainer" containerID="12c186c49834e242110e1b0aae47029fb501c7d28b78d1ee0258847de53b80d0" Dec 07 17:03:45 crc kubenswrapper[4716]: I1207 17:03:45.235979 4716 scope.go:117] "RemoveContainer" containerID="1203f8950b701419596da899fd57ec96550697997d7272e5a24ca12e8854e667" Dec 07 17:03:45 crc kubenswrapper[4716]: I1207 17:03:45.276341 4716 scope.go:117] "RemoveContainer" containerID="cde8dc95052812ab03e40b6810821174285d605f9edc140528dab6fab6870d24" Dec 07 17:03:45 crc kubenswrapper[4716]: E1207 17:03:45.276857 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cde8dc95052812ab03e40b6810821174285d605f9edc140528dab6fab6870d24\": container with ID starting with cde8dc95052812ab03e40b6810821174285d605f9edc140528dab6fab6870d24 not found: ID does not exist" containerID="cde8dc95052812ab03e40b6810821174285d605f9edc140528dab6fab6870d24" Dec 07 17:03:45 crc kubenswrapper[4716]: I1207 17:03:45.276895 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cde8dc95052812ab03e40b6810821174285d605f9edc140528dab6fab6870d24"} err="failed to get container status \"cde8dc95052812ab03e40b6810821174285d605f9edc140528dab6fab6870d24\": rpc error: code = NotFound desc = could not find container \"cde8dc95052812ab03e40b6810821174285d605f9edc140528dab6fab6870d24\": container with ID starting with cde8dc95052812ab03e40b6810821174285d605f9edc140528dab6fab6870d24 not found: ID does not exist" Dec 07 17:03:45 crc kubenswrapper[4716]: I1207 17:03:45.276921 4716 scope.go:117] "RemoveContainer" containerID="12c186c49834e242110e1b0aae47029fb501c7d28b78d1ee0258847de53b80d0" Dec 07 17:03:45 crc kubenswrapper[4716]: E1207 17:03:45.277378 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"12c186c49834e242110e1b0aae47029fb501c7d28b78d1ee0258847de53b80d0\": container with ID starting with 12c186c49834e242110e1b0aae47029fb501c7d28b78d1ee0258847de53b80d0 not found: ID does not exist" containerID="12c186c49834e242110e1b0aae47029fb501c7d28b78d1ee0258847de53b80d0" Dec 07 17:03:45 crc kubenswrapper[4716]: I1207 17:03:45.277414 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12c186c49834e242110e1b0aae47029fb501c7d28b78d1ee0258847de53b80d0"} err="failed to get container status \"12c186c49834e242110e1b0aae47029fb501c7d28b78d1ee0258847de53b80d0\": rpc error: code = NotFound desc = could not find container \"12c186c49834e242110e1b0aae47029fb501c7d28b78d1ee0258847de53b80d0\": container with ID starting with 12c186c49834e242110e1b0aae47029fb501c7d28b78d1ee0258847de53b80d0 not found: ID does not exist" Dec 07 17:03:45 crc kubenswrapper[4716]: I1207 17:03:45.277435 4716 scope.go:117] "RemoveContainer" containerID="1203f8950b701419596da899fd57ec96550697997d7272e5a24ca12e8854e667" Dec 07 17:03:45 crc kubenswrapper[4716]: E1207 17:03:45.277821 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1203f8950b701419596da899fd57ec96550697997d7272e5a24ca12e8854e667\": container with ID starting with 1203f8950b701419596da899fd57ec96550697997d7272e5a24ca12e8854e667 not found: ID does not exist" containerID="1203f8950b701419596da899fd57ec96550697997d7272e5a24ca12e8854e667" Dec 07 17:03:45 crc kubenswrapper[4716]: I1207 17:03:45.277841 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1203f8950b701419596da899fd57ec96550697997d7272e5a24ca12e8854e667"} err="failed to get container status \"1203f8950b701419596da899fd57ec96550697997d7272e5a24ca12e8854e667\": rpc error: code = NotFound desc = could not find container \"1203f8950b701419596da899fd57ec96550697997d7272e5a24ca12e8854e667\": container with ID starting with 1203f8950b701419596da899fd57ec96550697997d7272e5a24ca12e8854e667 not found: ID does not exist" Dec 07 17:03:45 crc kubenswrapper[4716]: I1207 17:03:45.300713 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jbhkl\" (UniqueName: \"kubernetes.io/projected/778133cd-af64-46f4-a87c-9c69a513a73b-kube-api-access-jbhkl\") pod \"778133cd-af64-46f4-a87c-9c69a513a73b\" (UID: \"778133cd-af64-46f4-a87c-9c69a513a73b\") " Dec 07 17:03:45 crc kubenswrapper[4716]: I1207 17:03:45.300796 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/778133cd-af64-46f4-a87c-9c69a513a73b-catalog-content\") pod \"778133cd-af64-46f4-a87c-9c69a513a73b\" (UID: \"778133cd-af64-46f4-a87c-9c69a513a73b\") " Dec 07 17:03:45 crc kubenswrapper[4716]: I1207 17:03:45.300913 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/778133cd-af64-46f4-a87c-9c69a513a73b-utilities\") pod \"778133cd-af64-46f4-a87c-9c69a513a73b\" (UID: \"778133cd-af64-46f4-a87c-9c69a513a73b\") " Dec 07 17:03:45 crc kubenswrapper[4716]: I1207 17:03:45.301826 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/778133cd-af64-46f4-a87c-9c69a513a73b-utilities" (OuterVolumeSpecName: "utilities") pod "778133cd-af64-46f4-a87c-9c69a513a73b" (UID: "778133cd-af64-46f4-a87c-9c69a513a73b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 17:03:45 crc kubenswrapper[4716]: I1207 17:03:45.306166 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/778133cd-af64-46f4-a87c-9c69a513a73b-kube-api-access-jbhkl" (OuterVolumeSpecName: "kube-api-access-jbhkl") pod "778133cd-af64-46f4-a87c-9c69a513a73b" (UID: "778133cd-af64-46f4-a87c-9c69a513a73b"). InnerVolumeSpecName "kube-api-access-jbhkl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 17:03:45 crc kubenswrapper[4716]: I1207 17:03:45.367520 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/778133cd-af64-46f4-a87c-9c69a513a73b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "778133cd-af64-46f4-a87c-9c69a513a73b" (UID: "778133cd-af64-46f4-a87c-9c69a513a73b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 17:03:45 crc kubenswrapper[4716]: I1207 17:03:45.404355 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jbhkl\" (UniqueName: \"kubernetes.io/projected/778133cd-af64-46f4-a87c-9c69a513a73b-kube-api-access-jbhkl\") on node \"crc\" DevicePath \"\"" Dec 07 17:03:45 crc kubenswrapper[4716]: I1207 17:03:45.404961 4716 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/778133cd-af64-46f4-a87c-9c69a513a73b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 07 17:03:45 crc kubenswrapper[4716]: I1207 17:03:45.405242 4716 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/778133cd-af64-46f4-a87c-9c69a513a73b-utilities\") on node \"crc\" DevicePath \"\"" Dec 07 17:03:45 crc kubenswrapper[4716]: I1207 17:03:45.549887 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2wfb4"] Dec 07 17:03:45 crc kubenswrapper[4716]: I1207 17:03:45.563829 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-2wfb4"] Dec 07 17:03:45 crc kubenswrapper[4716]: I1207 17:03:45.668163 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="778133cd-af64-46f4-a87c-9c69a513a73b" path="/var/lib/kubelet/pods/778133cd-af64-46f4-a87c-9c69a513a73b/volumes" Dec 07 17:04:46 crc kubenswrapper[4716]: I1207 17:04:46.964309 4716 generic.go:334] "Generic (PLEG): container finished" podID="68a27046-ae9a-4651-98fd-d5633838e785" containerID="7c6bdbfe919aa4baaebaa598b7bbb8f5e7594bb501a25a21f7e642540ca5eb48" exitCode=0 Dec 07 17:04:46 crc kubenswrapper[4716]: I1207 17:04:46.964364 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zlz52/must-gather-nx26c" event={"ID":"68a27046-ae9a-4651-98fd-d5633838e785","Type":"ContainerDied","Data":"7c6bdbfe919aa4baaebaa598b7bbb8f5e7594bb501a25a21f7e642540ca5eb48"} Dec 07 17:04:46 crc kubenswrapper[4716]: I1207 17:04:46.965739 4716 scope.go:117] "RemoveContainer" containerID="7c6bdbfe919aa4baaebaa598b7bbb8f5e7594bb501a25a21f7e642540ca5eb48" Dec 07 17:04:47 crc kubenswrapper[4716]: I1207 17:04:47.090776 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-zlz52_must-gather-nx26c_68a27046-ae9a-4651-98fd-d5633838e785/gather/0.log" Dec 07 17:04:55 crc kubenswrapper[4716]: I1207 17:04:55.102311 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-zlz52/must-gather-nx26c"] Dec 07 17:04:55 crc kubenswrapper[4716]: I1207 17:04:55.103675 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-zlz52/must-gather-nx26c" podUID="68a27046-ae9a-4651-98fd-d5633838e785" containerName="copy" containerID="cri-o://3313afc325e5011a91e073a0f91f05c04984c5290f76cca9a6a62a3c75663fa7" gracePeriod=2 Dec 07 17:04:55 crc kubenswrapper[4716]: I1207 17:04:55.118932 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-zlz52/must-gather-nx26c"] Dec 07 17:04:55 crc kubenswrapper[4716]: I1207 17:04:55.537821 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-zlz52_must-gather-nx26c_68a27046-ae9a-4651-98fd-d5633838e785/copy/0.log" Dec 07 17:04:55 crc kubenswrapper[4716]: I1207 17:04:55.538659 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zlz52/must-gather-nx26c" Dec 07 17:04:55 crc kubenswrapper[4716]: I1207 17:04:55.684236 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/68a27046-ae9a-4651-98fd-d5633838e785-must-gather-output\") pod \"68a27046-ae9a-4651-98fd-d5633838e785\" (UID: \"68a27046-ae9a-4651-98fd-d5633838e785\") " Dec 07 17:04:55 crc kubenswrapper[4716]: I1207 17:04:55.684362 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l4vc6\" (UniqueName: \"kubernetes.io/projected/68a27046-ae9a-4651-98fd-d5633838e785-kube-api-access-l4vc6\") pod \"68a27046-ae9a-4651-98fd-d5633838e785\" (UID: \"68a27046-ae9a-4651-98fd-d5633838e785\") " Dec 07 17:04:55 crc kubenswrapper[4716]: I1207 17:04:55.689464 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68a27046-ae9a-4651-98fd-d5633838e785-kube-api-access-l4vc6" (OuterVolumeSpecName: "kube-api-access-l4vc6") pod "68a27046-ae9a-4651-98fd-d5633838e785" (UID: "68a27046-ae9a-4651-98fd-d5633838e785"). InnerVolumeSpecName "kube-api-access-l4vc6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 17:04:55 crc kubenswrapper[4716]: I1207 17:04:55.787552 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l4vc6\" (UniqueName: \"kubernetes.io/projected/68a27046-ae9a-4651-98fd-d5633838e785-kube-api-access-l4vc6\") on node \"crc\" DevicePath \"\"" Dec 07 17:04:55 crc kubenswrapper[4716]: I1207 17:04:55.819200 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68a27046-ae9a-4651-98fd-d5633838e785-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "68a27046-ae9a-4651-98fd-d5633838e785" (UID: "68a27046-ae9a-4651-98fd-d5633838e785"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 17:04:55 crc kubenswrapper[4716]: I1207 17:04:55.889841 4716 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/68a27046-ae9a-4651-98fd-d5633838e785-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 07 17:04:56 crc kubenswrapper[4716]: I1207 17:04:56.063712 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-zlz52_must-gather-nx26c_68a27046-ae9a-4651-98fd-d5633838e785/copy/0.log" Dec 07 17:04:56 crc kubenswrapper[4716]: I1207 17:04:56.064451 4716 generic.go:334] "Generic (PLEG): container finished" podID="68a27046-ae9a-4651-98fd-d5633838e785" containerID="3313afc325e5011a91e073a0f91f05c04984c5290f76cca9a6a62a3c75663fa7" exitCode=143 Dec 07 17:04:56 crc kubenswrapper[4716]: I1207 17:04:56.064509 4716 scope.go:117] "RemoveContainer" containerID="3313afc325e5011a91e073a0f91f05c04984c5290f76cca9a6a62a3c75663fa7" Dec 07 17:04:56 crc kubenswrapper[4716]: I1207 17:04:56.064631 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zlz52/must-gather-nx26c" Dec 07 17:04:56 crc kubenswrapper[4716]: I1207 17:04:56.089172 4716 scope.go:117] "RemoveContainer" containerID="7c6bdbfe919aa4baaebaa598b7bbb8f5e7594bb501a25a21f7e642540ca5eb48" Dec 07 17:04:56 crc kubenswrapper[4716]: I1207 17:04:56.149007 4716 scope.go:117] "RemoveContainer" containerID="3313afc325e5011a91e073a0f91f05c04984c5290f76cca9a6a62a3c75663fa7" Dec 07 17:04:56 crc kubenswrapper[4716]: E1207 17:04:56.149628 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3313afc325e5011a91e073a0f91f05c04984c5290f76cca9a6a62a3c75663fa7\": container with ID starting with 3313afc325e5011a91e073a0f91f05c04984c5290f76cca9a6a62a3c75663fa7 not found: ID does not exist" containerID="3313afc325e5011a91e073a0f91f05c04984c5290f76cca9a6a62a3c75663fa7" Dec 07 17:04:56 crc kubenswrapper[4716]: I1207 17:04:56.149661 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3313afc325e5011a91e073a0f91f05c04984c5290f76cca9a6a62a3c75663fa7"} err="failed to get container status \"3313afc325e5011a91e073a0f91f05c04984c5290f76cca9a6a62a3c75663fa7\": rpc error: code = NotFound desc = could not find container \"3313afc325e5011a91e073a0f91f05c04984c5290f76cca9a6a62a3c75663fa7\": container with ID starting with 3313afc325e5011a91e073a0f91f05c04984c5290f76cca9a6a62a3c75663fa7 not found: ID does not exist" Dec 07 17:04:56 crc kubenswrapper[4716]: I1207 17:04:56.149686 4716 scope.go:117] "RemoveContainer" containerID="7c6bdbfe919aa4baaebaa598b7bbb8f5e7594bb501a25a21f7e642540ca5eb48" Dec 07 17:04:56 crc kubenswrapper[4716]: E1207 17:04:56.149999 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c6bdbfe919aa4baaebaa598b7bbb8f5e7594bb501a25a21f7e642540ca5eb48\": container with ID starting with 7c6bdbfe919aa4baaebaa598b7bbb8f5e7594bb501a25a21f7e642540ca5eb48 not found: ID does not exist" containerID="7c6bdbfe919aa4baaebaa598b7bbb8f5e7594bb501a25a21f7e642540ca5eb48" Dec 07 17:04:56 crc kubenswrapper[4716]: I1207 17:04:56.150019 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c6bdbfe919aa4baaebaa598b7bbb8f5e7594bb501a25a21f7e642540ca5eb48"} err="failed to get container status \"7c6bdbfe919aa4baaebaa598b7bbb8f5e7594bb501a25a21f7e642540ca5eb48\": rpc error: code = NotFound desc = could not find container \"7c6bdbfe919aa4baaebaa598b7bbb8f5e7594bb501a25a21f7e642540ca5eb48\": container with ID starting with 7c6bdbfe919aa4baaebaa598b7bbb8f5e7594bb501a25a21f7e642540ca5eb48 not found: ID does not exist" Dec 07 17:04:57 crc kubenswrapper[4716]: I1207 17:04:57.670156 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68a27046-ae9a-4651-98fd-d5633838e785" path="/var/lib/kubelet/pods/68a27046-ae9a-4651-98fd-d5633838e785/volumes" Dec 07 17:05:22 crc kubenswrapper[4716]: I1207 17:05:22.761351 4716 patch_prober.go:28] interesting pod/machine-config-daemon-zcxxp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 17:05:22 crc kubenswrapper[4716]: I1207 17:05:22.762434 4716 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 17:05:52 crc kubenswrapper[4716]: I1207 17:05:52.761434 4716 patch_prober.go:28] interesting pod/machine-config-daemon-zcxxp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 17:05:52 crc kubenswrapper[4716]: I1207 17:05:52.762190 4716 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 17:06:22 crc kubenswrapper[4716]: I1207 17:06:22.761954 4716 patch_prober.go:28] interesting pod/machine-config-daemon-zcxxp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 17:06:22 crc kubenswrapper[4716]: I1207 17:06:22.762653 4716 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 17:06:22 crc kubenswrapper[4716]: I1207 17:06:22.763043 4716 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" Dec 07 17:06:22 crc kubenswrapper[4716]: I1207 17:06:22.764508 4716 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0f9119a8d01b371a5ab06e6f944c6501025c243f659e26196309c46fc20c7a5d"} pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 07 17:06:22 crc kubenswrapper[4716]: I1207 17:06:22.764636 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" containerName="machine-config-daemon" containerID="cri-o://0f9119a8d01b371a5ab06e6f944c6501025c243f659e26196309c46fc20c7a5d" gracePeriod=600 Dec 07 17:06:22 crc kubenswrapper[4716]: E1207 17:06:22.907842 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 17:06:22 crc kubenswrapper[4716]: I1207 17:06:22.973770 4716 generic.go:334] "Generic (PLEG): container finished" podID="c15b59eb-565d-4556-a4ce-75afdf159dc8" containerID="0f9119a8d01b371a5ab06e6f944c6501025c243f659e26196309c46fc20c7a5d" exitCode=0 Dec 07 17:06:22 crc kubenswrapper[4716]: I1207 17:06:22.974176 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" event={"ID":"c15b59eb-565d-4556-a4ce-75afdf159dc8","Type":"ContainerDied","Data":"0f9119a8d01b371a5ab06e6f944c6501025c243f659e26196309c46fc20c7a5d"} Dec 07 17:06:22 crc kubenswrapper[4716]: I1207 17:06:22.974220 4716 scope.go:117] "RemoveContainer" containerID="79a8e8c2fa02cd8c837e2fe1d95aefde9b087d4864372423b2cc8afd41b4df75" Dec 07 17:06:22 crc kubenswrapper[4716]: I1207 17:06:22.975285 4716 scope.go:117] "RemoveContainer" containerID="0f9119a8d01b371a5ab06e6f944c6501025c243f659e26196309c46fc20c7a5d" Dec 07 17:06:22 crc kubenswrapper[4716]: E1207 17:06:22.975670 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 17:06:35 crc kubenswrapper[4716]: I1207 17:06:35.492360 4716 scope.go:117] "RemoveContainer" containerID="efd6def30ea2b205dacfcda73da7b1141d929863697261c072b99be752c3caab" Dec 07 17:06:36 crc kubenswrapper[4716]: I1207 17:06:36.658028 4716 scope.go:117] "RemoveContainer" containerID="0f9119a8d01b371a5ab06e6f944c6501025c243f659e26196309c46fc20c7a5d" Dec 07 17:06:36 crc kubenswrapper[4716]: E1207 17:06:36.658921 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 17:06:50 crc kubenswrapper[4716]: I1207 17:06:50.658378 4716 scope.go:117] "RemoveContainer" containerID="0f9119a8d01b371a5ab06e6f944c6501025c243f659e26196309c46fc20c7a5d" Dec 07 17:06:50 crc kubenswrapper[4716]: E1207 17:06:50.659403 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 17:07:03 crc kubenswrapper[4716]: I1207 17:07:03.657872 4716 scope.go:117] "RemoveContainer" containerID="0f9119a8d01b371a5ab06e6f944c6501025c243f659e26196309c46fc20c7a5d" Dec 07 17:07:03 crc kubenswrapper[4716]: E1207 17:07:03.658913 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 17:07:16 crc kubenswrapper[4716]: I1207 17:07:16.658104 4716 scope.go:117] "RemoveContainer" containerID="0f9119a8d01b371a5ab06e6f944c6501025c243f659e26196309c46fc20c7a5d" Dec 07 17:07:16 crc kubenswrapper[4716]: E1207 17:07:16.658957 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 17:07:30 crc kubenswrapper[4716]: I1207 17:07:30.658688 4716 scope.go:117] "RemoveContainer" containerID="0f9119a8d01b371a5ab06e6f944c6501025c243f659e26196309c46fc20c7a5d" Dec 07 17:07:30 crc kubenswrapper[4716]: E1207 17:07:30.659507 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 17:07:35 crc kubenswrapper[4716]: I1207 17:07:35.538152 4716 scope.go:117] "RemoveContainer" containerID="c251dea702de488ff8a2aab279c7651d0476faf38cd2a4eea7fc83db2d05ad2f" Dec 07 17:07:44 crc kubenswrapper[4716]: I1207 17:07:44.042478 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-x5c24/must-gather-tcwnq"] Dec 07 17:07:44 crc kubenswrapper[4716]: E1207 17:07:44.043317 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68a27046-ae9a-4651-98fd-d5633838e785" containerName="copy" Dec 07 17:07:44 crc kubenswrapper[4716]: I1207 17:07:44.043328 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="68a27046-ae9a-4651-98fd-d5633838e785" containerName="copy" Dec 07 17:07:44 crc kubenswrapper[4716]: E1207 17:07:44.043346 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68a27046-ae9a-4651-98fd-d5633838e785" containerName="gather" Dec 07 17:07:44 crc kubenswrapper[4716]: I1207 17:07:44.043351 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="68a27046-ae9a-4651-98fd-d5633838e785" containerName="gather" Dec 07 17:07:44 crc kubenswrapper[4716]: E1207 17:07:44.043360 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="778133cd-af64-46f4-a87c-9c69a513a73b" containerName="extract-utilities" Dec 07 17:07:44 crc kubenswrapper[4716]: I1207 17:07:44.043367 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="778133cd-af64-46f4-a87c-9c69a513a73b" containerName="extract-utilities" Dec 07 17:07:44 crc kubenswrapper[4716]: E1207 17:07:44.043387 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="778133cd-af64-46f4-a87c-9c69a513a73b" containerName="registry-server" Dec 07 17:07:44 crc kubenswrapper[4716]: I1207 17:07:44.043393 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="778133cd-af64-46f4-a87c-9c69a513a73b" containerName="registry-server" Dec 07 17:07:44 crc kubenswrapper[4716]: E1207 17:07:44.043406 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="778133cd-af64-46f4-a87c-9c69a513a73b" containerName="extract-content" Dec 07 17:07:44 crc kubenswrapper[4716]: I1207 17:07:44.043413 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="778133cd-af64-46f4-a87c-9c69a513a73b" containerName="extract-content" Dec 07 17:07:44 crc kubenswrapper[4716]: I1207 17:07:44.043588 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="68a27046-ae9a-4651-98fd-d5633838e785" containerName="copy" Dec 07 17:07:44 crc kubenswrapper[4716]: I1207 17:07:44.043603 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="68a27046-ae9a-4651-98fd-d5633838e785" containerName="gather" Dec 07 17:07:44 crc kubenswrapper[4716]: I1207 17:07:44.043616 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="778133cd-af64-46f4-a87c-9c69a513a73b" containerName="registry-server" Dec 07 17:07:44 crc kubenswrapper[4716]: I1207 17:07:44.044584 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x5c24/must-gather-tcwnq" Dec 07 17:07:44 crc kubenswrapper[4716]: I1207 17:07:44.047477 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-x5c24"/"kube-root-ca.crt" Dec 07 17:07:44 crc kubenswrapper[4716]: I1207 17:07:44.047620 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-x5c24"/"openshift-service-ca.crt" Dec 07 17:07:44 crc kubenswrapper[4716]: I1207 17:07:44.051801 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-x5c24"/"default-dockercfg-lzjcb" Dec 07 17:07:44 crc kubenswrapper[4716]: I1207 17:07:44.062599 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-x5c24/must-gather-tcwnq"] Dec 07 17:07:44 crc kubenswrapper[4716]: I1207 17:07:44.179992 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/6e77ed85-7b90-4d72-899c-13feb6b4615c-must-gather-output\") pod \"must-gather-tcwnq\" (UID: \"6e77ed85-7b90-4d72-899c-13feb6b4615c\") " pod="openshift-must-gather-x5c24/must-gather-tcwnq" Dec 07 17:07:44 crc kubenswrapper[4716]: I1207 17:07:44.180065 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wtftc\" (UniqueName: \"kubernetes.io/projected/6e77ed85-7b90-4d72-899c-13feb6b4615c-kube-api-access-wtftc\") pod \"must-gather-tcwnq\" (UID: \"6e77ed85-7b90-4d72-899c-13feb6b4615c\") " pod="openshift-must-gather-x5c24/must-gather-tcwnq" Dec 07 17:07:44 crc kubenswrapper[4716]: I1207 17:07:44.281975 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wtftc\" (UniqueName: \"kubernetes.io/projected/6e77ed85-7b90-4d72-899c-13feb6b4615c-kube-api-access-wtftc\") pod \"must-gather-tcwnq\" (UID: \"6e77ed85-7b90-4d72-899c-13feb6b4615c\") " pod="openshift-must-gather-x5c24/must-gather-tcwnq" Dec 07 17:07:44 crc kubenswrapper[4716]: I1207 17:07:44.282129 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/6e77ed85-7b90-4d72-899c-13feb6b4615c-must-gather-output\") pod \"must-gather-tcwnq\" (UID: \"6e77ed85-7b90-4d72-899c-13feb6b4615c\") " pod="openshift-must-gather-x5c24/must-gather-tcwnq" Dec 07 17:07:44 crc kubenswrapper[4716]: I1207 17:07:44.282458 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/6e77ed85-7b90-4d72-899c-13feb6b4615c-must-gather-output\") pod \"must-gather-tcwnq\" (UID: \"6e77ed85-7b90-4d72-899c-13feb6b4615c\") " pod="openshift-must-gather-x5c24/must-gather-tcwnq" Dec 07 17:07:44 crc kubenswrapper[4716]: I1207 17:07:44.302102 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wtftc\" (UniqueName: \"kubernetes.io/projected/6e77ed85-7b90-4d72-899c-13feb6b4615c-kube-api-access-wtftc\") pod \"must-gather-tcwnq\" (UID: \"6e77ed85-7b90-4d72-899c-13feb6b4615c\") " pod="openshift-must-gather-x5c24/must-gather-tcwnq" Dec 07 17:07:44 crc kubenswrapper[4716]: I1207 17:07:44.373934 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x5c24/must-gather-tcwnq" Dec 07 17:07:44 crc kubenswrapper[4716]: I1207 17:07:44.658151 4716 scope.go:117] "RemoveContainer" containerID="0f9119a8d01b371a5ab06e6f944c6501025c243f659e26196309c46fc20c7a5d" Dec 07 17:07:44 crc kubenswrapper[4716]: E1207 17:07:44.658700 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 17:07:44 crc kubenswrapper[4716]: I1207 17:07:44.813585 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-x5c24/must-gather-tcwnq"] Dec 07 17:07:45 crc kubenswrapper[4716]: I1207 17:07:45.750586 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x5c24/must-gather-tcwnq" event={"ID":"6e77ed85-7b90-4d72-899c-13feb6b4615c","Type":"ContainerStarted","Data":"d524df7d9598c1df472b479d6b590a726bfc512986882b2fd7ccb3c9ede5d1ab"} Dec 07 17:07:45 crc kubenswrapper[4716]: I1207 17:07:45.751104 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x5c24/must-gather-tcwnq" event={"ID":"6e77ed85-7b90-4d72-899c-13feb6b4615c","Type":"ContainerStarted","Data":"089983d99023ead848e9c460a5fefaaa5fbba5e6ae9dc60751cf8b8338fb5d97"} Dec 07 17:07:45 crc kubenswrapper[4716]: I1207 17:07:45.751116 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x5c24/must-gather-tcwnq" event={"ID":"6e77ed85-7b90-4d72-899c-13feb6b4615c","Type":"ContainerStarted","Data":"fbe75c5e3594738c68e8052845450297e1b5d9742d1524a7914ad67e20141c71"} Dec 07 17:07:45 crc kubenswrapper[4716]: I1207 17:07:45.765378 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-x5c24/must-gather-tcwnq" podStartSLOduration=1.765362667 podStartE2EDuration="1.765362667s" podCreationTimestamp="2025-12-07 17:07:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 17:07:45.762668544 +0000 UTC m=+3928.452953456" watchObservedRunningTime="2025-12-07 17:07:45.765362667 +0000 UTC m=+3928.455647579" Dec 07 17:07:48 crc kubenswrapper[4716]: I1207 17:07:48.697965 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-x5c24/crc-debug-jth5d"] Dec 07 17:07:48 crc kubenswrapper[4716]: I1207 17:07:48.699713 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x5c24/crc-debug-jth5d" Dec 07 17:07:48 crc kubenswrapper[4716]: I1207 17:07:48.858698 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/10d1bc7d-b49c-4221-87ae-4318d01367f2-host\") pod \"crc-debug-jth5d\" (UID: \"10d1bc7d-b49c-4221-87ae-4318d01367f2\") " pod="openshift-must-gather-x5c24/crc-debug-jth5d" Dec 07 17:07:48 crc kubenswrapper[4716]: I1207 17:07:48.859091 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6dz2k\" (UniqueName: \"kubernetes.io/projected/10d1bc7d-b49c-4221-87ae-4318d01367f2-kube-api-access-6dz2k\") pod \"crc-debug-jth5d\" (UID: \"10d1bc7d-b49c-4221-87ae-4318d01367f2\") " pod="openshift-must-gather-x5c24/crc-debug-jth5d" Dec 07 17:07:48 crc kubenswrapper[4716]: I1207 17:07:48.960865 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6dz2k\" (UniqueName: \"kubernetes.io/projected/10d1bc7d-b49c-4221-87ae-4318d01367f2-kube-api-access-6dz2k\") pod \"crc-debug-jth5d\" (UID: \"10d1bc7d-b49c-4221-87ae-4318d01367f2\") " pod="openshift-must-gather-x5c24/crc-debug-jth5d" Dec 07 17:07:48 crc kubenswrapper[4716]: I1207 17:07:48.960997 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/10d1bc7d-b49c-4221-87ae-4318d01367f2-host\") pod \"crc-debug-jth5d\" (UID: \"10d1bc7d-b49c-4221-87ae-4318d01367f2\") " pod="openshift-must-gather-x5c24/crc-debug-jth5d" Dec 07 17:07:48 crc kubenswrapper[4716]: I1207 17:07:48.961070 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/10d1bc7d-b49c-4221-87ae-4318d01367f2-host\") pod \"crc-debug-jth5d\" (UID: \"10d1bc7d-b49c-4221-87ae-4318d01367f2\") " pod="openshift-must-gather-x5c24/crc-debug-jth5d" Dec 07 17:07:48 crc kubenswrapper[4716]: I1207 17:07:48.980172 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6dz2k\" (UniqueName: \"kubernetes.io/projected/10d1bc7d-b49c-4221-87ae-4318d01367f2-kube-api-access-6dz2k\") pod \"crc-debug-jth5d\" (UID: \"10d1bc7d-b49c-4221-87ae-4318d01367f2\") " pod="openshift-must-gather-x5c24/crc-debug-jth5d" Dec 07 17:07:49 crc kubenswrapper[4716]: I1207 17:07:49.018387 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x5c24/crc-debug-jth5d" Dec 07 17:07:49 crc kubenswrapper[4716]: W1207 17:07:49.076820 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod10d1bc7d_b49c_4221_87ae_4318d01367f2.slice/crio-a3837af8c9589404ec6327a27fc5607b7ac2e6defb1c770b13c7b30caeec7cd5 WatchSource:0}: Error finding container a3837af8c9589404ec6327a27fc5607b7ac2e6defb1c770b13c7b30caeec7cd5: Status 404 returned error can't find the container with id a3837af8c9589404ec6327a27fc5607b7ac2e6defb1c770b13c7b30caeec7cd5 Dec 07 17:07:49 crc kubenswrapper[4716]: I1207 17:07:49.788683 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x5c24/crc-debug-jth5d" event={"ID":"10d1bc7d-b49c-4221-87ae-4318d01367f2","Type":"ContainerStarted","Data":"cee5e066fc428f048cddcdf7009e2528466a64db0dc946ab9d354efbefe3967b"} Dec 07 17:07:49 crc kubenswrapper[4716]: I1207 17:07:49.789352 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x5c24/crc-debug-jth5d" event={"ID":"10d1bc7d-b49c-4221-87ae-4318d01367f2","Type":"ContainerStarted","Data":"a3837af8c9589404ec6327a27fc5607b7ac2e6defb1c770b13c7b30caeec7cd5"} Dec 07 17:07:59 crc kubenswrapper[4716]: I1207 17:07:59.658494 4716 scope.go:117] "RemoveContainer" containerID="0f9119a8d01b371a5ab06e6f944c6501025c243f659e26196309c46fc20c7a5d" Dec 07 17:07:59 crc kubenswrapper[4716]: E1207 17:07:59.659499 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 17:08:09 crc kubenswrapper[4716]: I1207 17:08:09.406658 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-x5c24/crc-debug-jth5d" podStartSLOduration=21.406637353 podStartE2EDuration="21.406637353s" podCreationTimestamp="2025-12-07 17:07:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-07 17:07:49.801876002 +0000 UTC m=+3932.492160914" watchObservedRunningTime="2025-12-07 17:08:09.406637353 +0000 UTC m=+3952.096922265" Dec 07 17:08:09 crc kubenswrapper[4716]: I1207 17:08:09.412720 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-bcw8l"] Dec 07 17:08:09 crc kubenswrapper[4716]: I1207 17:08:09.415362 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bcw8l" Dec 07 17:08:09 crc kubenswrapper[4716]: I1207 17:08:09.427334 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bcw8l"] Dec 07 17:08:09 crc kubenswrapper[4716]: I1207 17:08:09.547983 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16f66cf5-edc0-432b-b7a9-179a750e95e1-catalog-content\") pod \"redhat-operators-bcw8l\" (UID: \"16f66cf5-edc0-432b-b7a9-179a750e95e1\") " pod="openshift-marketplace/redhat-operators-bcw8l" Dec 07 17:08:09 crc kubenswrapper[4716]: I1207 17:08:09.548399 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16f66cf5-edc0-432b-b7a9-179a750e95e1-utilities\") pod \"redhat-operators-bcw8l\" (UID: \"16f66cf5-edc0-432b-b7a9-179a750e95e1\") " pod="openshift-marketplace/redhat-operators-bcw8l" Dec 07 17:08:09 crc kubenswrapper[4716]: I1207 17:08:09.548475 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6jdl\" (UniqueName: \"kubernetes.io/projected/16f66cf5-edc0-432b-b7a9-179a750e95e1-kube-api-access-x6jdl\") pod \"redhat-operators-bcw8l\" (UID: \"16f66cf5-edc0-432b-b7a9-179a750e95e1\") " pod="openshift-marketplace/redhat-operators-bcw8l" Dec 07 17:08:09 crc kubenswrapper[4716]: I1207 17:08:09.649730 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16f66cf5-edc0-432b-b7a9-179a750e95e1-catalog-content\") pod \"redhat-operators-bcw8l\" (UID: \"16f66cf5-edc0-432b-b7a9-179a750e95e1\") " pod="openshift-marketplace/redhat-operators-bcw8l" Dec 07 17:08:09 crc kubenswrapper[4716]: I1207 17:08:09.649894 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16f66cf5-edc0-432b-b7a9-179a750e95e1-utilities\") pod \"redhat-operators-bcw8l\" (UID: \"16f66cf5-edc0-432b-b7a9-179a750e95e1\") " pod="openshift-marketplace/redhat-operators-bcw8l" Dec 07 17:08:09 crc kubenswrapper[4716]: I1207 17:08:09.649974 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6jdl\" (UniqueName: \"kubernetes.io/projected/16f66cf5-edc0-432b-b7a9-179a750e95e1-kube-api-access-x6jdl\") pod \"redhat-operators-bcw8l\" (UID: \"16f66cf5-edc0-432b-b7a9-179a750e95e1\") " pod="openshift-marketplace/redhat-operators-bcw8l" Dec 07 17:08:09 crc kubenswrapper[4716]: I1207 17:08:09.650288 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16f66cf5-edc0-432b-b7a9-179a750e95e1-catalog-content\") pod \"redhat-operators-bcw8l\" (UID: \"16f66cf5-edc0-432b-b7a9-179a750e95e1\") " pod="openshift-marketplace/redhat-operators-bcw8l" Dec 07 17:08:09 crc kubenswrapper[4716]: I1207 17:08:09.650360 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16f66cf5-edc0-432b-b7a9-179a750e95e1-utilities\") pod \"redhat-operators-bcw8l\" (UID: \"16f66cf5-edc0-432b-b7a9-179a750e95e1\") " pod="openshift-marketplace/redhat-operators-bcw8l" Dec 07 17:08:09 crc kubenswrapper[4716]: I1207 17:08:09.685316 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6jdl\" (UniqueName: \"kubernetes.io/projected/16f66cf5-edc0-432b-b7a9-179a750e95e1-kube-api-access-x6jdl\") pod \"redhat-operators-bcw8l\" (UID: \"16f66cf5-edc0-432b-b7a9-179a750e95e1\") " pod="openshift-marketplace/redhat-operators-bcw8l" Dec 07 17:08:09 crc kubenswrapper[4716]: I1207 17:08:09.783799 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bcw8l" Dec 07 17:08:10 crc kubenswrapper[4716]: I1207 17:08:10.338836 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bcw8l"] Dec 07 17:08:10 crc kubenswrapper[4716]: I1207 17:08:10.658303 4716 scope.go:117] "RemoveContainer" containerID="0f9119a8d01b371a5ab06e6f944c6501025c243f659e26196309c46fc20c7a5d" Dec 07 17:08:10 crc kubenswrapper[4716]: E1207 17:08:10.658585 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 17:08:10 crc kubenswrapper[4716]: I1207 17:08:10.978601 4716 generic.go:334] "Generic (PLEG): container finished" podID="16f66cf5-edc0-432b-b7a9-179a750e95e1" containerID="95bae6adc41154455bc811d8db543b56dc81c68814d8f78ebef60cab4940345b" exitCode=0 Dec 07 17:08:10 crc kubenswrapper[4716]: I1207 17:08:10.978717 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bcw8l" event={"ID":"16f66cf5-edc0-432b-b7a9-179a750e95e1","Type":"ContainerDied","Data":"95bae6adc41154455bc811d8db543b56dc81c68814d8f78ebef60cab4940345b"} Dec 07 17:08:10 crc kubenswrapper[4716]: I1207 17:08:10.979196 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bcw8l" event={"ID":"16f66cf5-edc0-432b-b7a9-179a750e95e1","Type":"ContainerStarted","Data":"75e78d987d77b8184a577b64ec41682b6ab25f1f211f9e465d210efa83ae5988"} Dec 07 17:08:11 crc kubenswrapper[4716]: I1207 17:08:11.988208 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bcw8l" event={"ID":"16f66cf5-edc0-432b-b7a9-179a750e95e1","Type":"ContainerStarted","Data":"16cda4989cfc2485735dc8144701ba792ad633d787296676cb3d48f29da764bb"} Dec 07 17:08:12 crc kubenswrapper[4716]: I1207 17:08:12.998678 4716 generic.go:334] "Generic (PLEG): container finished" podID="16f66cf5-edc0-432b-b7a9-179a750e95e1" containerID="16cda4989cfc2485735dc8144701ba792ad633d787296676cb3d48f29da764bb" exitCode=0 Dec 07 17:08:12 crc kubenswrapper[4716]: I1207 17:08:12.998880 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bcw8l" event={"ID":"16f66cf5-edc0-432b-b7a9-179a750e95e1","Type":"ContainerDied","Data":"16cda4989cfc2485735dc8144701ba792ad633d787296676cb3d48f29da764bb"} Dec 07 17:08:14 crc kubenswrapper[4716]: I1207 17:08:14.014788 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bcw8l" event={"ID":"16f66cf5-edc0-432b-b7a9-179a750e95e1","Type":"ContainerStarted","Data":"be92928d43c5da5e8b4289c6cbff430ff6915d61c7d7ace27e2e95ae0246c46e"} Dec 07 17:08:14 crc kubenswrapper[4716]: I1207 17:08:14.052204 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-bcw8l" podStartSLOduration=2.6476501580000003 podStartE2EDuration="5.052187787s" podCreationTimestamp="2025-12-07 17:08:09 +0000 UTC" firstStartedPulling="2025-12-07 17:08:10.980920678 +0000 UTC m=+3953.671205590" lastFinishedPulling="2025-12-07 17:08:13.385458307 +0000 UTC m=+3956.075743219" observedRunningTime="2025-12-07 17:08:14.042635368 +0000 UTC m=+3956.732920290" watchObservedRunningTime="2025-12-07 17:08:14.052187787 +0000 UTC m=+3956.742472709" Dec 07 17:08:19 crc kubenswrapper[4716]: I1207 17:08:19.785580 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-bcw8l" Dec 07 17:08:19 crc kubenswrapper[4716]: I1207 17:08:19.786242 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-bcw8l" Dec 07 17:08:19 crc kubenswrapper[4716]: I1207 17:08:19.844503 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-bcw8l" Dec 07 17:08:20 crc kubenswrapper[4716]: I1207 17:08:20.111708 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-bcw8l" Dec 07 17:08:20 crc kubenswrapper[4716]: I1207 17:08:20.168548 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bcw8l"] Dec 07 17:08:22 crc kubenswrapper[4716]: I1207 17:08:22.079352 4716 generic.go:334] "Generic (PLEG): container finished" podID="10d1bc7d-b49c-4221-87ae-4318d01367f2" containerID="cee5e066fc428f048cddcdf7009e2528466a64db0dc946ab9d354efbefe3967b" exitCode=0 Dec 07 17:08:22 crc kubenswrapper[4716]: I1207 17:08:22.079420 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x5c24/crc-debug-jth5d" event={"ID":"10d1bc7d-b49c-4221-87ae-4318d01367f2","Type":"ContainerDied","Data":"cee5e066fc428f048cddcdf7009e2528466a64db0dc946ab9d354efbefe3967b"} Dec 07 17:08:22 crc kubenswrapper[4716]: I1207 17:08:22.080314 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-bcw8l" podUID="16f66cf5-edc0-432b-b7a9-179a750e95e1" containerName="registry-server" containerID="cri-o://be92928d43c5da5e8b4289c6cbff430ff6915d61c7d7ace27e2e95ae0246c46e" gracePeriod=2 Dec 07 17:08:22 crc kubenswrapper[4716]: I1207 17:08:22.782850 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bcw8l" Dec 07 17:08:22 crc kubenswrapper[4716]: I1207 17:08:22.818818 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x6jdl\" (UniqueName: \"kubernetes.io/projected/16f66cf5-edc0-432b-b7a9-179a750e95e1-kube-api-access-x6jdl\") pod \"16f66cf5-edc0-432b-b7a9-179a750e95e1\" (UID: \"16f66cf5-edc0-432b-b7a9-179a750e95e1\") " Dec 07 17:08:22 crc kubenswrapper[4716]: I1207 17:08:22.818993 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16f66cf5-edc0-432b-b7a9-179a750e95e1-catalog-content\") pod \"16f66cf5-edc0-432b-b7a9-179a750e95e1\" (UID: \"16f66cf5-edc0-432b-b7a9-179a750e95e1\") " Dec 07 17:08:22 crc kubenswrapper[4716]: I1207 17:08:22.819113 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16f66cf5-edc0-432b-b7a9-179a750e95e1-utilities\") pod \"16f66cf5-edc0-432b-b7a9-179a750e95e1\" (UID: \"16f66cf5-edc0-432b-b7a9-179a750e95e1\") " Dec 07 17:08:22 crc kubenswrapper[4716]: I1207 17:08:22.820673 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/16f66cf5-edc0-432b-b7a9-179a750e95e1-utilities" (OuterVolumeSpecName: "utilities") pod "16f66cf5-edc0-432b-b7a9-179a750e95e1" (UID: "16f66cf5-edc0-432b-b7a9-179a750e95e1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 17:08:22 crc kubenswrapper[4716]: I1207 17:08:22.828525 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16f66cf5-edc0-432b-b7a9-179a750e95e1-kube-api-access-x6jdl" (OuterVolumeSpecName: "kube-api-access-x6jdl") pod "16f66cf5-edc0-432b-b7a9-179a750e95e1" (UID: "16f66cf5-edc0-432b-b7a9-179a750e95e1"). InnerVolumeSpecName "kube-api-access-x6jdl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 17:08:22 crc kubenswrapper[4716]: I1207 17:08:22.921174 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x6jdl\" (UniqueName: \"kubernetes.io/projected/16f66cf5-edc0-432b-b7a9-179a750e95e1-kube-api-access-x6jdl\") on node \"crc\" DevicePath \"\"" Dec 07 17:08:22 crc kubenswrapper[4716]: I1207 17:08:22.921449 4716 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16f66cf5-edc0-432b-b7a9-179a750e95e1-utilities\") on node \"crc\" DevicePath \"\"" Dec 07 17:08:23 crc kubenswrapper[4716]: I1207 17:08:23.090735 4716 generic.go:334] "Generic (PLEG): container finished" podID="16f66cf5-edc0-432b-b7a9-179a750e95e1" containerID="be92928d43c5da5e8b4289c6cbff430ff6915d61c7d7ace27e2e95ae0246c46e" exitCode=0 Dec 07 17:08:23 crc kubenswrapper[4716]: I1207 17:08:23.090808 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bcw8l" Dec 07 17:08:23 crc kubenswrapper[4716]: I1207 17:08:23.090792 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bcw8l" event={"ID":"16f66cf5-edc0-432b-b7a9-179a750e95e1","Type":"ContainerDied","Data":"be92928d43c5da5e8b4289c6cbff430ff6915d61c7d7ace27e2e95ae0246c46e"} Dec 07 17:08:23 crc kubenswrapper[4716]: I1207 17:08:23.090902 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bcw8l" event={"ID":"16f66cf5-edc0-432b-b7a9-179a750e95e1","Type":"ContainerDied","Data":"75e78d987d77b8184a577b64ec41682b6ab25f1f211f9e465d210efa83ae5988"} Dec 07 17:08:23 crc kubenswrapper[4716]: I1207 17:08:23.090951 4716 scope.go:117] "RemoveContainer" containerID="be92928d43c5da5e8b4289c6cbff430ff6915d61c7d7ace27e2e95ae0246c46e" Dec 07 17:08:23 crc kubenswrapper[4716]: I1207 17:08:23.496446 4716 scope.go:117] "RemoveContainer" containerID="16cda4989cfc2485735dc8144701ba792ad633d787296676cb3d48f29da764bb" Dec 07 17:08:23 crc kubenswrapper[4716]: I1207 17:08:23.631539 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x5c24/crc-debug-jth5d" Dec 07 17:08:23 crc kubenswrapper[4716]: I1207 17:08:23.642434 4716 scope.go:117] "RemoveContainer" containerID="95bae6adc41154455bc811d8db543b56dc81c68814d8f78ebef60cab4940345b" Dec 07 17:08:23 crc kubenswrapper[4716]: I1207 17:08:23.664420 4716 scope.go:117] "RemoveContainer" containerID="0f9119a8d01b371a5ab06e6f944c6501025c243f659e26196309c46fc20c7a5d" Dec 07 17:08:23 crc kubenswrapper[4716]: E1207 17:08:23.664689 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 17:08:23 crc kubenswrapper[4716]: I1207 17:08:23.686214 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-x5c24/crc-debug-jth5d"] Dec 07 17:08:23 crc kubenswrapper[4716]: I1207 17:08:23.692433 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-x5c24/crc-debug-jth5d"] Dec 07 17:08:23 crc kubenswrapper[4716]: I1207 17:08:23.702916 4716 scope.go:117] "RemoveContainer" containerID="be92928d43c5da5e8b4289c6cbff430ff6915d61c7d7ace27e2e95ae0246c46e" Dec 07 17:08:23 crc kubenswrapper[4716]: E1207 17:08:23.703384 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be92928d43c5da5e8b4289c6cbff430ff6915d61c7d7ace27e2e95ae0246c46e\": container with ID starting with be92928d43c5da5e8b4289c6cbff430ff6915d61c7d7ace27e2e95ae0246c46e not found: ID does not exist" containerID="be92928d43c5da5e8b4289c6cbff430ff6915d61c7d7ace27e2e95ae0246c46e" Dec 07 17:08:23 crc kubenswrapper[4716]: I1207 17:08:23.703453 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be92928d43c5da5e8b4289c6cbff430ff6915d61c7d7ace27e2e95ae0246c46e"} err="failed to get container status \"be92928d43c5da5e8b4289c6cbff430ff6915d61c7d7ace27e2e95ae0246c46e\": rpc error: code = NotFound desc = could not find container \"be92928d43c5da5e8b4289c6cbff430ff6915d61c7d7ace27e2e95ae0246c46e\": container with ID starting with be92928d43c5da5e8b4289c6cbff430ff6915d61c7d7ace27e2e95ae0246c46e not found: ID does not exist" Dec 07 17:08:23 crc kubenswrapper[4716]: I1207 17:08:23.703481 4716 scope.go:117] "RemoveContainer" containerID="16cda4989cfc2485735dc8144701ba792ad633d787296676cb3d48f29da764bb" Dec 07 17:08:23 crc kubenswrapper[4716]: E1207 17:08:23.703749 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"16cda4989cfc2485735dc8144701ba792ad633d787296676cb3d48f29da764bb\": container with ID starting with 16cda4989cfc2485735dc8144701ba792ad633d787296676cb3d48f29da764bb not found: ID does not exist" containerID="16cda4989cfc2485735dc8144701ba792ad633d787296676cb3d48f29da764bb" Dec 07 17:08:23 crc kubenswrapper[4716]: I1207 17:08:23.703770 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16cda4989cfc2485735dc8144701ba792ad633d787296676cb3d48f29da764bb"} err="failed to get container status \"16cda4989cfc2485735dc8144701ba792ad633d787296676cb3d48f29da764bb\": rpc error: code = NotFound desc = could not find container \"16cda4989cfc2485735dc8144701ba792ad633d787296676cb3d48f29da764bb\": container with ID starting with 16cda4989cfc2485735dc8144701ba792ad633d787296676cb3d48f29da764bb not found: ID does not exist" Dec 07 17:08:23 crc kubenswrapper[4716]: I1207 17:08:23.703783 4716 scope.go:117] "RemoveContainer" containerID="95bae6adc41154455bc811d8db543b56dc81c68814d8f78ebef60cab4940345b" Dec 07 17:08:23 crc kubenswrapper[4716]: E1207 17:08:23.704052 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95bae6adc41154455bc811d8db543b56dc81c68814d8f78ebef60cab4940345b\": container with ID starting with 95bae6adc41154455bc811d8db543b56dc81c68814d8f78ebef60cab4940345b not found: ID does not exist" containerID="95bae6adc41154455bc811d8db543b56dc81c68814d8f78ebef60cab4940345b" Dec 07 17:08:23 crc kubenswrapper[4716]: I1207 17:08:23.704284 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95bae6adc41154455bc811d8db543b56dc81c68814d8f78ebef60cab4940345b"} err="failed to get container status \"95bae6adc41154455bc811d8db543b56dc81c68814d8f78ebef60cab4940345b\": rpc error: code = NotFound desc = could not find container \"95bae6adc41154455bc811d8db543b56dc81c68814d8f78ebef60cab4940345b\": container with ID starting with 95bae6adc41154455bc811d8db543b56dc81c68814d8f78ebef60cab4940345b not found: ID does not exist" Dec 07 17:08:23 crc kubenswrapper[4716]: I1207 17:08:23.737180 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/10d1bc7d-b49c-4221-87ae-4318d01367f2-host\") pod \"10d1bc7d-b49c-4221-87ae-4318d01367f2\" (UID: \"10d1bc7d-b49c-4221-87ae-4318d01367f2\") " Dec 07 17:08:23 crc kubenswrapper[4716]: I1207 17:08:23.737271 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/10d1bc7d-b49c-4221-87ae-4318d01367f2-host" (OuterVolumeSpecName: "host") pod "10d1bc7d-b49c-4221-87ae-4318d01367f2" (UID: "10d1bc7d-b49c-4221-87ae-4318d01367f2"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 17:08:23 crc kubenswrapper[4716]: I1207 17:08:23.737599 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6dz2k\" (UniqueName: \"kubernetes.io/projected/10d1bc7d-b49c-4221-87ae-4318d01367f2-kube-api-access-6dz2k\") pod \"10d1bc7d-b49c-4221-87ae-4318d01367f2\" (UID: \"10d1bc7d-b49c-4221-87ae-4318d01367f2\") " Dec 07 17:08:23 crc kubenswrapper[4716]: I1207 17:08:23.739514 4716 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/10d1bc7d-b49c-4221-87ae-4318d01367f2-host\") on node \"crc\" DevicePath \"\"" Dec 07 17:08:23 crc kubenswrapper[4716]: I1207 17:08:23.743495 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10d1bc7d-b49c-4221-87ae-4318d01367f2-kube-api-access-6dz2k" (OuterVolumeSpecName: "kube-api-access-6dz2k") pod "10d1bc7d-b49c-4221-87ae-4318d01367f2" (UID: "10d1bc7d-b49c-4221-87ae-4318d01367f2"). InnerVolumeSpecName "kube-api-access-6dz2k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 17:08:23 crc kubenswrapper[4716]: I1207 17:08:23.807366 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/16f66cf5-edc0-432b-b7a9-179a750e95e1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "16f66cf5-edc0-432b-b7a9-179a750e95e1" (UID: "16f66cf5-edc0-432b-b7a9-179a750e95e1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 17:08:23 crc kubenswrapper[4716]: I1207 17:08:23.841766 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6dz2k\" (UniqueName: \"kubernetes.io/projected/10d1bc7d-b49c-4221-87ae-4318d01367f2-kube-api-access-6dz2k\") on node \"crc\" DevicePath \"\"" Dec 07 17:08:23 crc kubenswrapper[4716]: I1207 17:08:23.841801 4716 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16f66cf5-edc0-432b-b7a9-179a750e95e1-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 07 17:08:24 crc kubenswrapper[4716]: I1207 17:08:24.026817 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bcw8l"] Dec 07 17:08:24 crc kubenswrapper[4716]: I1207 17:08:24.040902 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-bcw8l"] Dec 07 17:08:24 crc kubenswrapper[4716]: I1207 17:08:24.101147 4716 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a3837af8c9589404ec6327a27fc5607b7ac2e6defb1c770b13c7b30caeec7cd5" Dec 07 17:08:24 crc kubenswrapper[4716]: I1207 17:08:24.101203 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x5c24/crc-debug-jth5d" Dec 07 17:08:24 crc kubenswrapper[4716]: I1207 17:08:24.901590 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-x5c24/crc-debug-fdscl"] Dec 07 17:08:24 crc kubenswrapper[4716]: E1207 17:08:24.901989 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16f66cf5-edc0-432b-b7a9-179a750e95e1" containerName="registry-server" Dec 07 17:08:24 crc kubenswrapper[4716]: I1207 17:08:24.902004 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="16f66cf5-edc0-432b-b7a9-179a750e95e1" containerName="registry-server" Dec 07 17:08:24 crc kubenswrapper[4716]: E1207 17:08:24.902024 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10d1bc7d-b49c-4221-87ae-4318d01367f2" containerName="container-00" Dec 07 17:08:24 crc kubenswrapper[4716]: I1207 17:08:24.902031 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="10d1bc7d-b49c-4221-87ae-4318d01367f2" containerName="container-00" Dec 07 17:08:24 crc kubenswrapper[4716]: E1207 17:08:24.902043 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16f66cf5-edc0-432b-b7a9-179a750e95e1" containerName="extract-content" Dec 07 17:08:24 crc kubenswrapper[4716]: I1207 17:08:24.902049 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="16f66cf5-edc0-432b-b7a9-179a750e95e1" containerName="extract-content" Dec 07 17:08:24 crc kubenswrapper[4716]: E1207 17:08:24.902056 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16f66cf5-edc0-432b-b7a9-179a750e95e1" containerName="extract-utilities" Dec 07 17:08:24 crc kubenswrapper[4716]: I1207 17:08:24.902061 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="16f66cf5-edc0-432b-b7a9-179a750e95e1" containerName="extract-utilities" Dec 07 17:08:24 crc kubenswrapper[4716]: I1207 17:08:24.902264 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="10d1bc7d-b49c-4221-87ae-4318d01367f2" containerName="container-00" Dec 07 17:08:24 crc kubenswrapper[4716]: I1207 17:08:24.902290 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="16f66cf5-edc0-432b-b7a9-179a750e95e1" containerName="registry-server" Dec 07 17:08:24 crc kubenswrapper[4716]: I1207 17:08:24.902885 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x5c24/crc-debug-fdscl" Dec 07 17:08:24 crc kubenswrapper[4716]: I1207 17:08:24.998912 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rw2n4\" (UniqueName: \"kubernetes.io/projected/97972967-79eb-49ed-917a-4559ef0c9e65-kube-api-access-rw2n4\") pod \"crc-debug-fdscl\" (UID: \"97972967-79eb-49ed-917a-4559ef0c9e65\") " pod="openshift-must-gather-x5c24/crc-debug-fdscl" Dec 07 17:08:24 crc kubenswrapper[4716]: I1207 17:08:24.999213 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/97972967-79eb-49ed-917a-4559ef0c9e65-host\") pod \"crc-debug-fdscl\" (UID: \"97972967-79eb-49ed-917a-4559ef0c9e65\") " pod="openshift-must-gather-x5c24/crc-debug-fdscl" Dec 07 17:08:25 crc kubenswrapper[4716]: I1207 17:08:25.101387 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rw2n4\" (UniqueName: \"kubernetes.io/projected/97972967-79eb-49ed-917a-4559ef0c9e65-kube-api-access-rw2n4\") pod \"crc-debug-fdscl\" (UID: \"97972967-79eb-49ed-917a-4559ef0c9e65\") " pod="openshift-must-gather-x5c24/crc-debug-fdscl" Dec 07 17:08:25 crc kubenswrapper[4716]: I1207 17:08:25.101761 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/97972967-79eb-49ed-917a-4559ef0c9e65-host\") pod \"crc-debug-fdscl\" (UID: \"97972967-79eb-49ed-917a-4559ef0c9e65\") " pod="openshift-must-gather-x5c24/crc-debug-fdscl" Dec 07 17:08:25 crc kubenswrapper[4716]: I1207 17:08:25.101893 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/97972967-79eb-49ed-917a-4559ef0c9e65-host\") pod \"crc-debug-fdscl\" (UID: \"97972967-79eb-49ed-917a-4559ef0c9e65\") " pod="openshift-must-gather-x5c24/crc-debug-fdscl" Dec 07 17:08:25 crc kubenswrapper[4716]: I1207 17:08:25.475514 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rw2n4\" (UniqueName: \"kubernetes.io/projected/97972967-79eb-49ed-917a-4559ef0c9e65-kube-api-access-rw2n4\") pod \"crc-debug-fdscl\" (UID: \"97972967-79eb-49ed-917a-4559ef0c9e65\") " pod="openshift-must-gather-x5c24/crc-debug-fdscl" Dec 07 17:08:25 crc kubenswrapper[4716]: I1207 17:08:25.521738 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x5c24/crc-debug-fdscl" Dec 07 17:08:25 crc kubenswrapper[4716]: W1207 17:08:25.544855 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod97972967_79eb_49ed_917a_4559ef0c9e65.slice/crio-7e08f5b3c9d2c6fa26044101553f98a62671770f33c43ce760b9ff3d26a4c192 WatchSource:0}: Error finding container 7e08f5b3c9d2c6fa26044101553f98a62671770f33c43ce760b9ff3d26a4c192: Status 404 returned error can't find the container with id 7e08f5b3c9d2c6fa26044101553f98a62671770f33c43ce760b9ff3d26a4c192 Dec 07 17:08:25 crc kubenswrapper[4716]: I1207 17:08:25.666779 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10d1bc7d-b49c-4221-87ae-4318d01367f2" path="/var/lib/kubelet/pods/10d1bc7d-b49c-4221-87ae-4318d01367f2/volumes" Dec 07 17:08:25 crc kubenswrapper[4716]: I1207 17:08:25.667295 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16f66cf5-edc0-432b-b7a9-179a750e95e1" path="/var/lib/kubelet/pods/16f66cf5-edc0-432b-b7a9-179a750e95e1/volumes" Dec 07 17:08:26 crc kubenswrapper[4716]: I1207 17:08:26.118272 4716 generic.go:334] "Generic (PLEG): container finished" podID="97972967-79eb-49ed-917a-4559ef0c9e65" containerID="3897d5a11286928aa04f1ec34443f3c50369fb0d76edb44e2759cee0a15b4760" exitCode=0 Dec 07 17:08:26 crc kubenswrapper[4716]: I1207 17:08:26.118363 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x5c24/crc-debug-fdscl" event={"ID":"97972967-79eb-49ed-917a-4559ef0c9e65","Type":"ContainerDied","Data":"3897d5a11286928aa04f1ec34443f3c50369fb0d76edb44e2759cee0a15b4760"} Dec 07 17:08:26 crc kubenswrapper[4716]: I1207 17:08:26.118561 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x5c24/crc-debug-fdscl" event={"ID":"97972967-79eb-49ed-917a-4559ef0c9e65","Type":"ContainerStarted","Data":"7e08f5b3c9d2c6fa26044101553f98a62671770f33c43ce760b9ff3d26a4c192"} Dec 07 17:08:26 crc kubenswrapper[4716]: I1207 17:08:26.502644 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-x5c24/crc-debug-fdscl"] Dec 07 17:08:26 crc kubenswrapper[4716]: I1207 17:08:26.510600 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-x5c24/crc-debug-fdscl"] Dec 07 17:08:27 crc kubenswrapper[4716]: I1207 17:08:27.214025 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x5c24/crc-debug-fdscl" Dec 07 17:08:27 crc kubenswrapper[4716]: I1207 17:08:27.338045 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/97972967-79eb-49ed-917a-4559ef0c9e65-host\") pod \"97972967-79eb-49ed-917a-4559ef0c9e65\" (UID: \"97972967-79eb-49ed-917a-4559ef0c9e65\") " Dec 07 17:08:27 crc kubenswrapper[4716]: I1207 17:08:27.338141 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/97972967-79eb-49ed-917a-4559ef0c9e65-host" (OuterVolumeSpecName: "host") pod "97972967-79eb-49ed-917a-4559ef0c9e65" (UID: "97972967-79eb-49ed-917a-4559ef0c9e65"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 17:08:27 crc kubenswrapper[4716]: I1207 17:08:27.338694 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rw2n4\" (UniqueName: \"kubernetes.io/projected/97972967-79eb-49ed-917a-4559ef0c9e65-kube-api-access-rw2n4\") pod \"97972967-79eb-49ed-917a-4559ef0c9e65\" (UID: \"97972967-79eb-49ed-917a-4559ef0c9e65\") " Dec 07 17:08:27 crc kubenswrapper[4716]: I1207 17:08:27.339609 4716 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/97972967-79eb-49ed-917a-4559ef0c9e65-host\") on node \"crc\" DevicePath \"\"" Dec 07 17:08:27 crc kubenswrapper[4716]: I1207 17:08:27.344992 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97972967-79eb-49ed-917a-4559ef0c9e65-kube-api-access-rw2n4" (OuterVolumeSpecName: "kube-api-access-rw2n4") pod "97972967-79eb-49ed-917a-4559ef0c9e65" (UID: "97972967-79eb-49ed-917a-4559ef0c9e65"). InnerVolumeSpecName "kube-api-access-rw2n4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 17:08:27 crc kubenswrapper[4716]: I1207 17:08:27.441542 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rw2n4\" (UniqueName: \"kubernetes.io/projected/97972967-79eb-49ed-917a-4559ef0c9e65-kube-api-access-rw2n4\") on node \"crc\" DevicePath \"\"" Dec 07 17:08:27 crc kubenswrapper[4716]: I1207 17:08:27.652570 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-x5c24/crc-debug-ztkmp"] Dec 07 17:08:27 crc kubenswrapper[4716]: E1207 17:08:27.653508 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97972967-79eb-49ed-917a-4559ef0c9e65" containerName="container-00" Dec 07 17:08:27 crc kubenswrapper[4716]: I1207 17:08:27.653536 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="97972967-79eb-49ed-917a-4559ef0c9e65" containerName="container-00" Dec 07 17:08:27 crc kubenswrapper[4716]: I1207 17:08:27.653782 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="97972967-79eb-49ed-917a-4559ef0c9e65" containerName="container-00" Dec 07 17:08:27 crc kubenswrapper[4716]: I1207 17:08:27.654437 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x5c24/crc-debug-ztkmp" Dec 07 17:08:27 crc kubenswrapper[4716]: I1207 17:08:27.669143 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="97972967-79eb-49ed-917a-4559ef0c9e65" path="/var/lib/kubelet/pods/97972967-79eb-49ed-917a-4559ef0c9e65/volumes" Dec 07 17:08:27 crc kubenswrapper[4716]: I1207 17:08:27.745558 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/07c5f01b-87a0-43f6-b9ee-fe63a8659330-host\") pod \"crc-debug-ztkmp\" (UID: \"07c5f01b-87a0-43f6-b9ee-fe63a8659330\") " pod="openshift-must-gather-x5c24/crc-debug-ztkmp" Dec 07 17:08:27 crc kubenswrapper[4716]: I1207 17:08:27.745651 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjd9t\" (UniqueName: \"kubernetes.io/projected/07c5f01b-87a0-43f6-b9ee-fe63a8659330-kube-api-access-pjd9t\") pod \"crc-debug-ztkmp\" (UID: \"07c5f01b-87a0-43f6-b9ee-fe63a8659330\") " pod="openshift-must-gather-x5c24/crc-debug-ztkmp" Dec 07 17:08:27 crc kubenswrapper[4716]: I1207 17:08:27.847029 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/07c5f01b-87a0-43f6-b9ee-fe63a8659330-host\") pod \"crc-debug-ztkmp\" (UID: \"07c5f01b-87a0-43f6-b9ee-fe63a8659330\") " pod="openshift-must-gather-x5c24/crc-debug-ztkmp" Dec 07 17:08:27 crc kubenswrapper[4716]: I1207 17:08:27.847342 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjd9t\" (UniqueName: \"kubernetes.io/projected/07c5f01b-87a0-43f6-b9ee-fe63a8659330-kube-api-access-pjd9t\") pod \"crc-debug-ztkmp\" (UID: \"07c5f01b-87a0-43f6-b9ee-fe63a8659330\") " pod="openshift-must-gather-x5c24/crc-debug-ztkmp" Dec 07 17:08:27 crc kubenswrapper[4716]: I1207 17:08:27.847190 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/07c5f01b-87a0-43f6-b9ee-fe63a8659330-host\") pod \"crc-debug-ztkmp\" (UID: \"07c5f01b-87a0-43f6-b9ee-fe63a8659330\") " pod="openshift-must-gather-x5c24/crc-debug-ztkmp" Dec 07 17:08:27 crc kubenswrapper[4716]: I1207 17:08:27.870225 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjd9t\" (UniqueName: \"kubernetes.io/projected/07c5f01b-87a0-43f6-b9ee-fe63a8659330-kube-api-access-pjd9t\") pod \"crc-debug-ztkmp\" (UID: \"07c5f01b-87a0-43f6-b9ee-fe63a8659330\") " pod="openshift-must-gather-x5c24/crc-debug-ztkmp" Dec 07 17:08:27 crc kubenswrapper[4716]: I1207 17:08:27.973347 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x5c24/crc-debug-ztkmp" Dec 07 17:08:28 crc kubenswrapper[4716]: W1207 17:08:28.005193 4716 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod07c5f01b_87a0_43f6_b9ee_fe63a8659330.slice/crio-03f03375c23432719e78e2641876adc1e777f2ae262e586b10328cf12bbdacaa WatchSource:0}: Error finding container 03f03375c23432719e78e2641876adc1e777f2ae262e586b10328cf12bbdacaa: Status 404 returned error can't find the container with id 03f03375c23432719e78e2641876adc1e777f2ae262e586b10328cf12bbdacaa Dec 07 17:08:28 crc kubenswrapper[4716]: I1207 17:08:28.138999 4716 scope.go:117] "RemoveContainer" containerID="3897d5a11286928aa04f1ec34443f3c50369fb0d76edb44e2759cee0a15b4760" Dec 07 17:08:28 crc kubenswrapper[4716]: I1207 17:08:28.139105 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x5c24/crc-debug-fdscl" Dec 07 17:08:28 crc kubenswrapper[4716]: I1207 17:08:28.140519 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x5c24/crc-debug-ztkmp" event={"ID":"07c5f01b-87a0-43f6-b9ee-fe63a8659330","Type":"ContainerStarted","Data":"03f03375c23432719e78e2641876adc1e777f2ae262e586b10328cf12bbdacaa"} Dec 07 17:08:29 crc kubenswrapper[4716]: I1207 17:08:29.150315 4716 generic.go:334] "Generic (PLEG): container finished" podID="07c5f01b-87a0-43f6-b9ee-fe63a8659330" containerID="07f50e2064c5508e85bdfd02b88b8e9b9a0d9d597a339226f356252855c79440" exitCode=0 Dec 07 17:08:29 crc kubenswrapper[4716]: I1207 17:08:29.150376 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x5c24/crc-debug-ztkmp" event={"ID":"07c5f01b-87a0-43f6-b9ee-fe63a8659330","Type":"ContainerDied","Data":"07f50e2064c5508e85bdfd02b88b8e9b9a0d9d597a339226f356252855c79440"} Dec 07 17:08:29 crc kubenswrapper[4716]: I1207 17:08:29.191457 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-x5c24/crc-debug-ztkmp"] Dec 07 17:08:29 crc kubenswrapper[4716]: I1207 17:08:29.201757 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-x5c24/crc-debug-ztkmp"] Dec 07 17:08:30 crc kubenswrapper[4716]: I1207 17:08:30.267868 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x5c24/crc-debug-ztkmp" Dec 07 17:08:30 crc kubenswrapper[4716]: I1207 17:08:30.300426 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjd9t\" (UniqueName: \"kubernetes.io/projected/07c5f01b-87a0-43f6-b9ee-fe63a8659330-kube-api-access-pjd9t\") pod \"07c5f01b-87a0-43f6-b9ee-fe63a8659330\" (UID: \"07c5f01b-87a0-43f6-b9ee-fe63a8659330\") " Dec 07 17:08:30 crc kubenswrapper[4716]: I1207 17:08:30.300554 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/07c5f01b-87a0-43f6-b9ee-fe63a8659330-host\") pod \"07c5f01b-87a0-43f6-b9ee-fe63a8659330\" (UID: \"07c5f01b-87a0-43f6-b9ee-fe63a8659330\") " Dec 07 17:08:30 crc kubenswrapper[4716]: I1207 17:08:30.300751 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/07c5f01b-87a0-43f6-b9ee-fe63a8659330-host" (OuterVolumeSpecName: "host") pod "07c5f01b-87a0-43f6-b9ee-fe63a8659330" (UID: "07c5f01b-87a0-43f6-b9ee-fe63a8659330"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 07 17:08:30 crc kubenswrapper[4716]: I1207 17:08:30.301365 4716 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/07c5f01b-87a0-43f6-b9ee-fe63a8659330-host\") on node \"crc\" DevicePath \"\"" Dec 07 17:08:30 crc kubenswrapper[4716]: I1207 17:08:30.315162 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07c5f01b-87a0-43f6-b9ee-fe63a8659330-kube-api-access-pjd9t" (OuterVolumeSpecName: "kube-api-access-pjd9t") pod "07c5f01b-87a0-43f6-b9ee-fe63a8659330" (UID: "07c5f01b-87a0-43f6-b9ee-fe63a8659330"). InnerVolumeSpecName "kube-api-access-pjd9t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 17:08:30 crc kubenswrapper[4716]: I1207 17:08:30.402620 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjd9t\" (UniqueName: \"kubernetes.io/projected/07c5f01b-87a0-43f6-b9ee-fe63a8659330-kube-api-access-pjd9t\") on node \"crc\" DevicePath \"\"" Dec 07 17:08:31 crc kubenswrapper[4716]: I1207 17:08:31.174820 4716 scope.go:117] "RemoveContainer" containerID="07f50e2064c5508e85bdfd02b88b8e9b9a0d9d597a339226f356252855c79440" Dec 07 17:08:31 crc kubenswrapper[4716]: I1207 17:08:31.174993 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x5c24/crc-debug-ztkmp" Dec 07 17:08:31 crc kubenswrapper[4716]: I1207 17:08:31.671635 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="07c5f01b-87a0-43f6-b9ee-fe63a8659330" path="/var/lib/kubelet/pods/07c5f01b-87a0-43f6-b9ee-fe63a8659330/volumes" Dec 07 17:08:34 crc kubenswrapper[4716]: I1207 17:08:34.658366 4716 scope.go:117] "RemoveContainer" containerID="0f9119a8d01b371a5ab06e6f944c6501025c243f659e26196309c46fc20c7a5d" Dec 07 17:08:34 crc kubenswrapper[4716]: E1207 17:08:34.659051 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 17:08:45 crc kubenswrapper[4716]: I1207 17:08:45.658324 4716 scope.go:117] "RemoveContainer" containerID="0f9119a8d01b371a5ab06e6f944c6501025c243f659e26196309c46fc20c7a5d" Dec 07 17:08:45 crc kubenswrapper[4716]: E1207 17:08:45.659269 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 17:08:56 crc kubenswrapper[4716]: I1207 17:08:56.868773 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-64c898bd6-wgw8m_19efaa0a-3cda-4ef8-a845-bbcc78c747ec/barbican-api/0.log" Dec 07 17:08:56 crc kubenswrapper[4716]: I1207 17:08:56.978773 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-64c898bd6-wgw8m_19efaa0a-3cda-4ef8-a845-bbcc78c747ec/barbican-api-log/0.log" Dec 07 17:08:57 crc kubenswrapper[4716]: I1207 17:08:57.077448 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-75c8ccd7fb-bhzds_214231d4-11de-4827-afd3-e5169b138b7b/barbican-keystone-listener/0.log" Dec 07 17:08:57 crc kubenswrapper[4716]: I1207 17:08:57.136860 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-75c8ccd7fb-bhzds_214231d4-11de-4827-afd3-e5169b138b7b/barbican-keystone-listener-log/0.log" Dec 07 17:08:57 crc kubenswrapper[4716]: I1207 17:08:57.250369 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-587c8bb479-4d8m7_cc929236-8385-4f05-8ca5-37315e852be6/barbican-worker/0.log" Dec 07 17:08:57 crc kubenswrapper[4716]: I1207 17:08:57.262279 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-587c8bb479-4d8m7_cc929236-8385-4f05-8ca5-37315e852be6/barbican-worker-log/0.log" Dec 07 17:08:57 crc kubenswrapper[4716]: I1207 17:08:57.326067 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-7sr89"] Dec 07 17:08:57 crc kubenswrapper[4716]: E1207 17:08:57.326451 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07c5f01b-87a0-43f6-b9ee-fe63a8659330" containerName="container-00" Dec 07 17:08:57 crc kubenswrapper[4716]: I1207 17:08:57.326465 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="07c5f01b-87a0-43f6-b9ee-fe63a8659330" containerName="container-00" Dec 07 17:08:57 crc kubenswrapper[4716]: I1207 17:08:57.326658 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="07c5f01b-87a0-43f6-b9ee-fe63a8659330" containerName="container-00" Dec 07 17:08:57 crc kubenswrapper[4716]: I1207 17:08:57.327966 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7sr89" Dec 07 17:08:57 crc kubenswrapper[4716]: I1207 17:08:57.336410 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7sr89"] Dec 07 17:08:57 crc kubenswrapper[4716]: I1207 17:08:57.426891 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dad2a9a6-cfc0-4e44-84f0-f5a352c9e611-catalog-content\") pod \"redhat-marketplace-7sr89\" (UID: \"dad2a9a6-cfc0-4e44-84f0-f5a352c9e611\") " pod="openshift-marketplace/redhat-marketplace-7sr89" Dec 07 17:08:57 crc kubenswrapper[4716]: I1207 17:08:57.427203 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jfkst\" (UniqueName: \"kubernetes.io/projected/dad2a9a6-cfc0-4e44-84f0-f5a352c9e611-kube-api-access-jfkst\") pod \"redhat-marketplace-7sr89\" (UID: \"dad2a9a6-cfc0-4e44-84f0-f5a352c9e611\") " pod="openshift-marketplace/redhat-marketplace-7sr89" Dec 07 17:08:57 crc kubenswrapper[4716]: I1207 17:08:57.427263 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dad2a9a6-cfc0-4e44-84f0-f5a352c9e611-utilities\") pod \"redhat-marketplace-7sr89\" (UID: \"dad2a9a6-cfc0-4e44-84f0-f5a352c9e611\") " pod="openshift-marketplace/redhat-marketplace-7sr89" Dec 07 17:08:57 crc kubenswrapper[4716]: I1207 17:08:57.463991 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-bvvcz_1c545d8b-e4c4-411c-b122-68f22c6befa4/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 07 17:08:57 crc kubenswrapper[4716]: I1207 17:08:57.494471 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_5f86be30-3a49-4354-a6e9-059592ee5ebc/ceilometer-central-agent/0.log" Dec 07 17:08:57 crc kubenswrapper[4716]: I1207 17:08:57.528434 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dad2a9a6-cfc0-4e44-84f0-f5a352c9e611-utilities\") pod \"redhat-marketplace-7sr89\" (UID: \"dad2a9a6-cfc0-4e44-84f0-f5a352c9e611\") " pod="openshift-marketplace/redhat-marketplace-7sr89" Dec 07 17:08:57 crc kubenswrapper[4716]: I1207 17:08:57.528592 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dad2a9a6-cfc0-4e44-84f0-f5a352c9e611-catalog-content\") pod \"redhat-marketplace-7sr89\" (UID: \"dad2a9a6-cfc0-4e44-84f0-f5a352c9e611\") " pod="openshift-marketplace/redhat-marketplace-7sr89" Dec 07 17:08:57 crc kubenswrapper[4716]: I1207 17:08:57.528613 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jfkst\" (UniqueName: \"kubernetes.io/projected/dad2a9a6-cfc0-4e44-84f0-f5a352c9e611-kube-api-access-jfkst\") pod \"redhat-marketplace-7sr89\" (UID: \"dad2a9a6-cfc0-4e44-84f0-f5a352c9e611\") " pod="openshift-marketplace/redhat-marketplace-7sr89" Dec 07 17:08:57 crc kubenswrapper[4716]: I1207 17:08:57.528958 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dad2a9a6-cfc0-4e44-84f0-f5a352c9e611-utilities\") pod \"redhat-marketplace-7sr89\" (UID: \"dad2a9a6-cfc0-4e44-84f0-f5a352c9e611\") " pod="openshift-marketplace/redhat-marketplace-7sr89" Dec 07 17:08:57 crc kubenswrapper[4716]: I1207 17:08:57.529030 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dad2a9a6-cfc0-4e44-84f0-f5a352c9e611-catalog-content\") pod \"redhat-marketplace-7sr89\" (UID: \"dad2a9a6-cfc0-4e44-84f0-f5a352c9e611\") " pod="openshift-marketplace/redhat-marketplace-7sr89" Dec 07 17:08:57 crc kubenswrapper[4716]: I1207 17:08:57.549568 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jfkst\" (UniqueName: \"kubernetes.io/projected/dad2a9a6-cfc0-4e44-84f0-f5a352c9e611-kube-api-access-jfkst\") pod \"redhat-marketplace-7sr89\" (UID: \"dad2a9a6-cfc0-4e44-84f0-f5a352c9e611\") " pod="openshift-marketplace/redhat-marketplace-7sr89" Dec 07 17:08:57 crc kubenswrapper[4716]: I1207 17:08:57.644761 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7sr89" Dec 07 17:08:57 crc kubenswrapper[4716]: I1207 17:08:57.648533 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_5f86be30-3a49-4354-a6e9-059592ee5ebc/ceilometer-notification-agent/0.log" Dec 07 17:08:57 crc kubenswrapper[4716]: I1207 17:08:57.763892 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_5f86be30-3a49-4354-a6e9-059592ee5ebc/proxy-httpd/0.log" Dec 07 17:08:57 crc kubenswrapper[4716]: I1207 17:08:57.861863 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_5f86be30-3a49-4354-a6e9-059592ee5ebc/sg-core/0.log" Dec 07 17:08:57 crc kubenswrapper[4716]: I1207 17:08:57.980670 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_9237e175-c046-4f00-9535-474448e79076/cinder-api/0.log" Dec 07 17:08:58 crc kubenswrapper[4716]: I1207 17:08:58.012788 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_9237e175-c046-4f00-9535-474448e79076/cinder-api-log/0.log" Dec 07 17:08:58 crc kubenswrapper[4716]: I1207 17:08:58.116589 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7sr89"] Dec 07 17:08:58 crc kubenswrapper[4716]: I1207 17:08:58.171256 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_6db216a8-54ce-4596-a4ab-bb24a787b027/cinder-scheduler/0.log" Dec 07 17:08:58 crc kubenswrapper[4716]: I1207 17:08:58.226058 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_6db216a8-54ce-4596-a4ab-bb24a787b027/probe/0.log" Dec 07 17:08:58 crc kubenswrapper[4716]: I1207 17:08:58.324927 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-zsnd4_d2fcfd64-6b34-4f9c-9e8b-3715df35f92d/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 07 17:08:58 crc kubenswrapper[4716]: I1207 17:08:58.426206 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7sr89" event={"ID":"dad2a9a6-cfc0-4e44-84f0-f5a352c9e611","Type":"ContainerStarted","Data":"5d4d81b0930298566b3a80ac2d152b668814fd4c5f3833ce8241377b11421521"} Dec 07 17:08:58 crc kubenswrapper[4716]: I1207 17:08:58.469966 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-kcj2p_9d00d0ca-8b9a-455b-a7a5-f3d17a6e363a/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 07 17:08:58 crc kubenswrapper[4716]: I1207 17:08:58.487179 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-sw78r_27293e5d-5955-4ccb-b78b-63433622f073/init/0.log" Dec 07 17:08:58 crc kubenswrapper[4716]: I1207 17:08:58.686652 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-sw78r_27293e5d-5955-4ccb-b78b-63433622f073/init/0.log" Dec 07 17:08:58 crc kubenswrapper[4716]: I1207 17:08:58.755960 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-9d8h7_7495cf42-56aa-43c9-9d24-a022a3e50505/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Dec 07 17:08:58 crc kubenswrapper[4716]: I1207 17:08:58.760089 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-sw78r_27293e5d-5955-4ccb-b78b-63433622f073/dnsmasq-dns/0.log" Dec 07 17:08:58 crc kubenswrapper[4716]: I1207 17:08:58.942626 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_9f4f8c7b-38d4-4054-bb66-c51f32448567/glance-httpd/0.log" Dec 07 17:08:58 crc kubenswrapper[4716]: I1207 17:08:58.949904 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_9f4f8c7b-38d4-4054-bb66-c51f32448567/glance-log/0.log" Dec 07 17:08:59 crc kubenswrapper[4716]: I1207 17:08:59.086561 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_0f287d7e-4b08-4c16-8a23-cf9e69225d3c/glance-httpd/0.log" Dec 07 17:08:59 crc kubenswrapper[4716]: I1207 17:08:59.134026 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_0f287d7e-4b08-4c16-8a23-cf9e69225d3c/glance-log/0.log" Dec 07 17:08:59 crc kubenswrapper[4716]: I1207 17:08:59.337236 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7f9b78866d-9g9kv_d845ae0b-5d45-4021-a1e9-4b124298b65b/horizon/0.log" Dec 07 17:08:59 crc kubenswrapper[4716]: I1207 17:08:59.435488 4716 generic.go:334] "Generic (PLEG): container finished" podID="dad2a9a6-cfc0-4e44-84f0-f5a352c9e611" containerID="032e735c2ca329ea227a0fed156105ba64b353b9c4fc73e494f7a67bd19d9e75" exitCode=0 Dec 07 17:08:59 crc kubenswrapper[4716]: I1207 17:08:59.435530 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7sr89" event={"ID":"dad2a9a6-cfc0-4e44-84f0-f5a352c9e611","Type":"ContainerDied","Data":"032e735c2ca329ea227a0fed156105ba64b353b9c4fc73e494f7a67bd19d9e75"} Dec 07 17:08:59 crc kubenswrapper[4716]: I1207 17:08:59.438367 4716 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 07 17:08:59 crc kubenswrapper[4716]: I1207 17:08:59.476975 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-52qx7_6dd92737-9433-4fe4-ab78-03ebaeb31a24/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 07 17:08:59 crc kubenswrapper[4716]: I1207 17:08:59.642883 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-bd27b_9e126137-6235-4a75-8d46-05bb3cf60acc/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 07 17:08:59 crc kubenswrapper[4716]: I1207 17:08:59.657474 4716 scope.go:117] "RemoveContainer" containerID="0f9119a8d01b371a5ab06e6f944c6501025c243f659e26196309c46fc20c7a5d" Dec 07 17:08:59 crc kubenswrapper[4716]: E1207 17:08:59.657798 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 17:08:59 crc kubenswrapper[4716]: I1207 17:08:59.698502 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7f9b78866d-9g9kv_d845ae0b-5d45-4021-a1e9-4b124298b65b/horizon-log/0.log" Dec 07 17:08:59 crc kubenswrapper[4716]: I1207 17:08:59.789339 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-6c4445878c-bnpkh_ade9101f-5435-455f-807c-d277918cbb46/keystone-api/0.log" Dec 07 17:09:00 crc kubenswrapper[4716]: I1207 17:09:00.006427 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29418781-qddct_1c7bbcbc-7ce9-419d-b63d-8bcc377e4264/keystone-cron/0.log" Dec 07 17:09:00 crc kubenswrapper[4716]: I1207 17:09:00.137126 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_7953b7b7-a0c6-452c-868a-dcc4a1bda1fa/kube-state-metrics/0.log" Dec 07 17:09:00 crc kubenswrapper[4716]: I1207 17:09:00.203263 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-66mjc_0c94fa45-a0ac-401f-87c6-4f9c65cf9ff3/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 07 17:09:00 crc kubenswrapper[4716]: I1207 17:09:00.452555 4716 generic.go:334] "Generic (PLEG): container finished" podID="dad2a9a6-cfc0-4e44-84f0-f5a352c9e611" containerID="fea9aa453d0c45ecb1b1b8f6769cd011b4d84a1cc7c2c7206727e00f47e6dec5" exitCode=0 Dec 07 17:09:00 crc kubenswrapper[4716]: I1207 17:09:00.452609 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7sr89" event={"ID":"dad2a9a6-cfc0-4e44-84f0-f5a352c9e611","Type":"ContainerDied","Data":"fea9aa453d0c45ecb1b1b8f6769cd011b4d84a1cc7c2c7206727e00f47e6dec5"} Dec 07 17:09:00 crc kubenswrapper[4716]: I1207 17:09:00.570370 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-bd7b79585-wwp6b_dd279760-9178-4ad8-ae1e-ae4e6fab3f3c/neutron-httpd/0.log" Dec 07 17:09:00 crc kubenswrapper[4716]: I1207 17:09:00.613027 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-n265m_b4f56131-2845-43e1-84ca-52db74bf2b08/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 07 17:09:00 crc kubenswrapper[4716]: I1207 17:09:00.627061 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-bd7b79585-wwp6b_dd279760-9178-4ad8-ae1e-ae4e6fab3f3c/neutron-api/0.log" Dec 07 17:09:01 crc kubenswrapper[4716]: I1207 17:09:01.329758 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_3953832e-aa6d-49f2-980d-e2308f0d1ece/nova-api-log/0.log" Dec 07 17:09:01 crc kubenswrapper[4716]: I1207 17:09:01.472807 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7sr89" event={"ID":"dad2a9a6-cfc0-4e44-84f0-f5a352c9e611","Type":"ContainerStarted","Data":"09c4c07ec47e7636fa56fbacfd96493338786a439c5c3a28517802067e917e5b"} Dec 07 17:09:01 crc kubenswrapper[4716]: I1207 17:09:01.542152 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_b3643cef-3c95-4ecc-9225-ebf929947127/nova-cell0-conductor-conductor/0.log" Dec 07 17:09:01 crc kubenswrapper[4716]: I1207 17:09:01.690508 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_098f28a4-4852-4c3a-ada9-7b8b9272c3ae/nova-cell1-conductor-conductor/0.log" Dec 07 17:09:01 crc kubenswrapper[4716]: I1207 17:09:01.921632 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_3953832e-aa6d-49f2-980d-e2308f0d1ece/nova-api-api/0.log" Dec 07 17:09:01 crc kubenswrapper[4716]: I1207 17:09:01.991887 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_c0c9c313-94be-404d-849d-3eb23c5eb5e5/nova-cell1-novncproxy-novncproxy/0.log" Dec 07 17:09:01 crc kubenswrapper[4716]: I1207 17:09:01.996285 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-8jm5n_01472d87-913d-4565-8d83-40966b88a630/nova-edpm-deployment-openstack-edpm-ipam/0.log" Dec 07 17:09:02 crc kubenswrapper[4716]: I1207 17:09:02.221775 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_2ac894fd-e293-4f28-bdb5-e80e01601903/nova-metadata-log/0.log" Dec 07 17:09:02 crc kubenswrapper[4716]: I1207 17:09:02.555581 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_f5bcaa65-381c-40e7-97bd-4fc0eafc572a/mysql-bootstrap/0.log" Dec 07 17:09:02 crc kubenswrapper[4716]: I1207 17:09:02.622390 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_794a9d02-75dc-4ad6-bdb1-51bccaab7c9a/nova-scheduler-scheduler/0.log" Dec 07 17:09:02 crc kubenswrapper[4716]: I1207 17:09:02.717558 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_f5bcaa65-381c-40e7-97bd-4fc0eafc572a/mysql-bootstrap/0.log" Dec 07 17:09:02 crc kubenswrapper[4716]: I1207 17:09:02.778257 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_f5bcaa65-381c-40e7-97bd-4fc0eafc572a/galera/0.log" Dec 07 17:09:02 crc kubenswrapper[4716]: I1207 17:09:02.926865 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_c450ef40-3a23-4bab-9d64-4475f12d0490/mysql-bootstrap/0.log" Dec 07 17:09:03 crc kubenswrapper[4716]: I1207 17:09:03.139434 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_c450ef40-3a23-4bab-9d64-4475f12d0490/galera/0.log" Dec 07 17:09:03 crc kubenswrapper[4716]: I1207 17:09:03.145313 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_c450ef40-3a23-4bab-9d64-4475f12d0490/mysql-bootstrap/0.log" Dec 07 17:09:03 crc kubenswrapper[4716]: I1207 17:09:03.348702 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_6a6ecf77-dd60-497a-8670-919d23db95dd/openstackclient/0.log" Dec 07 17:09:03 crc kubenswrapper[4716]: I1207 17:09:03.432859 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-8r8gk_1fcf6527-b54d-4239-a4c4-a9cd6fe35bd2/openstack-network-exporter/0.log" Dec 07 17:09:03 crc kubenswrapper[4716]: I1207 17:09:03.549454 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_2ac894fd-e293-4f28-bdb5-e80e01601903/nova-metadata-metadata/0.log" Dec 07 17:09:03 crc kubenswrapper[4716]: I1207 17:09:03.843673 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-g6nwl_0fc0f0b8-b60a-4aff-84dd-bbec37b12d81/ovsdb-server-init/0.log" Dec 07 17:09:04 crc kubenswrapper[4716]: I1207 17:09:04.130328 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-g6nwl_0fc0f0b8-b60a-4aff-84dd-bbec37b12d81/ovsdb-server-init/0.log" Dec 07 17:09:04 crc kubenswrapper[4716]: I1207 17:09:04.213143 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-g6nwl_0fc0f0b8-b60a-4aff-84dd-bbec37b12d81/ovs-vswitchd/0.log" Dec 07 17:09:04 crc kubenswrapper[4716]: I1207 17:09:04.257754 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-g6nwl_0fc0f0b8-b60a-4aff-84dd-bbec37b12d81/ovsdb-server/0.log" Dec 07 17:09:04 crc kubenswrapper[4716]: I1207 17:09:04.377555 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-tgl27_399984bf-94af-4630-a7c3-9375a388159a/ovn-controller/0.log" Dec 07 17:09:04 crc kubenswrapper[4716]: I1207 17:09:04.516068 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-26hn5_b9e8242a-2734-47b9-9f14-7eea8d9ad3c4/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 07 17:09:04 crc kubenswrapper[4716]: I1207 17:09:04.582450 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_3ae6bded-8084-4e07-afcc-a7fafb718e3a/openstack-network-exporter/0.log" Dec 07 17:09:04 crc kubenswrapper[4716]: I1207 17:09:04.650431 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_3ae6bded-8084-4e07-afcc-a7fafb718e3a/ovn-northd/0.log" Dec 07 17:09:04 crc kubenswrapper[4716]: I1207 17:09:04.797337 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_771a3c3e-34e4-43f6-a3e2-8e7156e081d1/openstack-network-exporter/0.log" Dec 07 17:09:04 crc kubenswrapper[4716]: I1207 17:09:04.819051 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_771a3c3e-34e4-43f6-a3e2-8e7156e081d1/ovsdbserver-nb/0.log" Dec 07 17:09:05 crc kubenswrapper[4716]: I1207 17:09:05.051572 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_5c11a9b9-34de-42f7-9399-3cddce1af7c9/openstack-network-exporter/0.log" Dec 07 17:09:05 crc kubenswrapper[4716]: I1207 17:09:05.058962 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_5c11a9b9-34de-42f7-9399-3cddce1af7c9/ovsdbserver-sb/0.log" Dec 07 17:09:05 crc kubenswrapper[4716]: I1207 17:09:05.537401 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_bb3eb7dd-3210-452a-adf4-92bb2eb2447f/setup-container/0.log" Dec 07 17:09:05 crc kubenswrapper[4716]: I1207 17:09:05.568948 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-746cf47744-w9nm7_0320eb6a-f473-4c4c-ae36-f74080dcdaa5/placement-api/0.log" Dec 07 17:09:05 crc kubenswrapper[4716]: I1207 17:09:05.704873 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_bb3eb7dd-3210-452a-adf4-92bb2eb2447f/setup-container/0.log" Dec 07 17:09:05 crc kubenswrapper[4716]: I1207 17:09:05.749896 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_bb3eb7dd-3210-452a-adf4-92bb2eb2447f/rabbitmq/0.log" Dec 07 17:09:05 crc kubenswrapper[4716]: I1207 17:09:05.846037 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-746cf47744-w9nm7_0320eb6a-f473-4c4c-ae36-f74080dcdaa5/placement-log/0.log" Dec 07 17:09:05 crc kubenswrapper[4716]: I1207 17:09:05.898265 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_db1d1fd3-b1ed-472b-b615-b866706b28a9/setup-container/0.log" Dec 07 17:09:06 crc kubenswrapper[4716]: I1207 17:09:06.082236 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_db1d1fd3-b1ed-472b-b615-b866706b28a9/setup-container/0.log" Dec 07 17:09:06 crc kubenswrapper[4716]: I1207 17:09:06.141978 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_db1d1fd3-b1ed-472b-b615-b866706b28a9/rabbitmq/0.log" Dec 07 17:09:06 crc kubenswrapper[4716]: I1207 17:09:06.206281 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-xsl26_3a9def06-fe00-45a7-9599-2fada793ba25/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 07 17:09:06 crc kubenswrapper[4716]: I1207 17:09:06.330664 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-l5phd_2d823bb1-cb32-4c09-af7c-6ca81d0277bf/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Dec 07 17:09:06 crc kubenswrapper[4716]: I1207 17:09:06.433801 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-dv7q9_2c9bedba-2a2e-4e68-839a-646d5747911c/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 07 17:09:06 crc kubenswrapper[4716]: I1207 17:09:06.535369 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-9c27f_0d4b7b56-033c-41be-b46f-e4090775ab5c/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 07 17:09:06 crc kubenswrapper[4716]: I1207 17:09:06.687064 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-vpm5k_92420ce0-b95e-4857-b6f3-f6d4519c3d86/ssh-known-hosts-edpm-deployment/0.log" Dec 07 17:09:06 crc kubenswrapper[4716]: I1207 17:09:06.868475 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-5488d7fd99-pwmf5_826c0644-3046-453f-a139-5bd9c95216d0/proxy-server/0.log" Dec 07 17:09:06 crc kubenswrapper[4716]: I1207 17:09:06.940674 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-5488d7fd99-pwmf5_826c0644-3046-453f-a139-5bd9c95216d0/proxy-httpd/0.log" Dec 07 17:09:07 crc kubenswrapper[4716]: I1207 17:09:07.015182 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-jgrbf_2c371669-443a-492e-ad04-ab79ae978e5f/swift-ring-rebalance/0.log" Dec 07 17:09:07 crc kubenswrapper[4716]: I1207 17:09:07.150381 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a714bbaa-9626-4b49-92ae-2abb029408d3/account-auditor/0.log" Dec 07 17:09:07 crc kubenswrapper[4716]: I1207 17:09:07.189996 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a714bbaa-9626-4b49-92ae-2abb029408d3/account-reaper/0.log" Dec 07 17:09:07 crc kubenswrapper[4716]: I1207 17:09:07.300110 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a714bbaa-9626-4b49-92ae-2abb029408d3/account-replicator/0.log" Dec 07 17:09:07 crc kubenswrapper[4716]: I1207 17:09:07.307428 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a714bbaa-9626-4b49-92ae-2abb029408d3/account-server/0.log" Dec 07 17:09:07 crc kubenswrapper[4716]: I1207 17:09:07.378425 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a714bbaa-9626-4b49-92ae-2abb029408d3/container-auditor/0.log" Dec 07 17:09:07 crc kubenswrapper[4716]: I1207 17:09:07.410798 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a714bbaa-9626-4b49-92ae-2abb029408d3/container-replicator/0.log" Dec 07 17:09:07 crc kubenswrapper[4716]: I1207 17:09:07.484631 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a714bbaa-9626-4b49-92ae-2abb029408d3/container-server/0.log" Dec 07 17:09:07 crc kubenswrapper[4716]: I1207 17:09:07.506726 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a714bbaa-9626-4b49-92ae-2abb029408d3/container-updater/0.log" Dec 07 17:09:07 crc kubenswrapper[4716]: I1207 17:09:07.597740 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a714bbaa-9626-4b49-92ae-2abb029408d3/object-auditor/0.log" Dec 07 17:09:07 crc kubenswrapper[4716]: I1207 17:09:07.632530 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a714bbaa-9626-4b49-92ae-2abb029408d3/object-expirer/0.log" Dec 07 17:09:07 crc kubenswrapper[4716]: I1207 17:09:07.646499 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-7sr89" Dec 07 17:09:07 crc kubenswrapper[4716]: I1207 17:09:07.646533 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-7sr89" Dec 07 17:09:07 crc kubenswrapper[4716]: I1207 17:09:07.671849 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a714bbaa-9626-4b49-92ae-2abb029408d3/object-replicator/0.log" Dec 07 17:09:07 crc kubenswrapper[4716]: I1207 17:09:07.700234 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-7sr89" Dec 07 17:09:07 crc kubenswrapper[4716]: I1207 17:09:07.723968 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-7sr89" podStartSLOduration=9.338406572 podStartE2EDuration="10.72394942s" podCreationTimestamp="2025-12-07 17:08:57 +0000 UTC" firstStartedPulling="2025-12-07 17:08:59.438119762 +0000 UTC m=+4002.128404674" lastFinishedPulling="2025-12-07 17:09:00.82366261 +0000 UTC m=+4003.513947522" observedRunningTime="2025-12-07 17:09:01.500115743 +0000 UTC m=+4004.190400655" watchObservedRunningTime="2025-12-07 17:09:07.72394942 +0000 UTC m=+4010.414234342" Dec 07 17:09:07 crc kubenswrapper[4716]: I1207 17:09:07.745346 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a714bbaa-9626-4b49-92ae-2abb029408d3/object-server/0.log" Dec 07 17:09:07 crc kubenswrapper[4716]: I1207 17:09:07.801645 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a714bbaa-9626-4b49-92ae-2abb029408d3/object-updater/0.log" Dec 07 17:09:07 crc kubenswrapper[4716]: I1207 17:09:07.852236 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a714bbaa-9626-4b49-92ae-2abb029408d3/swift-recon-cron/0.log" Dec 07 17:09:07 crc kubenswrapper[4716]: I1207 17:09:07.864362 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a714bbaa-9626-4b49-92ae-2abb029408d3/rsync/0.log" Dec 07 17:09:08 crc kubenswrapper[4716]: I1207 17:09:08.044404 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-69qz6_c9735df5-27ef-41b1-84c1-4cab55f23b3a/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 07 17:09:08 crc kubenswrapper[4716]: I1207 17:09:08.058024 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_7c4bb2e7-29d0-4548-863b-ea89d11d68a7/tempest-tests-tempest-tests-runner/0.log" Dec 07 17:09:08 crc kubenswrapper[4716]: I1207 17:09:08.278536 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_fa15bae7-7145-486c-8f55-53797c60c01c/test-operator-logs-container/0.log" Dec 07 17:09:08 crc kubenswrapper[4716]: I1207 17:09:08.353564 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-kwl9v_fa50585d-9042-4f6e-b926-fdff8344e0ad/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 07 17:09:08 crc kubenswrapper[4716]: I1207 17:09:08.579208 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-7sr89" Dec 07 17:09:08 crc kubenswrapper[4716]: I1207 17:09:08.639647 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7sr89"] Dec 07 17:09:10 crc kubenswrapper[4716]: I1207 17:09:10.540543 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-7sr89" podUID="dad2a9a6-cfc0-4e44-84f0-f5a352c9e611" containerName="registry-server" containerID="cri-o://09c4c07ec47e7636fa56fbacfd96493338786a439c5c3a28517802067e917e5b" gracePeriod=2 Dec 07 17:09:11 crc kubenswrapper[4716]: I1207 17:09:11.550120 4716 generic.go:334] "Generic (PLEG): container finished" podID="dad2a9a6-cfc0-4e44-84f0-f5a352c9e611" containerID="09c4c07ec47e7636fa56fbacfd96493338786a439c5c3a28517802067e917e5b" exitCode=0 Dec 07 17:09:11 crc kubenswrapper[4716]: I1207 17:09:11.550395 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7sr89" event={"ID":"dad2a9a6-cfc0-4e44-84f0-f5a352c9e611","Type":"ContainerDied","Data":"09c4c07ec47e7636fa56fbacfd96493338786a439c5c3a28517802067e917e5b"} Dec 07 17:09:11 crc kubenswrapper[4716]: I1207 17:09:11.550420 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7sr89" event={"ID":"dad2a9a6-cfc0-4e44-84f0-f5a352c9e611","Type":"ContainerDied","Data":"5d4d81b0930298566b3a80ac2d152b668814fd4c5f3833ce8241377b11421521"} Dec 07 17:09:11 crc kubenswrapper[4716]: I1207 17:09:11.550432 4716 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5d4d81b0930298566b3a80ac2d152b668814fd4c5f3833ce8241377b11421521" Dec 07 17:09:11 crc kubenswrapper[4716]: I1207 17:09:11.574514 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7sr89" Dec 07 17:09:11 crc kubenswrapper[4716]: I1207 17:09:11.702936 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dad2a9a6-cfc0-4e44-84f0-f5a352c9e611-utilities\") pod \"dad2a9a6-cfc0-4e44-84f0-f5a352c9e611\" (UID: \"dad2a9a6-cfc0-4e44-84f0-f5a352c9e611\") " Dec 07 17:09:11 crc kubenswrapper[4716]: I1207 17:09:11.702975 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jfkst\" (UniqueName: \"kubernetes.io/projected/dad2a9a6-cfc0-4e44-84f0-f5a352c9e611-kube-api-access-jfkst\") pod \"dad2a9a6-cfc0-4e44-84f0-f5a352c9e611\" (UID: \"dad2a9a6-cfc0-4e44-84f0-f5a352c9e611\") " Dec 07 17:09:11 crc kubenswrapper[4716]: I1207 17:09:11.703017 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dad2a9a6-cfc0-4e44-84f0-f5a352c9e611-catalog-content\") pod \"dad2a9a6-cfc0-4e44-84f0-f5a352c9e611\" (UID: \"dad2a9a6-cfc0-4e44-84f0-f5a352c9e611\") " Dec 07 17:09:11 crc kubenswrapper[4716]: I1207 17:09:11.703752 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dad2a9a6-cfc0-4e44-84f0-f5a352c9e611-utilities" (OuterVolumeSpecName: "utilities") pod "dad2a9a6-cfc0-4e44-84f0-f5a352c9e611" (UID: "dad2a9a6-cfc0-4e44-84f0-f5a352c9e611"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 17:09:11 crc kubenswrapper[4716]: I1207 17:09:11.704427 4716 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dad2a9a6-cfc0-4e44-84f0-f5a352c9e611-utilities\") on node \"crc\" DevicePath \"\"" Dec 07 17:09:11 crc kubenswrapper[4716]: I1207 17:09:11.709185 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dad2a9a6-cfc0-4e44-84f0-f5a352c9e611-kube-api-access-jfkst" (OuterVolumeSpecName: "kube-api-access-jfkst") pod "dad2a9a6-cfc0-4e44-84f0-f5a352c9e611" (UID: "dad2a9a6-cfc0-4e44-84f0-f5a352c9e611"). InnerVolumeSpecName "kube-api-access-jfkst". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 17:09:11 crc kubenswrapper[4716]: I1207 17:09:11.724330 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dad2a9a6-cfc0-4e44-84f0-f5a352c9e611-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dad2a9a6-cfc0-4e44-84f0-f5a352c9e611" (UID: "dad2a9a6-cfc0-4e44-84f0-f5a352c9e611"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 17:09:11 crc kubenswrapper[4716]: I1207 17:09:11.805647 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jfkst\" (UniqueName: \"kubernetes.io/projected/dad2a9a6-cfc0-4e44-84f0-f5a352c9e611-kube-api-access-jfkst\") on node \"crc\" DevicePath \"\"" Dec 07 17:09:11 crc kubenswrapper[4716]: I1207 17:09:11.805685 4716 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dad2a9a6-cfc0-4e44-84f0-f5a352c9e611-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 07 17:09:12 crc kubenswrapper[4716]: I1207 17:09:12.557009 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7sr89" Dec 07 17:09:12 crc kubenswrapper[4716]: I1207 17:09:12.623890 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7sr89"] Dec 07 17:09:12 crc kubenswrapper[4716]: I1207 17:09:12.641934 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-7sr89"] Dec 07 17:09:13 crc kubenswrapper[4716]: I1207 17:09:13.662931 4716 scope.go:117] "RemoveContainer" containerID="0f9119a8d01b371a5ab06e6f944c6501025c243f659e26196309c46fc20c7a5d" Dec 07 17:09:13 crc kubenswrapper[4716]: E1207 17:09:13.663544 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 17:09:13 crc kubenswrapper[4716]: I1207 17:09:13.668913 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dad2a9a6-cfc0-4e44-84f0-f5a352c9e611" path="/var/lib/kubelet/pods/dad2a9a6-cfc0-4e44-84f0-f5a352c9e611/volumes" Dec 07 17:09:17 crc kubenswrapper[4716]: I1207 17:09:16.999722 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_0f4b5de9-9760-4930-bc8e-d7e47a13ad2e/memcached/0.log" Dec 07 17:09:24 crc kubenswrapper[4716]: I1207 17:09:24.659021 4716 scope.go:117] "RemoveContainer" containerID="0f9119a8d01b371a5ab06e6f944c6501025c243f659e26196309c46fc20c7a5d" Dec 07 17:09:24 crc kubenswrapper[4716]: E1207 17:09:24.660778 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 17:09:34 crc kubenswrapper[4716]: I1207 17:09:34.850563 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4cb94d7f59637b410c6ea06aeac57a499ffab903391b8415798aca23986vcrp_5bdf0591-fc57-4545-8ab4-e89de972df50/util/0.log" Dec 07 17:09:35 crc kubenswrapper[4716]: I1207 17:09:35.170596 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4cb94d7f59637b410c6ea06aeac57a499ffab903391b8415798aca23986vcrp_5bdf0591-fc57-4545-8ab4-e89de972df50/util/0.log" Dec 07 17:09:35 crc kubenswrapper[4716]: I1207 17:09:35.233871 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4cb94d7f59637b410c6ea06aeac57a499ffab903391b8415798aca23986vcrp_5bdf0591-fc57-4545-8ab4-e89de972df50/pull/0.log" Dec 07 17:09:35 crc kubenswrapper[4716]: I1207 17:09:35.242417 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4cb94d7f59637b410c6ea06aeac57a499ffab903391b8415798aca23986vcrp_5bdf0591-fc57-4545-8ab4-e89de972df50/pull/0.log" Dec 07 17:09:35 crc kubenswrapper[4716]: I1207 17:09:35.842322 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4cb94d7f59637b410c6ea06aeac57a499ffab903391b8415798aca23986vcrp_5bdf0591-fc57-4545-8ab4-e89de972df50/pull/0.log" Dec 07 17:09:35 crc kubenswrapper[4716]: I1207 17:09:35.893475 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4cb94d7f59637b410c6ea06aeac57a499ffab903391b8415798aca23986vcrp_5bdf0591-fc57-4545-8ab4-e89de972df50/extract/0.log" Dec 07 17:09:35 crc kubenswrapper[4716]: I1207 17:09:35.903142 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4cb94d7f59637b410c6ea06aeac57a499ffab903391b8415798aca23986vcrp_5bdf0591-fc57-4545-8ab4-e89de972df50/util/0.log" Dec 07 17:09:36 crc kubenswrapper[4716]: I1207 17:09:36.040185 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-m957c_8765c8b2-a50b-4417-adbe-1174dcdfe172/kube-rbac-proxy/0.log" Dec 07 17:09:36 crc kubenswrapper[4716]: I1207 17:09:36.097134 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6c677c69b-4nrlz_62368956-27d2-41fa-a2b2-0fb49d869f11/kube-rbac-proxy/0.log" Dec 07 17:09:36 crc kubenswrapper[4716]: I1207 17:09:36.181539 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-m957c_8765c8b2-a50b-4417-adbe-1174dcdfe172/manager/0.log" Dec 07 17:09:36 crc kubenswrapper[4716]: I1207 17:09:36.259223 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6c677c69b-4nrlz_62368956-27d2-41fa-a2b2-0fb49d869f11/manager/0.log" Dec 07 17:09:36 crc kubenswrapper[4716]: I1207 17:09:36.319366 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-697fb699cf-zts6p_79562d20-950f-428e-ac3a-f78979053266/kube-rbac-proxy/0.log" Dec 07 17:09:36 crc kubenswrapper[4716]: I1207 17:09:36.400522 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-697fb699cf-zts6p_79562d20-950f-428e-ac3a-f78979053266/manager/0.log" Dec 07 17:09:36 crc kubenswrapper[4716]: I1207 17:09:36.486489 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5697bb5779-rcmz5_6f24594b-ebe7-4518-a067-45891924abe5/kube-rbac-proxy/0.log" Dec 07 17:09:36 crc kubenswrapper[4716]: I1207 17:09:36.625243 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5697bb5779-rcmz5_6f24594b-ebe7-4518-a067-45891924abe5/manager/0.log" Dec 07 17:09:36 crc kubenswrapper[4716]: I1207 17:09:36.726503 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-xqnzw_d9d5c8b8-240b-4c88-ad6e-b04881842f0b/manager/0.log" Dec 07 17:09:36 crc kubenswrapper[4716]: I1207 17:09:36.731263 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-xqnzw_d9d5c8b8-240b-4c88-ad6e-b04881842f0b/kube-rbac-proxy/0.log" Dec 07 17:09:36 crc kubenswrapper[4716]: I1207 17:09:36.896719 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-nx5mv_20067e63-712d-4f94-9019-627219d4299f/kube-rbac-proxy/0.log" Dec 07 17:09:36 crc kubenswrapper[4716]: I1207 17:09:36.937031 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-nx5mv_20067e63-712d-4f94-9019-627219d4299f/manager/0.log" Dec 07 17:09:37 crc kubenswrapper[4716]: I1207 17:09:37.117275 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-78d48bff9d-pxs5r_01686458-debd-419a-90ec-cd27cc6953ec/kube-rbac-proxy/0.log" Dec 07 17:09:37 crc kubenswrapper[4716]: I1207 17:09:37.225461 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-967d97867-hx48c_a929d511-1d07-429d-b302-8843a6834f52/kube-rbac-proxy/0.log" Dec 07 17:09:37 crc kubenswrapper[4716]: I1207 17:09:37.343256 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-78d48bff9d-pxs5r_01686458-debd-419a-90ec-cd27cc6953ec/manager/0.log" Dec 07 17:09:37 crc kubenswrapper[4716]: I1207 17:09:37.361266 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-967d97867-hx48c_a929d511-1d07-429d-b302-8843a6834f52/manager/0.log" Dec 07 17:09:37 crc kubenswrapper[4716]: I1207 17:09:37.413068 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-wctdp_2babb856-a5c2-4b60-a6bc-00d3f510f014/kube-rbac-proxy/0.log" Dec 07 17:09:37 crc kubenswrapper[4716]: I1207 17:09:37.605595 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-wctdp_2babb856-a5c2-4b60-a6bc-00d3f510f014/manager/0.log" Dec 07 17:09:37 crc kubenswrapper[4716]: I1207 17:09:37.633640 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5b5fd79c9c-kbzpp_d3ef7831-ac99-410d-90bf-42d87042964b/kube-rbac-proxy/0.log" Dec 07 17:09:37 crc kubenswrapper[4716]: I1207 17:09:37.658362 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5b5fd79c9c-kbzpp_d3ef7831-ac99-410d-90bf-42d87042964b/manager/0.log" Dec 07 17:09:37 crc kubenswrapper[4716]: I1207 17:09:37.665917 4716 scope.go:117] "RemoveContainer" containerID="0f9119a8d01b371a5ab06e6f944c6501025c243f659e26196309c46fc20c7a5d" Dec 07 17:09:37 crc kubenswrapper[4716]: E1207 17:09:37.666271 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 17:09:37 crc kubenswrapper[4716]: I1207 17:09:37.863531 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-79c8c4686c-9dsmr_5be39543-890e-479c-a041-d864922e038f/kube-rbac-proxy/0.log" Dec 07 17:09:37 crc kubenswrapper[4716]: I1207 17:09:37.882607 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-79c8c4686c-9dsmr_5be39543-890e-479c-a041-d864922e038f/manager/0.log" Dec 07 17:09:38 crc kubenswrapper[4716]: I1207 17:09:38.020297 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-srq27_463e6bed-ce60-4e35-98a3-cd837b9066f2/kube-rbac-proxy/0.log" Dec 07 17:09:38 crc kubenswrapper[4716]: I1207 17:09:38.127108 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-z7lcv_69c2e54f-9af6-44e8-abdd-226a29b64da6/kube-rbac-proxy/0.log" Dec 07 17:09:38 crc kubenswrapper[4716]: I1207 17:09:38.136113 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-srq27_463e6bed-ce60-4e35-98a3-cd837b9066f2/manager/0.log" Dec 07 17:09:38 crc kubenswrapper[4716]: I1207 17:09:38.157633 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-z7lcv_69c2e54f-9af6-44e8-abdd-226a29b64da6/manager/0.log" Dec 07 17:09:38 crc kubenswrapper[4716]: I1207 17:09:38.346022 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-kkdgl_57c9705a-9bf1-45b4-aefb-d224a9d72a4a/kube-rbac-proxy/0.log" Dec 07 17:09:38 crc kubenswrapper[4716]: I1207 17:09:38.364839 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-84b575879fdkpkx_34e0d69c-2f36-4ba6-a4e3-114cabe016c2/kube-rbac-proxy/0.log" Dec 07 17:09:38 crc kubenswrapper[4716]: I1207 17:09:38.375447 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-kkdgl_57c9705a-9bf1-45b4-aefb-d224a9d72a4a/manager/0.log" Dec 07 17:09:38 crc kubenswrapper[4716]: I1207 17:09:38.542580 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-84b575879fdkpkx_34e0d69c-2f36-4ba6-a4e3-114cabe016c2/manager/0.log" Dec 07 17:09:38 crc kubenswrapper[4716]: I1207 17:09:38.744023 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-64d95d5f5f-6t4rp_93975489-d46c-4ea7-a11b-65202433b51f/operator/0.log" Dec 07 17:09:38 crc kubenswrapper[4716]: I1207 17:09:38.775400 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-qcnsg_12e91e68-d5ea-4b85-81ca-88f8d00e06bc/registry-server/0.log" Dec 07 17:09:39 crc kubenswrapper[4716]: I1207 17:09:39.248361 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-h95ls_1bf3b0b2-ea38-4fa3-b07c-4467dc111476/manager/0.log" Dec 07 17:09:39 crc kubenswrapper[4716]: I1207 17:09:39.296735 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-h95ls_1bf3b0b2-ea38-4fa3-b07c-4467dc111476/kube-rbac-proxy/0.log" Dec 07 17:09:39 crc kubenswrapper[4716]: I1207 17:09:39.355354 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-8xrch_0f05e35d-f818-4f65-bac6-f6a22006627e/kube-rbac-proxy/0.log" Dec 07 17:09:39 crc kubenswrapper[4716]: I1207 17:09:39.529964 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-8xrch_0f05e35d-f818-4f65-bac6-f6a22006627e/manager/0.log" Dec 07 17:09:39 crc kubenswrapper[4716]: I1207 17:09:39.596040 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-7mjmm_79aa5696-34a1-45cf-b965-56f64fe63f4e/operator/0.log" Dec 07 17:09:39 crc kubenswrapper[4716]: I1207 17:09:39.601337 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-7db7d7886d-z6n25_9d840ba3-d5ef-48fc-87b2-6af4d164c9ba/manager/0.log" Dec 07 17:09:39 crc kubenswrapper[4716]: I1207 17:09:39.737448 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-9d58d64bc-p2hbf_d9d5a1f4-edb7-4630-8907-5d29c4678f24/manager/0.log" Dec 07 17:09:39 crc kubenswrapper[4716]: I1207 17:09:39.750188 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-9d58d64bc-p2hbf_d9d5a1f4-edb7-4630-8907-5d29c4678f24/kube-rbac-proxy/0.log" Dec 07 17:09:39 crc kubenswrapper[4716]: I1207 17:09:39.844512 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-58d5ff84df-ggcr7_e4c2f2f6-3285-4f48-a428-b2065f33c046/kube-rbac-proxy/0.log" Dec 07 17:09:39 crc kubenswrapper[4716]: I1207 17:09:39.858369 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-58d5ff84df-ggcr7_e4c2f2f6-3285-4f48-a428-b2065f33c046/manager/0.log" Dec 07 17:09:39 crc kubenswrapper[4716]: I1207 17:09:39.948516 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-4dnl8_0c6a49cd-1cfe-49c8-a068-f85c735b46b1/kube-rbac-proxy/0.log" Dec 07 17:09:40 crc kubenswrapper[4716]: I1207 17:09:40.046067 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-4dnl8_0c6a49cd-1cfe-49c8-a068-f85c735b46b1/manager/0.log" Dec 07 17:09:40 crc kubenswrapper[4716]: I1207 17:09:40.081947 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-667bd8d554-8zwtk_2ef47713-a73c-4c63-8fa0-ffc530832285/kube-rbac-proxy/0.log" Dec 07 17:09:40 crc kubenswrapper[4716]: I1207 17:09:40.164842 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-667bd8d554-8zwtk_2ef47713-a73c-4c63-8fa0-ffc530832285/manager/0.log" Dec 07 17:09:48 crc kubenswrapper[4716]: I1207 17:09:48.657621 4716 scope.go:117] "RemoveContainer" containerID="0f9119a8d01b371a5ab06e6f944c6501025c243f659e26196309c46fc20c7a5d" Dec 07 17:09:48 crc kubenswrapper[4716]: E1207 17:09:48.659196 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 17:09:59 crc kubenswrapper[4716]: I1207 17:09:59.658002 4716 scope.go:117] "RemoveContainer" containerID="0f9119a8d01b371a5ab06e6f944c6501025c243f659e26196309c46fc20c7a5d" Dec 07 17:09:59 crc kubenswrapper[4716]: E1207 17:09:59.658950 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 17:10:00 crc kubenswrapper[4716]: I1207 17:10:00.413116 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-mmfjq_db812416-33e4-462c-a28f-ec40b629dc37/control-plane-machine-set-operator/0.log" Dec 07 17:10:00 crc kubenswrapper[4716]: I1207 17:10:00.654504 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-j7zhr_4df2ba74-d18e-4e94-bee1-703645b1a463/kube-rbac-proxy/0.log" Dec 07 17:10:00 crc kubenswrapper[4716]: I1207 17:10:00.654617 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-j7zhr_4df2ba74-d18e-4e94-bee1-703645b1a463/machine-api-operator/0.log" Dec 07 17:10:13 crc kubenswrapper[4716]: I1207 17:10:13.956651 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-xb4z7_55db1fd5-6c4f-4f27-888c-4d4b2445e603/cert-manager-controller/0.log" Dec 07 17:10:14 crc kubenswrapper[4716]: I1207 17:10:14.143585 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-cgnm7_64d9b2cc-19be-4cd5-81d8-fda3cbf2228e/cert-manager-cainjector/0.log" Dec 07 17:10:14 crc kubenswrapper[4716]: I1207 17:10:14.229432 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-ft9cq_badea5ba-19c9-4716-87df-a4ef7248b50a/cert-manager-webhook/0.log" Dec 07 17:10:14 crc kubenswrapper[4716]: I1207 17:10:14.657975 4716 scope.go:117] "RemoveContainer" containerID="0f9119a8d01b371a5ab06e6f944c6501025c243f659e26196309c46fc20c7a5d" Dec 07 17:10:14 crc kubenswrapper[4716]: E1207 17:10:14.658248 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 17:10:19 crc kubenswrapper[4716]: I1207 17:10:19.033302 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-gmh8g"] Dec 07 17:10:19 crc kubenswrapper[4716]: E1207 17:10:19.034393 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dad2a9a6-cfc0-4e44-84f0-f5a352c9e611" containerName="extract-content" Dec 07 17:10:19 crc kubenswrapper[4716]: I1207 17:10:19.034408 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="dad2a9a6-cfc0-4e44-84f0-f5a352c9e611" containerName="extract-content" Dec 07 17:10:19 crc kubenswrapper[4716]: E1207 17:10:19.034443 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dad2a9a6-cfc0-4e44-84f0-f5a352c9e611" containerName="registry-server" Dec 07 17:10:19 crc kubenswrapper[4716]: I1207 17:10:19.034449 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="dad2a9a6-cfc0-4e44-84f0-f5a352c9e611" containerName="registry-server" Dec 07 17:10:19 crc kubenswrapper[4716]: E1207 17:10:19.034471 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dad2a9a6-cfc0-4e44-84f0-f5a352c9e611" containerName="extract-utilities" Dec 07 17:10:19 crc kubenswrapper[4716]: I1207 17:10:19.034477 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="dad2a9a6-cfc0-4e44-84f0-f5a352c9e611" containerName="extract-utilities" Dec 07 17:10:19 crc kubenswrapper[4716]: I1207 17:10:19.034709 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="dad2a9a6-cfc0-4e44-84f0-f5a352c9e611" containerName="registry-server" Dec 07 17:10:19 crc kubenswrapper[4716]: I1207 17:10:19.036383 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gmh8g" Dec 07 17:10:19 crc kubenswrapper[4716]: I1207 17:10:19.046102 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gmh8g"] Dec 07 17:10:19 crc kubenswrapper[4716]: I1207 17:10:19.162405 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52f0efae-9ebe-4069-b586-029b47cc07ae-utilities\") pod \"certified-operators-gmh8g\" (UID: \"52f0efae-9ebe-4069-b586-029b47cc07ae\") " pod="openshift-marketplace/certified-operators-gmh8g" Dec 07 17:10:19 crc kubenswrapper[4716]: I1207 17:10:19.162632 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2znxm\" (UniqueName: \"kubernetes.io/projected/52f0efae-9ebe-4069-b586-029b47cc07ae-kube-api-access-2znxm\") pod \"certified-operators-gmh8g\" (UID: \"52f0efae-9ebe-4069-b586-029b47cc07ae\") " pod="openshift-marketplace/certified-operators-gmh8g" Dec 07 17:10:19 crc kubenswrapper[4716]: I1207 17:10:19.162844 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52f0efae-9ebe-4069-b586-029b47cc07ae-catalog-content\") pod \"certified-operators-gmh8g\" (UID: \"52f0efae-9ebe-4069-b586-029b47cc07ae\") " pod="openshift-marketplace/certified-operators-gmh8g" Dec 07 17:10:19 crc kubenswrapper[4716]: I1207 17:10:19.265032 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52f0efae-9ebe-4069-b586-029b47cc07ae-utilities\") pod \"certified-operators-gmh8g\" (UID: \"52f0efae-9ebe-4069-b586-029b47cc07ae\") " pod="openshift-marketplace/certified-operators-gmh8g" Dec 07 17:10:19 crc kubenswrapper[4716]: I1207 17:10:19.265185 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2znxm\" (UniqueName: \"kubernetes.io/projected/52f0efae-9ebe-4069-b586-029b47cc07ae-kube-api-access-2znxm\") pod \"certified-operators-gmh8g\" (UID: \"52f0efae-9ebe-4069-b586-029b47cc07ae\") " pod="openshift-marketplace/certified-operators-gmh8g" Dec 07 17:10:19 crc kubenswrapper[4716]: I1207 17:10:19.265277 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52f0efae-9ebe-4069-b586-029b47cc07ae-catalog-content\") pod \"certified-operators-gmh8g\" (UID: \"52f0efae-9ebe-4069-b586-029b47cc07ae\") " pod="openshift-marketplace/certified-operators-gmh8g" Dec 07 17:10:19 crc kubenswrapper[4716]: I1207 17:10:19.265700 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52f0efae-9ebe-4069-b586-029b47cc07ae-utilities\") pod \"certified-operators-gmh8g\" (UID: \"52f0efae-9ebe-4069-b586-029b47cc07ae\") " pod="openshift-marketplace/certified-operators-gmh8g" Dec 07 17:10:19 crc kubenswrapper[4716]: I1207 17:10:19.265735 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52f0efae-9ebe-4069-b586-029b47cc07ae-catalog-content\") pod \"certified-operators-gmh8g\" (UID: \"52f0efae-9ebe-4069-b586-029b47cc07ae\") " pod="openshift-marketplace/certified-operators-gmh8g" Dec 07 17:10:19 crc kubenswrapper[4716]: I1207 17:10:19.290773 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2znxm\" (UniqueName: \"kubernetes.io/projected/52f0efae-9ebe-4069-b586-029b47cc07ae-kube-api-access-2znxm\") pod \"certified-operators-gmh8g\" (UID: \"52f0efae-9ebe-4069-b586-029b47cc07ae\") " pod="openshift-marketplace/certified-operators-gmh8g" Dec 07 17:10:19 crc kubenswrapper[4716]: I1207 17:10:19.355417 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gmh8g" Dec 07 17:10:19 crc kubenswrapper[4716]: I1207 17:10:19.868441 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gmh8g"] Dec 07 17:10:20 crc kubenswrapper[4716]: I1207 17:10:20.302034 4716 generic.go:334] "Generic (PLEG): container finished" podID="52f0efae-9ebe-4069-b586-029b47cc07ae" containerID="d807c57f4af64df58e7d07d841ac71a3a316375fb1da007294608fb00e87324a" exitCode=0 Dec 07 17:10:20 crc kubenswrapper[4716]: I1207 17:10:20.302126 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gmh8g" event={"ID":"52f0efae-9ebe-4069-b586-029b47cc07ae","Type":"ContainerDied","Data":"d807c57f4af64df58e7d07d841ac71a3a316375fb1da007294608fb00e87324a"} Dec 07 17:10:20 crc kubenswrapper[4716]: I1207 17:10:20.303198 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gmh8g" event={"ID":"52f0efae-9ebe-4069-b586-029b47cc07ae","Type":"ContainerStarted","Data":"344370e17eeaa4c0f24ae87d5ac7d4dad5192209322141b48f89559dde37e084"} Dec 07 17:10:21 crc kubenswrapper[4716]: I1207 17:10:21.315021 4716 generic.go:334] "Generic (PLEG): container finished" podID="52f0efae-9ebe-4069-b586-029b47cc07ae" containerID="056429bc0f75f4283005bf8dae5802acbe6f5e1985da6fd877b04d2c2a47e5ed" exitCode=0 Dec 07 17:10:21 crc kubenswrapper[4716]: I1207 17:10:21.315215 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gmh8g" event={"ID":"52f0efae-9ebe-4069-b586-029b47cc07ae","Type":"ContainerDied","Data":"056429bc0f75f4283005bf8dae5802acbe6f5e1985da6fd877b04d2c2a47e5ed"} Dec 07 17:10:22 crc kubenswrapper[4716]: I1207 17:10:22.327432 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gmh8g" event={"ID":"52f0efae-9ebe-4069-b586-029b47cc07ae","Type":"ContainerStarted","Data":"3941f3e337b0ed2fdc59351f79120436255f05af860d8b78a964493161b0f1a7"} Dec 07 17:10:22 crc kubenswrapper[4716]: I1207 17:10:22.344607 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-gmh8g" podStartSLOduration=1.922276477 podStartE2EDuration="3.344585611s" podCreationTimestamp="2025-12-07 17:10:19 +0000 UTC" firstStartedPulling="2025-12-07 17:10:20.304329017 +0000 UTC m=+4082.994613929" lastFinishedPulling="2025-12-07 17:10:21.726638151 +0000 UTC m=+4084.416923063" observedRunningTime="2025-12-07 17:10:22.342626967 +0000 UTC m=+4085.032911879" watchObservedRunningTime="2025-12-07 17:10:22.344585611 +0000 UTC m=+4085.034870523" Dec 07 17:10:26 crc kubenswrapper[4716]: I1207 17:10:26.657927 4716 scope.go:117] "RemoveContainer" containerID="0f9119a8d01b371a5ab06e6f944c6501025c243f659e26196309c46fc20c7a5d" Dec 07 17:10:26 crc kubenswrapper[4716]: E1207 17:10:26.659144 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 17:10:29 crc kubenswrapper[4716]: I1207 17:10:29.119316 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-rqbh5_682c523b-c5ad-4e1c-9fd9-969911ef2242/nmstate-console-plugin/0.log" Dec 07 17:10:29 crc kubenswrapper[4716]: I1207 17:10:29.316957 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-89zc4_e4dfc168-4ec1-40d0-9d1a-d7f04063f189/nmstate-handler/0.log" Dec 07 17:10:29 crc kubenswrapper[4716]: I1207 17:10:29.356176 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-gmh8g" Dec 07 17:10:29 crc kubenswrapper[4716]: I1207 17:10:29.356230 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-gmh8g" Dec 07 17:10:29 crc kubenswrapper[4716]: I1207 17:10:29.396619 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-slf82_f54315f3-de9b-48b5-9baf-94e78414c0e2/kube-rbac-proxy/0.log" Dec 07 17:10:29 crc kubenswrapper[4716]: I1207 17:10:29.406333 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-gmh8g" Dec 07 17:10:29 crc kubenswrapper[4716]: I1207 17:10:29.449386 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-slf82_f54315f3-de9b-48b5-9baf-94e78414c0e2/nmstate-metrics/0.log" Dec 07 17:10:29 crc kubenswrapper[4716]: I1207 17:10:29.455611 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-gmh8g" Dec 07 17:10:29 crc kubenswrapper[4716]: I1207 17:10:29.583956 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-s55qn_08d6dea5-f7a4-4895-b768-f6feebceab45/nmstate-operator/0.log" Dec 07 17:10:29 crc kubenswrapper[4716]: I1207 17:10:29.647549 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gmh8g"] Dec 07 17:10:29 crc kubenswrapper[4716]: I1207 17:10:29.648426 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-8tpxz_3546fcc6-820f-4601-9c0e-b652481582d3/nmstate-webhook/0.log" Dec 07 17:10:31 crc kubenswrapper[4716]: I1207 17:10:31.411497 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-gmh8g" podUID="52f0efae-9ebe-4069-b586-029b47cc07ae" containerName="registry-server" containerID="cri-o://3941f3e337b0ed2fdc59351f79120436255f05af860d8b78a964493161b0f1a7" gracePeriod=2 Dec 07 17:10:32 crc kubenswrapper[4716]: I1207 17:10:32.419305 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gmh8g" Dec 07 17:10:32 crc kubenswrapper[4716]: I1207 17:10:32.422617 4716 generic.go:334] "Generic (PLEG): container finished" podID="52f0efae-9ebe-4069-b586-029b47cc07ae" containerID="3941f3e337b0ed2fdc59351f79120436255f05af860d8b78a964493161b0f1a7" exitCode=0 Dec 07 17:10:32 crc kubenswrapper[4716]: I1207 17:10:32.422650 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gmh8g" event={"ID":"52f0efae-9ebe-4069-b586-029b47cc07ae","Type":"ContainerDied","Data":"3941f3e337b0ed2fdc59351f79120436255f05af860d8b78a964493161b0f1a7"} Dec 07 17:10:32 crc kubenswrapper[4716]: I1207 17:10:32.422676 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gmh8g" event={"ID":"52f0efae-9ebe-4069-b586-029b47cc07ae","Type":"ContainerDied","Data":"344370e17eeaa4c0f24ae87d5ac7d4dad5192209322141b48f89559dde37e084"} Dec 07 17:10:32 crc kubenswrapper[4716]: I1207 17:10:32.422678 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gmh8g" Dec 07 17:10:32 crc kubenswrapper[4716]: I1207 17:10:32.422691 4716 scope.go:117] "RemoveContainer" containerID="3941f3e337b0ed2fdc59351f79120436255f05af860d8b78a964493161b0f1a7" Dec 07 17:10:32 crc kubenswrapper[4716]: I1207 17:10:32.451948 4716 scope.go:117] "RemoveContainer" containerID="056429bc0f75f4283005bf8dae5802acbe6f5e1985da6fd877b04d2c2a47e5ed" Dec 07 17:10:32 crc kubenswrapper[4716]: I1207 17:10:32.491428 4716 scope.go:117] "RemoveContainer" containerID="d807c57f4af64df58e7d07d841ac71a3a316375fb1da007294608fb00e87324a" Dec 07 17:10:32 crc kubenswrapper[4716]: I1207 17:10:32.525877 4716 scope.go:117] "RemoveContainer" containerID="3941f3e337b0ed2fdc59351f79120436255f05af860d8b78a964493161b0f1a7" Dec 07 17:10:32 crc kubenswrapper[4716]: E1207 17:10:32.526344 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3941f3e337b0ed2fdc59351f79120436255f05af860d8b78a964493161b0f1a7\": container with ID starting with 3941f3e337b0ed2fdc59351f79120436255f05af860d8b78a964493161b0f1a7 not found: ID does not exist" containerID="3941f3e337b0ed2fdc59351f79120436255f05af860d8b78a964493161b0f1a7" Dec 07 17:10:32 crc kubenswrapper[4716]: I1207 17:10:32.526411 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3941f3e337b0ed2fdc59351f79120436255f05af860d8b78a964493161b0f1a7"} err="failed to get container status \"3941f3e337b0ed2fdc59351f79120436255f05af860d8b78a964493161b0f1a7\": rpc error: code = NotFound desc = could not find container \"3941f3e337b0ed2fdc59351f79120436255f05af860d8b78a964493161b0f1a7\": container with ID starting with 3941f3e337b0ed2fdc59351f79120436255f05af860d8b78a964493161b0f1a7 not found: ID does not exist" Dec 07 17:10:32 crc kubenswrapper[4716]: I1207 17:10:32.526441 4716 scope.go:117] "RemoveContainer" containerID="056429bc0f75f4283005bf8dae5802acbe6f5e1985da6fd877b04d2c2a47e5ed" Dec 07 17:10:32 crc kubenswrapper[4716]: E1207 17:10:32.526790 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"056429bc0f75f4283005bf8dae5802acbe6f5e1985da6fd877b04d2c2a47e5ed\": container with ID starting with 056429bc0f75f4283005bf8dae5802acbe6f5e1985da6fd877b04d2c2a47e5ed not found: ID does not exist" containerID="056429bc0f75f4283005bf8dae5802acbe6f5e1985da6fd877b04d2c2a47e5ed" Dec 07 17:10:32 crc kubenswrapper[4716]: I1207 17:10:32.526850 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"056429bc0f75f4283005bf8dae5802acbe6f5e1985da6fd877b04d2c2a47e5ed"} err="failed to get container status \"056429bc0f75f4283005bf8dae5802acbe6f5e1985da6fd877b04d2c2a47e5ed\": rpc error: code = NotFound desc = could not find container \"056429bc0f75f4283005bf8dae5802acbe6f5e1985da6fd877b04d2c2a47e5ed\": container with ID starting with 056429bc0f75f4283005bf8dae5802acbe6f5e1985da6fd877b04d2c2a47e5ed not found: ID does not exist" Dec 07 17:10:32 crc kubenswrapper[4716]: I1207 17:10:32.526891 4716 scope.go:117] "RemoveContainer" containerID="d807c57f4af64df58e7d07d841ac71a3a316375fb1da007294608fb00e87324a" Dec 07 17:10:32 crc kubenswrapper[4716]: E1207 17:10:32.527534 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d807c57f4af64df58e7d07d841ac71a3a316375fb1da007294608fb00e87324a\": container with ID starting with d807c57f4af64df58e7d07d841ac71a3a316375fb1da007294608fb00e87324a not found: ID does not exist" containerID="d807c57f4af64df58e7d07d841ac71a3a316375fb1da007294608fb00e87324a" Dec 07 17:10:32 crc kubenswrapper[4716]: I1207 17:10:32.527607 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d807c57f4af64df58e7d07d841ac71a3a316375fb1da007294608fb00e87324a"} err="failed to get container status \"d807c57f4af64df58e7d07d841ac71a3a316375fb1da007294608fb00e87324a\": rpc error: code = NotFound desc = could not find container \"d807c57f4af64df58e7d07d841ac71a3a316375fb1da007294608fb00e87324a\": container with ID starting with d807c57f4af64df58e7d07d841ac71a3a316375fb1da007294608fb00e87324a not found: ID does not exist" Dec 07 17:10:32 crc kubenswrapper[4716]: I1207 17:10:32.541354 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52f0efae-9ebe-4069-b586-029b47cc07ae-catalog-content\") pod \"52f0efae-9ebe-4069-b586-029b47cc07ae\" (UID: \"52f0efae-9ebe-4069-b586-029b47cc07ae\") " Dec 07 17:10:32 crc kubenswrapper[4716]: I1207 17:10:32.541401 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52f0efae-9ebe-4069-b586-029b47cc07ae-utilities\") pod \"52f0efae-9ebe-4069-b586-029b47cc07ae\" (UID: \"52f0efae-9ebe-4069-b586-029b47cc07ae\") " Dec 07 17:10:32 crc kubenswrapper[4716]: I1207 17:10:32.541476 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2znxm\" (UniqueName: \"kubernetes.io/projected/52f0efae-9ebe-4069-b586-029b47cc07ae-kube-api-access-2znxm\") pod \"52f0efae-9ebe-4069-b586-029b47cc07ae\" (UID: \"52f0efae-9ebe-4069-b586-029b47cc07ae\") " Dec 07 17:10:32 crc kubenswrapper[4716]: I1207 17:10:32.542599 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/52f0efae-9ebe-4069-b586-029b47cc07ae-utilities" (OuterVolumeSpecName: "utilities") pod "52f0efae-9ebe-4069-b586-029b47cc07ae" (UID: "52f0efae-9ebe-4069-b586-029b47cc07ae"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 17:10:32 crc kubenswrapper[4716]: I1207 17:10:32.549985 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52f0efae-9ebe-4069-b586-029b47cc07ae-kube-api-access-2znxm" (OuterVolumeSpecName: "kube-api-access-2znxm") pod "52f0efae-9ebe-4069-b586-029b47cc07ae" (UID: "52f0efae-9ebe-4069-b586-029b47cc07ae"). InnerVolumeSpecName "kube-api-access-2znxm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 17:10:32 crc kubenswrapper[4716]: I1207 17:10:32.596674 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/52f0efae-9ebe-4069-b586-029b47cc07ae-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "52f0efae-9ebe-4069-b586-029b47cc07ae" (UID: "52f0efae-9ebe-4069-b586-029b47cc07ae"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 17:10:32 crc kubenswrapper[4716]: I1207 17:10:32.643371 4716 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52f0efae-9ebe-4069-b586-029b47cc07ae-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 07 17:10:32 crc kubenswrapper[4716]: I1207 17:10:32.643403 4716 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52f0efae-9ebe-4069-b586-029b47cc07ae-utilities\") on node \"crc\" DevicePath \"\"" Dec 07 17:10:32 crc kubenswrapper[4716]: I1207 17:10:32.643413 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2znxm\" (UniqueName: \"kubernetes.io/projected/52f0efae-9ebe-4069-b586-029b47cc07ae-kube-api-access-2znxm\") on node \"crc\" DevicePath \"\"" Dec 07 17:10:32 crc kubenswrapper[4716]: I1207 17:10:32.758652 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gmh8g"] Dec 07 17:10:32 crc kubenswrapper[4716]: I1207 17:10:32.767309 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-gmh8g"] Dec 07 17:10:33 crc kubenswrapper[4716]: I1207 17:10:33.668778 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="52f0efae-9ebe-4069-b586-029b47cc07ae" path="/var/lib/kubelet/pods/52f0efae-9ebe-4069-b586-029b47cc07ae/volumes" Dec 07 17:10:39 crc kubenswrapper[4716]: I1207 17:10:39.658786 4716 scope.go:117] "RemoveContainer" containerID="0f9119a8d01b371a5ab06e6f944c6501025c243f659e26196309c46fc20c7a5d" Dec 07 17:10:39 crc kubenswrapper[4716]: E1207 17:10:39.659550 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 17:10:44 crc kubenswrapper[4716]: I1207 17:10:44.614422 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-t728x_f233fcdd-5c35-4ffb-bb33-7e2249f1c6c0/kube-rbac-proxy/0.log" Dec 07 17:10:44 crc kubenswrapper[4716]: I1207 17:10:44.740257 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-t728x_f233fcdd-5c35-4ffb-bb33-7e2249f1c6c0/controller/0.log" Dec 07 17:10:44 crc kubenswrapper[4716]: I1207 17:10:44.819519 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9zdjv_5e169ec0-a81a-45eb-b2a6-8ac97d905caa/cp-frr-files/0.log" Dec 07 17:10:44 crc kubenswrapper[4716]: I1207 17:10:44.980234 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9zdjv_5e169ec0-a81a-45eb-b2a6-8ac97d905caa/cp-reloader/0.log" Dec 07 17:10:44 crc kubenswrapper[4716]: I1207 17:10:44.986306 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9zdjv_5e169ec0-a81a-45eb-b2a6-8ac97d905caa/cp-metrics/0.log" Dec 07 17:10:45 crc kubenswrapper[4716]: I1207 17:10:45.025211 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9zdjv_5e169ec0-a81a-45eb-b2a6-8ac97d905caa/cp-reloader/0.log" Dec 07 17:10:45 crc kubenswrapper[4716]: I1207 17:10:45.051190 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9zdjv_5e169ec0-a81a-45eb-b2a6-8ac97d905caa/cp-frr-files/0.log" Dec 07 17:10:45 crc kubenswrapper[4716]: I1207 17:10:45.549491 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9zdjv_5e169ec0-a81a-45eb-b2a6-8ac97d905caa/cp-metrics/0.log" Dec 07 17:10:45 crc kubenswrapper[4716]: I1207 17:10:45.551638 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9zdjv_5e169ec0-a81a-45eb-b2a6-8ac97d905caa/cp-reloader/0.log" Dec 07 17:10:45 crc kubenswrapper[4716]: I1207 17:10:45.583683 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9zdjv_5e169ec0-a81a-45eb-b2a6-8ac97d905caa/cp-frr-files/0.log" Dec 07 17:10:45 crc kubenswrapper[4716]: I1207 17:10:45.600648 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9zdjv_5e169ec0-a81a-45eb-b2a6-8ac97d905caa/cp-metrics/0.log" Dec 07 17:10:45 crc kubenswrapper[4716]: I1207 17:10:45.817694 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9zdjv_5e169ec0-a81a-45eb-b2a6-8ac97d905caa/cp-metrics/0.log" Dec 07 17:10:45 crc kubenswrapper[4716]: I1207 17:10:45.826169 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9zdjv_5e169ec0-a81a-45eb-b2a6-8ac97d905caa/cp-frr-files/0.log" Dec 07 17:10:45 crc kubenswrapper[4716]: I1207 17:10:45.843673 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9zdjv_5e169ec0-a81a-45eb-b2a6-8ac97d905caa/cp-reloader/0.log" Dec 07 17:10:45 crc kubenswrapper[4716]: I1207 17:10:45.851006 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9zdjv_5e169ec0-a81a-45eb-b2a6-8ac97d905caa/controller/0.log" Dec 07 17:10:46 crc kubenswrapper[4716]: I1207 17:10:46.006174 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9zdjv_5e169ec0-a81a-45eb-b2a6-8ac97d905caa/kube-rbac-proxy/0.log" Dec 07 17:10:46 crc kubenswrapper[4716]: I1207 17:10:46.008807 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9zdjv_5e169ec0-a81a-45eb-b2a6-8ac97d905caa/frr-metrics/0.log" Dec 07 17:10:46 crc kubenswrapper[4716]: I1207 17:10:46.046212 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9zdjv_5e169ec0-a81a-45eb-b2a6-8ac97d905caa/kube-rbac-proxy-frr/0.log" Dec 07 17:10:46 crc kubenswrapper[4716]: I1207 17:10:46.180932 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9zdjv_5e169ec0-a81a-45eb-b2a6-8ac97d905caa/reloader/0.log" Dec 07 17:10:46 crc kubenswrapper[4716]: I1207 17:10:46.273239 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-dqzl5_1a4973da-40f0-4277-a6ff-06ff961b4815/frr-k8s-webhook-server/0.log" Dec 07 17:10:46 crc kubenswrapper[4716]: I1207 17:10:46.492711 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-655c466c6d-8pvn6_049c277a-93fc-42db-952b-b1549f6a599c/manager/0.log" Dec 07 17:10:46 crc kubenswrapper[4716]: I1207 17:10:46.665063 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-65fbf78689-nbmlk_b6bf63fb-cc3e-4349-b7f6-efb11c1ff14f/webhook-server/0.log" Dec 07 17:10:46 crc kubenswrapper[4716]: I1207 17:10:46.733750 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-xnsb6_82d3c78d-32ff-47ed-854a-d3786c962ffd/kube-rbac-proxy/0.log" Dec 07 17:10:47 crc kubenswrapper[4716]: I1207 17:10:47.214995 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9zdjv_5e169ec0-a81a-45eb-b2a6-8ac97d905caa/frr/0.log" Dec 07 17:10:47 crc kubenswrapper[4716]: I1207 17:10:47.561817 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-xnsb6_82d3c78d-32ff-47ed-854a-d3786c962ffd/speaker/0.log" Dec 07 17:10:54 crc kubenswrapper[4716]: I1207 17:10:54.658565 4716 scope.go:117] "RemoveContainer" containerID="0f9119a8d01b371a5ab06e6f944c6501025c243f659e26196309c46fc20c7a5d" Dec 07 17:10:54 crc kubenswrapper[4716]: E1207 17:10:54.659271 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 17:10:59 crc kubenswrapper[4716]: I1207 17:10:59.464402 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fh9qqw_7ecea701-6ded-4f27-844f-fd9dfcb92f29/util/0.log" Dec 07 17:10:59 crc kubenswrapper[4716]: I1207 17:10:59.614007 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fh9qqw_7ecea701-6ded-4f27-844f-fd9dfcb92f29/util/0.log" Dec 07 17:10:59 crc kubenswrapper[4716]: I1207 17:10:59.626909 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fh9qqw_7ecea701-6ded-4f27-844f-fd9dfcb92f29/pull/0.log" Dec 07 17:10:59 crc kubenswrapper[4716]: I1207 17:10:59.688860 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fh9qqw_7ecea701-6ded-4f27-844f-fd9dfcb92f29/pull/0.log" Dec 07 17:10:59 crc kubenswrapper[4716]: I1207 17:10:59.852050 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fh9qqw_7ecea701-6ded-4f27-844f-fd9dfcb92f29/pull/0.log" Dec 07 17:10:59 crc kubenswrapper[4716]: I1207 17:10:59.866870 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fh9qqw_7ecea701-6ded-4f27-844f-fd9dfcb92f29/util/0.log" Dec 07 17:10:59 crc kubenswrapper[4716]: I1207 17:10:59.880653 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fh9qqw_7ecea701-6ded-4f27-844f-fd9dfcb92f29/extract/0.log" Dec 07 17:11:00 crc kubenswrapper[4716]: I1207 17:11:00.012499 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c2vdd_19363c75-987f-4c3b-bbca-28a9cc33238b/util/0.log" Dec 07 17:11:00 crc kubenswrapper[4716]: I1207 17:11:00.189307 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c2vdd_19363c75-987f-4c3b-bbca-28a9cc33238b/pull/0.log" Dec 07 17:11:00 crc kubenswrapper[4716]: I1207 17:11:00.202555 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c2vdd_19363c75-987f-4c3b-bbca-28a9cc33238b/util/0.log" Dec 07 17:11:00 crc kubenswrapper[4716]: I1207 17:11:00.210730 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c2vdd_19363c75-987f-4c3b-bbca-28a9cc33238b/pull/0.log" Dec 07 17:11:00 crc kubenswrapper[4716]: I1207 17:11:00.395854 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c2vdd_19363c75-987f-4c3b-bbca-28a9cc33238b/util/0.log" Dec 07 17:11:00 crc kubenswrapper[4716]: I1207 17:11:00.404113 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c2vdd_19363c75-987f-4c3b-bbca-28a9cc33238b/extract/0.log" Dec 07 17:11:00 crc kubenswrapper[4716]: I1207 17:11:00.412601 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83c2vdd_19363c75-987f-4c3b-bbca-28a9cc33238b/pull/0.log" Dec 07 17:11:00 crc kubenswrapper[4716]: I1207 17:11:00.568826 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tgn9w_fc342361-554a-41c0-a512-06e046ff03eb/extract-utilities/0.log" Dec 07 17:11:00 crc kubenswrapper[4716]: I1207 17:11:00.748226 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tgn9w_fc342361-554a-41c0-a512-06e046ff03eb/extract-content/0.log" Dec 07 17:11:00 crc kubenswrapper[4716]: I1207 17:11:00.754841 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tgn9w_fc342361-554a-41c0-a512-06e046ff03eb/extract-content/0.log" Dec 07 17:11:00 crc kubenswrapper[4716]: I1207 17:11:00.779580 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tgn9w_fc342361-554a-41c0-a512-06e046ff03eb/extract-utilities/0.log" Dec 07 17:11:00 crc kubenswrapper[4716]: I1207 17:11:00.929674 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tgn9w_fc342361-554a-41c0-a512-06e046ff03eb/extract-utilities/0.log" Dec 07 17:11:00 crc kubenswrapper[4716]: I1207 17:11:00.957938 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tgn9w_fc342361-554a-41c0-a512-06e046ff03eb/extract-content/0.log" Dec 07 17:11:01 crc kubenswrapper[4716]: I1207 17:11:01.329596 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-vmnk2_33db3f8a-1f66-44c4-b563-2f07dbc8267e/extract-utilities/0.log" Dec 07 17:11:01 crc kubenswrapper[4716]: I1207 17:11:01.344252 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tgn9w_fc342361-554a-41c0-a512-06e046ff03eb/registry-server/0.log" Dec 07 17:11:01 crc kubenswrapper[4716]: I1207 17:11:01.416825 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-vmnk2_33db3f8a-1f66-44c4-b563-2f07dbc8267e/extract-utilities/0.log" Dec 07 17:11:01 crc kubenswrapper[4716]: I1207 17:11:01.452656 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-vmnk2_33db3f8a-1f66-44c4-b563-2f07dbc8267e/extract-content/0.log" Dec 07 17:11:01 crc kubenswrapper[4716]: I1207 17:11:01.479996 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-vmnk2_33db3f8a-1f66-44c4-b563-2f07dbc8267e/extract-content/0.log" Dec 07 17:11:01 crc kubenswrapper[4716]: I1207 17:11:01.657620 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-vmnk2_33db3f8a-1f66-44c4-b563-2f07dbc8267e/extract-utilities/0.log" Dec 07 17:11:01 crc kubenswrapper[4716]: I1207 17:11:01.680704 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-vmnk2_33db3f8a-1f66-44c4-b563-2f07dbc8267e/extract-content/0.log" Dec 07 17:11:01 crc kubenswrapper[4716]: I1207 17:11:01.919009 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f7pqt_0f030783-1303-4815-8ac1-fbecf8e036fe/extract-utilities/0.log" Dec 07 17:11:01 crc kubenswrapper[4716]: I1207 17:11:01.972707 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-dhjm4_845614d7-aa72-46de-b358-2e39f0209886/marketplace-operator/0.log" Dec 07 17:11:02 crc kubenswrapper[4716]: I1207 17:11:02.176287 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-vmnk2_33db3f8a-1f66-44c4-b563-2f07dbc8267e/registry-server/0.log" Dec 07 17:11:02 crc kubenswrapper[4716]: I1207 17:11:02.176334 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f7pqt_0f030783-1303-4815-8ac1-fbecf8e036fe/extract-utilities/0.log" Dec 07 17:11:02 crc kubenswrapper[4716]: I1207 17:11:02.176411 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f7pqt_0f030783-1303-4815-8ac1-fbecf8e036fe/extract-content/0.log" Dec 07 17:11:02 crc kubenswrapper[4716]: I1207 17:11:02.195260 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f7pqt_0f030783-1303-4815-8ac1-fbecf8e036fe/extract-content/0.log" Dec 07 17:11:02 crc kubenswrapper[4716]: I1207 17:11:02.355849 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f7pqt_0f030783-1303-4815-8ac1-fbecf8e036fe/extract-utilities/0.log" Dec 07 17:11:02 crc kubenswrapper[4716]: I1207 17:11:02.399227 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f7pqt_0f030783-1303-4815-8ac1-fbecf8e036fe/extract-content/0.log" Dec 07 17:11:02 crc kubenswrapper[4716]: I1207 17:11:02.523685 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f7pqt_0f030783-1303-4815-8ac1-fbecf8e036fe/registry-server/0.log" Dec 07 17:11:02 crc kubenswrapper[4716]: I1207 17:11:02.566292 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-nkjl4_af948d72-6e6c-42d9-9d0b-1821d7f47176/extract-utilities/0.log" Dec 07 17:11:02 crc kubenswrapper[4716]: I1207 17:11:02.714192 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-nkjl4_af948d72-6e6c-42d9-9d0b-1821d7f47176/extract-content/0.log" Dec 07 17:11:02 crc kubenswrapper[4716]: I1207 17:11:02.726770 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-nkjl4_af948d72-6e6c-42d9-9d0b-1821d7f47176/extract-content/0.log" Dec 07 17:11:02 crc kubenswrapper[4716]: I1207 17:11:02.737446 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-nkjl4_af948d72-6e6c-42d9-9d0b-1821d7f47176/extract-utilities/0.log" Dec 07 17:11:02 crc kubenswrapper[4716]: I1207 17:11:02.890451 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-nkjl4_af948d72-6e6c-42d9-9d0b-1821d7f47176/extract-content/0.log" Dec 07 17:11:02 crc kubenswrapper[4716]: I1207 17:11:02.909267 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-nkjl4_af948d72-6e6c-42d9-9d0b-1821d7f47176/extract-utilities/0.log" Dec 07 17:11:03 crc kubenswrapper[4716]: I1207 17:11:03.345789 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-nkjl4_af948d72-6e6c-42d9-9d0b-1821d7f47176/registry-server/0.log" Dec 07 17:11:07 crc kubenswrapper[4716]: I1207 17:11:07.665524 4716 scope.go:117] "RemoveContainer" containerID="0f9119a8d01b371a5ab06e6f944c6501025c243f659e26196309c46fc20c7a5d" Dec 07 17:11:07 crc kubenswrapper[4716]: E1207 17:11:07.666486 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 17:11:20 crc kubenswrapper[4716]: I1207 17:11:20.658226 4716 scope.go:117] "RemoveContainer" containerID="0f9119a8d01b371a5ab06e6f944c6501025c243f659e26196309c46fc20c7a5d" Dec 07 17:11:20 crc kubenswrapper[4716]: E1207 17:11:20.658965 4716 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zcxxp_openshift-machine-config-operator(c15b59eb-565d-4556-a4ce-75afdf159dc8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" Dec 07 17:11:35 crc kubenswrapper[4716]: I1207 17:11:35.659099 4716 scope.go:117] "RemoveContainer" containerID="0f9119a8d01b371a5ab06e6f944c6501025c243f659e26196309c46fc20c7a5d" Dec 07 17:11:35 crc kubenswrapper[4716]: I1207 17:11:35.976370 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" event={"ID":"c15b59eb-565d-4556-a4ce-75afdf159dc8","Type":"ContainerStarted","Data":"3bf73ccc579d3403208c446ebe67836eb58ddd9923228f63417db3fd236428b9"} Dec 07 17:12:41 crc kubenswrapper[4716]: I1207 17:12:41.711274 4716 generic.go:334] "Generic (PLEG): container finished" podID="6e77ed85-7b90-4d72-899c-13feb6b4615c" containerID="089983d99023ead848e9c460a5fefaaa5fbba5e6ae9dc60751cf8b8338fb5d97" exitCode=0 Dec 07 17:12:41 crc kubenswrapper[4716]: I1207 17:12:41.711364 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x5c24/must-gather-tcwnq" event={"ID":"6e77ed85-7b90-4d72-899c-13feb6b4615c","Type":"ContainerDied","Data":"089983d99023ead848e9c460a5fefaaa5fbba5e6ae9dc60751cf8b8338fb5d97"} Dec 07 17:12:41 crc kubenswrapper[4716]: I1207 17:12:41.712451 4716 scope.go:117] "RemoveContainer" containerID="089983d99023ead848e9c460a5fefaaa5fbba5e6ae9dc60751cf8b8338fb5d97" Dec 07 17:12:42 crc kubenswrapper[4716]: I1207 17:12:42.190590 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-x5c24_must-gather-tcwnq_6e77ed85-7b90-4d72-899c-13feb6b4615c/gather/0.log" Dec 07 17:12:52 crc kubenswrapper[4716]: I1207 17:12:52.947939 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-x5c24/must-gather-tcwnq"] Dec 07 17:12:52 crc kubenswrapper[4716]: I1207 17:12:52.948760 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-x5c24/must-gather-tcwnq" podUID="6e77ed85-7b90-4d72-899c-13feb6b4615c" containerName="copy" containerID="cri-o://d524df7d9598c1df472b479d6b590a726bfc512986882b2fd7ccb3c9ede5d1ab" gracePeriod=2 Dec 07 17:12:52 crc kubenswrapper[4716]: I1207 17:12:52.957061 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-x5c24/must-gather-tcwnq"] Dec 07 17:12:53 crc kubenswrapper[4716]: I1207 17:12:53.837880 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-x5c24_must-gather-tcwnq_6e77ed85-7b90-4d72-899c-13feb6b4615c/copy/0.log" Dec 07 17:12:53 crc kubenswrapper[4716]: I1207 17:12:53.838486 4716 generic.go:334] "Generic (PLEG): container finished" podID="6e77ed85-7b90-4d72-899c-13feb6b4615c" containerID="d524df7d9598c1df472b479d6b590a726bfc512986882b2fd7ccb3c9ede5d1ab" exitCode=143 Dec 07 17:12:54 crc kubenswrapper[4716]: I1207 17:12:54.273744 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-x5c24_must-gather-tcwnq_6e77ed85-7b90-4d72-899c-13feb6b4615c/copy/0.log" Dec 07 17:12:54 crc kubenswrapper[4716]: I1207 17:12:54.274643 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x5c24/must-gather-tcwnq" Dec 07 17:12:54 crc kubenswrapper[4716]: I1207 17:12:54.456444 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/6e77ed85-7b90-4d72-899c-13feb6b4615c-must-gather-output\") pod \"6e77ed85-7b90-4d72-899c-13feb6b4615c\" (UID: \"6e77ed85-7b90-4d72-899c-13feb6b4615c\") " Dec 07 17:12:54 crc kubenswrapper[4716]: I1207 17:12:54.456618 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wtftc\" (UniqueName: \"kubernetes.io/projected/6e77ed85-7b90-4d72-899c-13feb6b4615c-kube-api-access-wtftc\") pod \"6e77ed85-7b90-4d72-899c-13feb6b4615c\" (UID: \"6e77ed85-7b90-4d72-899c-13feb6b4615c\") " Dec 07 17:12:54 crc kubenswrapper[4716]: I1207 17:12:54.462818 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e77ed85-7b90-4d72-899c-13feb6b4615c-kube-api-access-wtftc" (OuterVolumeSpecName: "kube-api-access-wtftc") pod "6e77ed85-7b90-4d72-899c-13feb6b4615c" (UID: "6e77ed85-7b90-4d72-899c-13feb6b4615c"). InnerVolumeSpecName "kube-api-access-wtftc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 17:12:54 crc kubenswrapper[4716]: I1207 17:12:54.558714 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wtftc\" (UniqueName: \"kubernetes.io/projected/6e77ed85-7b90-4d72-899c-13feb6b4615c-kube-api-access-wtftc\") on node \"crc\" DevicePath \"\"" Dec 07 17:12:54 crc kubenswrapper[4716]: I1207 17:12:54.600323 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6e77ed85-7b90-4d72-899c-13feb6b4615c-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "6e77ed85-7b90-4d72-899c-13feb6b4615c" (UID: "6e77ed85-7b90-4d72-899c-13feb6b4615c"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 17:12:54 crc kubenswrapper[4716]: I1207 17:12:54.660247 4716 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/6e77ed85-7b90-4d72-899c-13feb6b4615c-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 07 17:12:54 crc kubenswrapper[4716]: I1207 17:12:54.852203 4716 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-x5c24_must-gather-tcwnq_6e77ed85-7b90-4d72-899c-13feb6b4615c/copy/0.log" Dec 07 17:12:54 crc kubenswrapper[4716]: I1207 17:12:54.853312 4716 scope.go:117] "RemoveContainer" containerID="d524df7d9598c1df472b479d6b590a726bfc512986882b2fd7ccb3c9ede5d1ab" Dec 07 17:12:54 crc kubenswrapper[4716]: I1207 17:12:54.853352 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x5c24/must-gather-tcwnq" Dec 07 17:12:54 crc kubenswrapper[4716]: I1207 17:12:54.882241 4716 scope.go:117] "RemoveContainer" containerID="089983d99023ead848e9c460a5fefaaa5fbba5e6ae9dc60751cf8b8338fb5d97" Dec 07 17:12:55 crc kubenswrapper[4716]: I1207 17:12:55.671468 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e77ed85-7b90-4d72-899c-13feb6b4615c" path="/var/lib/kubelet/pods/6e77ed85-7b90-4d72-899c-13feb6b4615c/volumes" Dec 07 17:13:52 crc kubenswrapper[4716]: I1207 17:13:52.761900 4716 patch_prober.go:28] interesting pod/machine-config-daemon-zcxxp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 17:13:52 crc kubenswrapper[4716]: I1207 17:13:52.762518 4716 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 17:14:22 crc kubenswrapper[4716]: I1207 17:14:22.761289 4716 patch_prober.go:28] interesting pod/machine-config-daemon-zcxxp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 17:14:22 crc kubenswrapper[4716]: I1207 17:14:22.762225 4716 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 17:14:35 crc kubenswrapper[4716]: I1207 17:14:35.820298 4716 scope.go:117] "RemoveContainer" containerID="cee5e066fc428f048cddcdf7009e2528466a64db0dc946ab9d354efbefe3967b" Dec 07 17:14:45 crc kubenswrapper[4716]: I1207 17:14:45.505751 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-h5jbn"] Dec 07 17:14:45 crc kubenswrapper[4716]: E1207 17:14:45.506788 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52f0efae-9ebe-4069-b586-029b47cc07ae" containerName="extract-content" Dec 07 17:14:45 crc kubenswrapper[4716]: I1207 17:14:45.506803 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="52f0efae-9ebe-4069-b586-029b47cc07ae" containerName="extract-content" Dec 07 17:14:45 crc kubenswrapper[4716]: E1207 17:14:45.506837 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e77ed85-7b90-4d72-899c-13feb6b4615c" containerName="gather" Dec 07 17:14:45 crc kubenswrapper[4716]: I1207 17:14:45.506846 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e77ed85-7b90-4d72-899c-13feb6b4615c" containerName="gather" Dec 07 17:14:45 crc kubenswrapper[4716]: E1207 17:14:45.506859 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52f0efae-9ebe-4069-b586-029b47cc07ae" containerName="registry-server" Dec 07 17:14:45 crc kubenswrapper[4716]: I1207 17:14:45.506868 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="52f0efae-9ebe-4069-b586-029b47cc07ae" containerName="registry-server" Dec 07 17:14:45 crc kubenswrapper[4716]: E1207 17:14:45.506897 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52f0efae-9ebe-4069-b586-029b47cc07ae" containerName="extract-utilities" Dec 07 17:14:45 crc kubenswrapper[4716]: I1207 17:14:45.506906 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="52f0efae-9ebe-4069-b586-029b47cc07ae" containerName="extract-utilities" Dec 07 17:14:45 crc kubenswrapper[4716]: E1207 17:14:45.506929 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e77ed85-7b90-4d72-899c-13feb6b4615c" containerName="copy" Dec 07 17:14:45 crc kubenswrapper[4716]: I1207 17:14:45.506937 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e77ed85-7b90-4d72-899c-13feb6b4615c" containerName="copy" Dec 07 17:14:45 crc kubenswrapper[4716]: I1207 17:14:45.507190 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e77ed85-7b90-4d72-899c-13feb6b4615c" containerName="copy" Dec 07 17:14:45 crc kubenswrapper[4716]: I1207 17:14:45.507203 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="52f0efae-9ebe-4069-b586-029b47cc07ae" containerName="registry-server" Dec 07 17:14:45 crc kubenswrapper[4716]: I1207 17:14:45.507228 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e77ed85-7b90-4d72-899c-13feb6b4615c" containerName="gather" Dec 07 17:14:45 crc kubenswrapper[4716]: I1207 17:14:45.509020 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h5jbn" Dec 07 17:14:45 crc kubenswrapper[4716]: I1207 17:14:45.509962 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-h5jbn"] Dec 07 17:14:45 crc kubenswrapper[4716]: I1207 17:14:45.554942 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09c7e427-ef3b-4816-805d-d811be5b4337-catalog-content\") pod \"community-operators-h5jbn\" (UID: \"09c7e427-ef3b-4816-805d-d811be5b4337\") " pod="openshift-marketplace/community-operators-h5jbn" Dec 07 17:14:45 crc kubenswrapper[4716]: I1207 17:14:45.555450 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09c7e427-ef3b-4816-805d-d811be5b4337-utilities\") pod \"community-operators-h5jbn\" (UID: \"09c7e427-ef3b-4816-805d-d811be5b4337\") " pod="openshift-marketplace/community-operators-h5jbn" Dec 07 17:14:45 crc kubenswrapper[4716]: I1207 17:14:45.555542 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-422j6\" (UniqueName: \"kubernetes.io/projected/09c7e427-ef3b-4816-805d-d811be5b4337-kube-api-access-422j6\") pod \"community-operators-h5jbn\" (UID: \"09c7e427-ef3b-4816-805d-d811be5b4337\") " pod="openshift-marketplace/community-operators-h5jbn" Dec 07 17:14:45 crc kubenswrapper[4716]: I1207 17:14:45.657045 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09c7e427-ef3b-4816-805d-d811be5b4337-catalog-content\") pod \"community-operators-h5jbn\" (UID: \"09c7e427-ef3b-4816-805d-d811be5b4337\") " pod="openshift-marketplace/community-operators-h5jbn" Dec 07 17:14:45 crc kubenswrapper[4716]: I1207 17:14:45.657191 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09c7e427-ef3b-4816-805d-d811be5b4337-utilities\") pod \"community-operators-h5jbn\" (UID: \"09c7e427-ef3b-4816-805d-d811be5b4337\") " pod="openshift-marketplace/community-operators-h5jbn" Dec 07 17:14:45 crc kubenswrapper[4716]: I1207 17:14:45.657226 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-422j6\" (UniqueName: \"kubernetes.io/projected/09c7e427-ef3b-4816-805d-d811be5b4337-kube-api-access-422j6\") pod \"community-operators-h5jbn\" (UID: \"09c7e427-ef3b-4816-805d-d811be5b4337\") " pod="openshift-marketplace/community-operators-h5jbn" Dec 07 17:14:45 crc kubenswrapper[4716]: I1207 17:14:45.658071 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09c7e427-ef3b-4816-805d-d811be5b4337-catalog-content\") pod \"community-operators-h5jbn\" (UID: \"09c7e427-ef3b-4816-805d-d811be5b4337\") " pod="openshift-marketplace/community-operators-h5jbn" Dec 07 17:14:45 crc kubenswrapper[4716]: I1207 17:14:45.658818 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09c7e427-ef3b-4816-805d-d811be5b4337-utilities\") pod \"community-operators-h5jbn\" (UID: \"09c7e427-ef3b-4816-805d-d811be5b4337\") " pod="openshift-marketplace/community-operators-h5jbn" Dec 07 17:14:45 crc kubenswrapper[4716]: I1207 17:14:45.690135 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-422j6\" (UniqueName: \"kubernetes.io/projected/09c7e427-ef3b-4816-805d-d811be5b4337-kube-api-access-422j6\") pod \"community-operators-h5jbn\" (UID: \"09c7e427-ef3b-4816-805d-d811be5b4337\") " pod="openshift-marketplace/community-operators-h5jbn" Dec 07 17:14:45 crc kubenswrapper[4716]: I1207 17:14:45.850350 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h5jbn" Dec 07 17:14:46 crc kubenswrapper[4716]: I1207 17:14:46.383603 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-h5jbn"] Dec 07 17:14:47 crc kubenswrapper[4716]: I1207 17:14:47.260425 4716 generic.go:334] "Generic (PLEG): container finished" podID="09c7e427-ef3b-4816-805d-d811be5b4337" containerID="e3fc6065725a0acadbcf14cf9930b9879eff82e15e623667dab42cfd3865d956" exitCode=0 Dec 07 17:14:47 crc kubenswrapper[4716]: I1207 17:14:47.260473 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h5jbn" event={"ID":"09c7e427-ef3b-4816-805d-d811be5b4337","Type":"ContainerDied","Data":"e3fc6065725a0acadbcf14cf9930b9879eff82e15e623667dab42cfd3865d956"} Dec 07 17:14:47 crc kubenswrapper[4716]: I1207 17:14:47.261972 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h5jbn" event={"ID":"09c7e427-ef3b-4816-805d-d811be5b4337","Type":"ContainerStarted","Data":"4b9061df2a3818fe85be9c7b3f14e78504b3e4b6892831340fbc8ce955e0261c"} Dec 07 17:14:47 crc kubenswrapper[4716]: I1207 17:14:47.265386 4716 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 07 17:14:48 crc kubenswrapper[4716]: I1207 17:14:48.300664 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h5jbn" event={"ID":"09c7e427-ef3b-4816-805d-d811be5b4337","Type":"ContainerStarted","Data":"95e647ce5c332e2dd3003af717199c35309de162794f5c314c33968464818561"} Dec 07 17:14:49 crc kubenswrapper[4716]: I1207 17:14:49.311543 4716 generic.go:334] "Generic (PLEG): container finished" podID="09c7e427-ef3b-4816-805d-d811be5b4337" containerID="95e647ce5c332e2dd3003af717199c35309de162794f5c314c33968464818561" exitCode=0 Dec 07 17:14:49 crc kubenswrapper[4716]: I1207 17:14:49.311589 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h5jbn" event={"ID":"09c7e427-ef3b-4816-805d-d811be5b4337","Type":"ContainerDied","Data":"95e647ce5c332e2dd3003af717199c35309de162794f5c314c33968464818561"} Dec 07 17:14:50 crc kubenswrapper[4716]: I1207 17:14:50.321589 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h5jbn" event={"ID":"09c7e427-ef3b-4816-805d-d811be5b4337","Type":"ContainerStarted","Data":"de129226a1ca306a25e24106e499754f5c6833838e7f052ecc0df93b296762b4"} Dec 07 17:14:50 crc kubenswrapper[4716]: I1207 17:14:50.353336 4716 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-h5jbn" podStartSLOduration=2.66330554 podStartE2EDuration="5.353307343s" podCreationTimestamp="2025-12-07 17:14:45 +0000 UTC" firstStartedPulling="2025-12-07 17:14:47.265140106 +0000 UTC m=+4349.955425018" lastFinishedPulling="2025-12-07 17:14:49.955141889 +0000 UTC m=+4352.645426821" observedRunningTime="2025-12-07 17:14:50.340477285 +0000 UTC m=+4353.030762207" watchObservedRunningTime="2025-12-07 17:14:50.353307343 +0000 UTC m=+4353.043592275" Dec 07 17:14:52 crc kubenswrapper[4716]: I1207 17:14:52.761958 4716 patch_prober.go:28] interesting pod/machine-config-daemon-zcxxp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 07 17:14:52 crc kubenswrapper[4716]: I1207 17:14:52.762547 4716 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 07 17:14:52 crc kubenswrapper[4716]: I1207 17:14:52.762603 4716 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" Dec 07 17:14:52 crc kubenswrapper[4716]: I1207 17:14:52.763401 4716 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3bf73ccc579d3403208c446ebe67836eb58ddd9923228f63417db3fd236428b9"} pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 07 17:14:52 crc kubenswrapper[4716]: I1207 17:14:52.763469 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" podUID="c15b59eb-565d-4556-a4ce-75afdf159dc8" containerName="machine-config-daemon" containerID="cri-o://3bf73ccc579d3403208c446ebe67836eb58ddd9923228f63417db3fd236428b9" gracePeriod=600 Dec 07 17:14:53 crc kubenswrapper[4716]: I1207 17:14:53.371820 4716 generic.go:334] "Generic (PLEG): container finished" podID="c15b59eb-565d-4556-a4ce-75afdf159dc8" containerID="3bf73ccc579d3403208c446ebe67836eb58ddd9923228f63417db3fd236428b9" exitCode=0 Dec 07 17:14:53 crc kubenswrapper[4716]: I1207 17:14:53.371877 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" event={"ID":"c15b59eb-565d-4556-a4ce-75afdf159dc8","Type":"ContainerDied","Data":"3bf73ccc579d3403208c446ebe67836eb58ddd9923228f63417db3fd236428b9"} Dec 07 17:14:53 crc kubenswrapper[4716]: I1207 17:14:53.372256 4716 scope.go:117] "RemoveContainer" containerID="0f9119a8d01b371a5ab06e6f944c6501025c243f659e26196309c46fc20c7a5d" Dec 07 17:14:54 crc kubenswrapper[4716]: I1207 17:14:54.383337 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zcxxp" event={"ID":"c15b59eb-565d-4556-a4ce-75afdf159dc8","Type":"ContainerStarted","Data":"b246348910474bf96dc36235fccd4500dcfcd2b4b79dae0428b3e710d254e5b9"} Dec 07 17:14:55 crc kubenswrapper[4716]: I1207 17:14:55.850528 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-h5jbn" Dec 07 17:14:55 crc kubenswrapper[4716]: I1207 17:14:55.850770 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-h5jbn" Dec 07 17:14:56 crc kubenswrapper[4716]: I1207 17:14:56.211899 4716 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-h5jbn" Dec 07 17:14:56 crc kubenswrapper[4716]: I1207 17:14:56.460031 4716 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-h5jbn" Dec 07 17:14:56 crc kubenswrapper[4716]: I1207 17:14:56.512309 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-h5jbn"] Dec 07 17:14:58 crc kubenswrapper[4716]: I1207 17:14:58.419931 4716 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-h5jbn" podUID="09c7e427-ef3b-4816-805d-d811be5b4337" containerName="registry-server" containerID="cri-o://de129226a1ca306a25e24106e499754f5c6833838e7f052ecc0df93b296762b4" gracePeriod=2 Dec 07 17:14:59 crc kubenswrapper[4716]: I1207 17:14:59.312553 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h5jbn" Dec 07 17:14:59 crc kubenswrapper[4716]: I1207 17:14:59.346944 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09c7e427-ef3b-4816-805d-d811be5b4337-catalog-content\") pod \"09c7e427-ef3b-4816-805d-d811be5b4337\" (UID: \"09c7e427-ef3b-4816-805d-d811be5b4337\") " Dec 07 17:14:59 crc kubenswrapper[4716]: I1207 17:14:59.347137 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-422j6\" (UniqueName: \"kubernetes.io/projected/09c7e427-ef3b-4816-805d-d811be5b4337-kube-api-access-422j6\") pod \"09c7e427-ef3b-4816-805d-d811be5b4337\" (UID: \"09c7e427-ef3b-4816-805d-d811be5b4337\") " Dec 07 17:14:59 crc kubenswrapper[4716]: I1207 17:14:59.347197 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09c7e427-ef3b-4816-805d-d811be5b4337-utilities\") pod \"09c7e427-ef3b-4816-805d-d811be5b4337\" (UID: \"09c7e427-ef3b-4816-805d-d811be5b4337\") " Dec 07 17:14:59 crc kubenswrapper[4716]: I1207 17:14:59.348401 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/09c7e427-ef3b-4816-805d-d811be5b4337-utilities" (OuterVolumeSpecName: "utilities") pod "09c7e427-ef3b-4816-805d-d811be5b4337" (UID: "09c7e427-ef3b-4816-805d-d811be5b4337"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 17:14:59 crc kubenswrapper[4716]: I1207 17:14:59.357232 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09c7e427-ef3b-4816-805d-d811be5b4337-kube-api-access-422j6" (OuterVolumeSpecName: "kube-api-access-422j6") pod "09c7e427-ef3b-4816-805d-d811be5b4337" (UID: "09c7e427-ef3b-4816-805d-d811be5b4337"). InnerVolumeSpecName "kube-api-access-422j6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 17:14:59 crc kubenswrapper[4716]: I1207 17:14:59.397395 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/09c7e427-ef3b-4816-805d-d811be5b4337-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "09c7e427-ef3b-4816-805d-d811be5b4337" (UID: "09c7e427-ef3b-4816-805d-d811be5b4337"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 07 17:14:59 crc kubenswrapper[4716]: I1207 17:14:59.431778 4716 generic.go:334] "Generic (PLEG): container finished" podID="09c7e427-ef3b-4816-805d-d811be5b4337" containerID="de129226a1ca306a25e24106e499754f5c6833838e7f052ecc0df93b296762b4" exitCode=0 Dec 07 17:14:59 crc kubenswrapper[4716]: I1207 17:14:59.431817 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h5jbn" event={"ID":"09c7e427-ef3b-4816-805d-d811be5b4337","Type":"ContainerDied","Data":"de129226a1ca306a25e24106e499754f5c6833838e7f052ecc0df93b296762b4"} Dec 07 17:14:59 crc kubenswrapper[4716]: I1207 17:14:59.431844 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h5jbn" event={"ID":"09c7e427-ef3b-4816-805d-d811be5b4337","Type":"ContainerDied","Data":"4b9061df2a3818fe85be9c7b3f14e78504b3e4b6892831340fbc8ce955e0261c"} Dec 07 17:14:59 crc kubenswrapper[4716]: I1207 17:14:59.431864 4716 scope.go:117] "RemoveContainer" containerID="de129226a1ca306a25e24106e499754f5c6833838e7f052ecc0df93b296762b4" Dec 07 17:14:59 crc kubenswrapper[4716]: I1207 17:14:59.431964 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h5jbn" Dec 07 17:14:59 crc kubenswrapper[4716]: I1207 17:14:59.448360 4716 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09c7e427-ef3b-4816-805d-d811be5b4337-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 07 17:14:59 crc kubenswrapper[4716]: I1207 17:14:59.448391 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-422j6\" (UniqueName: \"kubernetes.io/projected/09c7e427-ef3b-4816-805d-d811be5b4337-kube-api-access-422j6\") on node \"crc\" DevicePath \"\"" Dec 07 17:14:59 crc kubenswrapper[4716]: I1207 17:14:59.448403 4716 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09c7e427-ef3b-4816-805d-d811be5b4337-utilities\") on node \"crc\" DevicePath \"\"" Dec 07 17:14:59 crc kubenswrapper[4716]: I1207 17:14:59.454637 4716 scope.go:117] "RemoveContainer" containerID="95e647ce5c332e2dd3003af717199c35309de162794f5c314c33968464818561" Dec 07 17:14:59 crc kubenswrapper[4716]: I1207 17:14:59.475031 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-h5jbn"] Dec 07 17:14:59 crc kubenswrapper[4716]: I1207 17:14:59.484563 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-h5jbn"] Dec 07 17:14:59 crc kubenswrapper[4716]: I1207 17:14:59.492014 4716 scope.go:117] "RemoveContainer" containerID="e3fc6065725a0acadbcf14cf9930b9879eff82e15e623667dab42cfd3865d956" Dec 07 17:14:59 crc kubenswrapper[4716]: I1207 17:14:59.523565 4716 scope.go:117] "RemoveContainer" containerID="de129226a1ca306a25e24106e499754f5c6833838e7f052ecc0df93b296762b4" Dec 07 17:14:59 crc kubenswrapper[4716]: E1207 17:14:59.523900 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de129226a1ca306a25e24106e499754f5c6833838e7f052ecc0df93b296762b4\": container with ID starting with de129226a1ca306a25e24106e499754f5c6833838e7f052ecc0df93b296762b4 not found: ID does not exist" containerID="de129226a1ca306a25e24106e499754f5c6833838e7f052ecc0df93b296762b4" Dec 07 17:14:59 crc kubenswrapper[4716]: I1207 17:14:59.523932 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de129226a1ca306a25e24106e499754f5c6833838e7f052ecc0df93b296762b4"} err="failed to get container status \"de129226a1ca306a25e24106e499754f5c6833838e7f052ecc0df93b296762b4\": rpc error: code = NotFound desc = could not find container \"de129226a1ca306a25e24106e499754f5c6833838e7f052ecc0df93b296762b4\": container with ID starting with de129226a1ca306a25e24106e499754f5c6833838e7f052ecc0df93b296762b4 not found: ID does not exist" Dec 07 17:14:59 crc kubenswrapper[4716]: I1207 17:14:59.523957 4716 scope.go:117] "RemoveContainer" containerID="95e647ce5c332e2dd3003af717199c35309de162794f5c314c33968464818561" Dec 07 17:14:59 crc kubenswrapper[4716]: E1207 17:14:59.524222 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95e647ce5c332e2dd3003af717199c35309de162794f5c314c33968464818561\": container with ID starting with 95e647ce5c332e2dd3003af717199c35309de162794f5c314c33968464818561 not found: ID does not exist" containerID="95e647ce5c332e2dd3003af717199c35309de162794f5c314c33968464818561" Dec 07 17:14:59 crc kubenswrapper[4716]: I1207 17:14:59.524247 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95e647ce5c332e2dd3003af717199c35309de162794f5c314c33968464818561"} err="failed to get container status \"95e647ce5c332e2dd3003af717199c35309de162794f5c314c33968464818561\": rpc error: code = NotFound desc = could not find container \"95e647ce5c332e2dd3003af717199c35309de162794f5c314c33968464818561\": container with ID starting with 95e647ce5c332e2dd3003af717199c35309de162794f5c314c33968464818561 not found: ID does not exist" Dec 07 17:14:59 crc kubenswrapper[4716]: I1207 17:14:59.524263 4716 scope.go:117] "RemoveContainer" containerID="e3fc6065725a0acadbcf14cf9930b9879eff82e15e623667dab42cfd3865d956" Dec 07 17:14:59 crc kubenswrapper[4716]: E1207 17:14:59.524530 4716 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3fc6065725a0acadbcf14cf9930b9879eff82e15e623667dab42cfd3865d956\": container with ID starting with e3fc6065725a0acadbcf14cf9930b9879eff82e15e623667dab42cfd3865d956 not found: ID does not exist" containerID="e3fc6065725a0acadbcf14cf9930b9879eff82e15e623667dab42cfd3865d956" Dec 07 17:14:59 crc kubenswrapper[4716]: I1207 17:14:59.524645 4716 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3fc6065725a0acadbcf14cf9930b9879eff82e15e623667dab42cfd3865d956"} err="failed to get container status \"e3fc6065725a0acadbcf14cf9930b9879eff82e15e623667dab42cfd3865d956\": rpc error: code = NotFound desc = could not find container \"e3fc6065725a0acadbcf14cf9930b9879eff82e15e623667dab42cfd3865d956\": container with ID starting with e3fc6065725a0acadbcf14cf9930b9879eff82e15e623667dab42cfd3865d956 not found: ID does not exist" Dec 07 17:14:59 crc kubenswrapper[4716]: I1207 17:14:59.667114 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09c7e427-ef3b-4816-805d-d811be5b4337" path="/var/lib/kubelet/pods/09c7e427-ef3b-4816-805d-d811be5b4337/volumes" Dec 07 17:15:00 crc kubenswrapper[4716]: I1207 17:15:00.211296 4716 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29418795-rvk7r"] Dec 07 17:15:00 crc kubenswrapper[4716]: E1207 17:15:00.211799 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09c7e427-ef3b-4816-805d-d811be5b4337" containerName="registry-server" Dec 07 17:15:00 crc kubenswrapper[4716]: I1207 17:15:00.211831 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="09c7e427-ef3b-4816-805d-d811be5b4337" containerName="registry-server" Dec 07 17:15:00 crc kubenswrapper[4716]: E1207 17:15:00.211849 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09c7e427-ef3b-4816-805d-d811be5b4337" containerName="extract-utilities" Dec 07 17:15:00 crc kubenswrapper[4716]: I1207 17:15:00.211855 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="09c7e427-ef3b-4816-805d-d811be5b4337" containerName="extract-utilities" Dec 07 17:15:00 crc kubenswrapper[4716]: E1207 17:15:00.211873 4716 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09c7e427-ef3b-4816-805d-d811be5b4337" containerName="extract-content" Dec 07 17:15:00 crc kubenswrapper[4716]: I1207 17:15:00.211879 4716 state_mem.go:107] "Deleted CPUSet assignment" podUID="09c7e427-ef3b-4816-805d-d811be5b4337" containerName="extract-content" Dec 07 17:15:00 crc kubenswrapper[4716]: I1207 17:15:00.212105 4716 memory_manager.go:354] "RemoveStaleState removing state" podUID="09c7e427-ef3b-4816-805d-d811be5b4337" containerName="registry-server" Dec 07 17:15:00 crc kubenswrapper[4716]: I1207 17:15:00.213318 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29418795-rvk7r" Dec 07 17:15:00 crc kubenswrapper[4716]: I1207 17:15:00.215285 4716 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 07 17:15:00 crc kubenswrapper[4716]: I1207 17:15:00.215693 4716 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 07 17:15:00 crc kubenswrapper[4716]: I1207 17:15:00.219809 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29418795-rvk7r"] Dec 07 17:15:00 crc kubenswrapper[4716]: I1207 17:15:00.367474 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2c35ffc2-c0e8-482a-a670-1a51885a4609-config-volume\") pod \"collect-profiles-29418795-rvk7r\" (UID: \"2c35ffc2-c0e8-482a-a670-1a51885a4609\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418795-rvk7r" Dec 07 17:15:00 crc kubenswrapper[4716]: I1207 17:15:00.367570 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2c35ffc2-c0e8-482a-a670-1a51885a4609-secret-volume\") pod \"collect-profiles-29418795-rvk7r\" (UID: \"2c35ffc2-c0e8-482a-a670-1a51885a4609\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418795-rvk7r" Dec 07 17:15:00 crc kubenswrapper[4716]: I1207 17:15:00.367630 4716 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9hkt\" (UniqueName: \"kubernetes.io/projected/2c35ffc2-c0e8-482a-a670-1a51885a4609-kube-api-access-j9hkt\") pod \"collect-profiles-29418795-rvk7r\" (UID: \"2c35ffc2-c0e8-482a-a670-1a51885a4609\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418795-rvk7r" Dec 07 17:15:00 crc kubenswrapper[4716]: I1207 17:15:00.469054 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2c35ffc2-c0e8-482a-a670-1a51885a4609-config-volume\") pod \"collect-profiles-29418795-rvk7r\" (UID: \"2c35ffc2-c0e8-482a-a670-1a51885a4609\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418795-rvk7r" Dec 07 17:15:00 crc kubenswrapper[4716]: I1207 17:15:00.469637 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2c35ffc2-c0e8-482a-a670-1a51885a4609-secret-volume\") pod \"collect-profiles-29418795-rvk7r\" (UID: \"2c35ffc2-c0e8-482a-a670-1a51885a4609\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418795-rvk7r" Dec 07 17:15:00 crc kubenswrapper[4716]: I1207 17:15:00.469917 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2c35ffc2-c0e8-482a-a670-1a51885a4609-config-volume\") pod \"collect-profiles-29418795-rvk7r\" (UID: \"2c35ffc2-c0e8-482a-a670-1a51885a4609\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418795-rvk7r" Dec 07 17:15:00 crc kubenswrapper[4716]: I1207 17:15:00.470499 4716 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9hkt\" (UniqueName: \"kubernetes.io/projected/2c35ffc2-c0e8-482a-a670-1a51885a4609-kube-api-access-j9hkt\") pod \"collect-profiles-29418795-rvk7r\" (UID: \"2c35ffc2-c0e8-482a-a670-1a51885a4609\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418795-rvk7r" Dec 07 17:15:00 crc kubenswrapper[4716]: I1207 17:15:00.483865 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2c35ffc2-c0e8-482a-a670-1a51885a4609-secret-volume\") pod \"collect-profiles-29418795-rvk7r\" (UID: \"2c35ffc2-c0e8-482a-a670-1a51885a4609\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418795-rvk7r" Dec 07 17:15:00 crc kubenswrapper[4716]: I1207 17:15:00.492394 4716 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9hkt\" (UniqueName: \"kubernetes.io/projected/2c35ffc2-c0e8-482a-a670-1a51885a4609-kube-api-access-j9hkt\") pod \"collect-profiles-29418795-rvk7r\" (UID: \"2c35ffc2-c0e8-482a-a670-1a51885a4609\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29418795-rvk7r" Dec 07 17:15:00 crc kubenswrapper[4716]: I1207 17:15:00.533888 4716 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29418795-rvk7r" Dec 07 17:15:01 crc kubenswrapper[4716]: I1207 17:15:01.042571 4716 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29418795-rvk7r"] Dec 07 17:15:01 crc kubenswrapper[4716]: I1207 17:15:01.453171 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29418795-rvk7r" event={"ID":"2c35ffc2-c0e8-482a-a670-1a51885a4609","Type":"ContainerStarted","Data":"9fba3da65f1a2557a1647b4b55d1049c7657e632345b51b8596d3af13b8c289e"} Dec 07 17:15:01 crc kubenswrapper[4716]: I1207 17:15:01.453433 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29418795-rvk7r" event={"ID":"2c35ffc2-c0e8-482a-a670-1a51885a4609","Type":"ContainerStarted","Data":"0f727e6f1c35bc0b18163e4c14140d5706a4931f7f384be3ec7bfbc2815dff51"} Dec 07 17:15:02 crc kubenswrapper[4716]: I1207 17:15:02.463529 4716 generic.go:334] "Generic (PLEG): container finished" podID="2c35ffc2-c0e8-482a-a670-1a51885a4609" containerID="9fba3da65f1a2557a1647b4b55d1049c7657e632345b51b8596d3af13b8c289e" exitCode=0 Dec 07 17:15:02 crc kubenswrapper[4716]: I1207 17:15:02.463602 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29418795-rvk7r" event={"ID":"2c35ffc2-c0e8-482a-a670-1a51885a4609","Type":"ContainerDied","Data":"9fba3da65f1a2557a1647b4b55d1049c7657e632345b51b8596d3af13b8c289e"} Dec 07 17:15:02 crc kubenswrapper[4716]: I1207 17:15:02.856318 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29418795-rvk7r" Dec 07 17:15:03 crc kubenswrapper[4716]: I1207 17:15:03.039561 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j9hkt\" (UniqueName: \"kubernetes.io/projected/2c35ffc2-c0e8-482a-a670-1a51885a4609-kube-api-access-j9hkt\") pod \"2c35ffc2-c0e8-482a-a670-1a51885a4609\" (UID: \"2c35ffc2-c0e8-482a-a670-1a51885a4609\") " Dec 07 17:15:03 crc kubenswrapper[4716]: I1207 17:15:03.039632 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2c35ffc2-c0e8-482a-a670-1a51885a4609-secret-volume\") pod \"2c35ffc2-c0e8-482a-a670-1a51885a4609\" (UID: \"2c35ffc2-c0e8-482a-a670-1a51885a4609\") " Dec 07 17:15:03 crc kubenswrapper[4716]: I1207 17:15:03.039885 4716 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2c35ffc2-c0e8-482a-a670-1a51885a4609-config-volume\") pod \"2c35ffc2-c0e8-482a-a670-1a51885a4609\" (UID: \"2c35ffc2-c0e8-482a-a670-1a51885a4609\") " Dec 07 17:15:03 crc kubenswrapper[4716]: I1207 17:15:03.040478 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2c35ffc2-c0e8-482a-a670-1a51885a4609-config-volume" (OuterVolumeSpecName: "config-volume") pod "2c35ffc2-c0e8-482a-a670-1a51885a4609" (UID: "2c35ffc2-c0e8-482a-a670-1a51885a4609"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 07 17:15:03 crc kubenswrapper[4716]: I1207 17:15:03.044605 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c35ffc2-c0e8-482a-a670-1a51885a4609-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "2c35ffc2-c0e8-482a-a670-1a51885a4609" (UID: "2c35ffc2-c0e8-482a-a670-1a51885a4609"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 07 17:15:03 crc kubenswrapper[4716]: I1207 17:15:03.046311 4716 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c35ffc2-c0e8-482a-a670-1a51885a4609-kube-api-access-j9hkt" (OuterVolumeSpecName: "kube-api-access-j9hkt") pod "2c35ffc2-c0e8-482a-a670-1a51885a4609" (UID: "2c35ffc2-c0e8-482a-a670-1a51885a4609"). InnerVolumeSpecName "kube-api-access-j9hkt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 07 17:15:03 crc kubenswrapper[4716]: I1207 17:15:03.142649 4716 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2c35ffc2-c0e8-482a-a670-1a51885a4609-config-volume\") on node \"crc\" DevicePath \"\"" Dec 07 17:15:03 crc kubenswrapper[4716]: I1207 17:15:03.142694 4716 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j9hkt\" (UniqueName: \"kubernetes.io/projected/2c35ffc2-c0e8-482a-a670-1a51885a4609-kube-api-access-j9hkt\") on node \"crc\" DevicePath \"\"" Dec 07 17:15:03 crc kubenswrapper[4716]: I1207 17:15:03.142710 4716 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2c35ffc2-c0e8-482a-a670-1a51885a4609-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 07 17:15:03 crc kubenswrapper[4716]: I1207 17:15:03.473845 4716 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29418795-rvk7r" event={"ID":"2c35ffc2-c0e8-482a-a670-1a51885a4609","Type":"ContainerDied","Data":"0f727e6f1c35bc0b18163e4c14140d5706a4931f7f384be3ec7bfbc2815dff51"} Dec 07 17:15:03 crc kubenswrapper[4716]: I1207 17:15:03.474213 4716 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0f727e6f1c35bc0b18163e4c14140d5706a4931f7f384be3ec7bfbc2815dff51" Dec 07 17:15:03 crc kubenswrapper[4716]: I1207 17:15:03.474278 4716 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29418795-rvk7r" Dec 07 17:15:03 crc kubenswrapper[4716]: I1207 17:15:03.937932 4716 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29418750-c9svx"] Dec 07 17:15:03 crc kubenswrapper[4716]: I1207 17:15:03.945902 4716 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29418750-c9svx"] Dec 07 17:15:05 crc kubenswrapper[4716]: I1207 17:15:05.678047 4716 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15e3e08f-67c0-4d1b-8ecb-29a0f4d8ff44" path="/var/lib/kubelet/pods/15e3e08f-67c0-4d1b-8ecb-29a0f4d8ff44/volumes" Dec 07 17:15:35 crc kubenswrapper[4716]: I1207 17:15:35.874752 4716 scope.go:117] "RemoveContainer" containerID="daca886a24b8417a78bdbab31ee6a24649798d0954cfcb7350978fb0fb9b062c" Dec 07 17:15:35 crc kubenswrapper[4716]: I1207 17:15:35.919297 4716 scope.go:117] "RemoveContainer" containerID="09c4c07ec47e7636fa56fbacfd96493338786a439c5c3a28517802067e917e5b" Dec 07 17:15:35 crc kubenswrapper[4716]: I1207 17:15:35.966802 4716 scope.go:117] "RemoveContainer" containerID="fea9aa453d0c45ecb1b1b8f6769cd011b4d84a1cc7c2c7206727e00f47e6dec5" Dec 07 17:15:35 crc kubenswrapper[4716]: I1207 17:15:35.998711 4716 scope.go:117] "RemoveContainer" containerID="032e735c2ca329ea227a0fed156105ba64b353b9c4fc73e494f7a67bd19d9e75" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515115333145024446 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015115333146017364 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015115322113016477 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015115322113015447 5ustar corecore